Latest advancements in AI from Google's Gemini and DeepMind, OpenAI's memory and Sora, to SoftBank's ambitious chip venture and Reddit's smart licensing.

Last Week in AI: Episode 19

Welcome to this week’s Last Week in AI! We’ve got a bunch of cool AI stuff to talk about. From Google making moves with its file-identifying wizard Magika, to SoftBank getting ready to shake up the AI chip game, and even Reddit making a smart play with a new licensing deal. It’s been a busy week, and we’re here to break it all down for you.



Google’s latest AI, Gemini 1.5 Pro, outperforms its predecessor with improved efficiency and advanced capabilities. Here’s what stands out:

  1. More Efficient: Uses less compute power for the same quality.
  2. Longer Context: Handles up to 1 million tokens for deep understanding.
  3. Superior Performance: Beats the previous model on 87% of benchmarks.
Why It Matters

Gemini 1.5 Pro offers faster, deeper analysis of massive data. It enables complex problem-solving and innovation in AI applications, making advanced AI tools more accessible to developers and enterprises.


Google DeepMind and USC have developed SELF-DISCOVER, a new framework enhancing LLM reasoning abilities. Key points:

  1. Significant Performance Boost: Up to 32% better than traditional Chain of Thought methods (a technique that guides LLMs to follow a reasoning process when dealing with hard problems).
  2. Autonomous Reasoning: LLMs self-discover reasoning structures for complex problem-solving.
  3. Broad Implications: Marks progress towards general intelligence and advanced AI capabilities.
Why It Matters

SELF-DISCOVER represents a major advancement in AI, offering a more sophisticated approach to reasoning tasks. This framework could revolutionize how AI understands and interacts with the world, pushing closer to achieving general intelligence.


Google has released Magika, an AI-driven system for identifying file types, to the open-source community. Highlights include:

  1. High Performance: Utilizes a deep-learning model for rapid, accurate file-type identification on a CPU.
  2. Superior Accuracy: Achieves a 20% improvement over current tools on a diverse 1M files benchmark.
  3. Community Contribution: Available on GitHub under Apache2 License, enhancing file identification for software and cybersecurity.
Why It Matters

Magika’s open-sourcing represents a significant advancement in file identification, crucial for cybersecurity and data management. By offering a more precise tool freely, Google fosters innovation and security enhancements across the tech ecosystem.



OpenAI has introduced memory capabilities to ChatGPT for a select user group, enhancing personalization and context relevance. Key highlights include:

  1. User-Controlled Memory: Options to turn off, delete selectively, or clear all memories.
  2. Personalized Interactions: Memory evolves with user interactions, not tied to specific conversations.
  3. Selective Rollout: Available to a limited number of users, with plans for broader access.
Why It Matters

This feature marks a leap in AI conversational agents, promising more efficient, personalized interactions. It benefits enterprises, teams, and developers by retaining context and preferences, paving the way for advanced AI applications.

Credit: OpenAI


OpenAI’s Sora, an AI model, transforms prompts into realistic videos up to a minute long. Here’s the breakdown:

  1. Advanced Capabilities: Generates complex scenes with accurate motion and emotions.
  2. Language Understanding: Deeply interprets prompts for vivid character and scene creation.
  3. Safety Measures: Includes adversarial testing and tools to detect misleading content.
Why It Matters

Sora represents a major step towards AI that simulates real-world interactions, aiming for AGI. Its blend of visual quality and language understanding opens new possibilities for creative and problem-solving applications, despite challenges in physics simulation and cause-effect understanding.

Credit: OpenAI


Groq has teamed up with Samsung to create cutting-edge AI silicon. Here’s what you need to know:

  1. Advanced Manufacturing: Utilizing Samsung’s 4nm process for better performance and efficiency.
  2. Tensor Streaming Architecture: First-gen technology boosting power and memory capabilities.
  3. Scalability: Enables systems from 85,000 to over 600,000 chips without external switches.
Why It Matters

This collaboration pushes the envelope in AI and machine learning, promising revolutionary solutions for AI, HPC (High-Performance Computing), and data centers. It underscores Groq’s commitment to high-quality, fast-to-market innovations, leveraging Samsung’s manufacturing prowess.


Amazon researchers have developed BASE TTS, the largest text-to-speech model to date, with 980 million parameters. Highlights include:

  1. Massive Training: Leveraged up to 100,000 hours of speech data for training.
  2. Optimal Size Insights: Found that a 400 million parameter model showed significant improvements without further gains at 980 million parameters.
  3. Efficiency: Designed for low-bandwidth streaming, separating emotional and prosodic data.
Why It Matters

BASE TTS aims to refine text-to-speech technology, focusing on natural sound and efficiency. Despite its size, the quest for the optimal model size for emergent abilities continues, offering a path toward more accessible and versatile speech synthesis applications.

Project Izanagi

Masayoshi Son of SoftBank is eyeing a $100 billion venture, Izanagi, to enter the AI chip market, challenging Nvidia. Here’s the scoop:

  1. Massive Funding: Aiming for $100 billion, with $70 billion from Middle East investors and $30 billion from SoftBank.
  2. Arm Collaboration: Plans to partner with Arm for chip design, leveraging its recent public spin-off.
  3. Strategic Shift: Reflects SoftBank’s pivot towards AI, fueled by divesting Alibaba stakes for AI investments.
Why It Matters

Son’s ambitious venture signals a significant shift in the AI landscape, aiming to offer an alternative to Nvidia’s dominance. With AI’s growing importance, Izanagi represents a strategic move to capitalize on this burgeoning market, amidst SoftBank’s broader focus on AI and its return to profitability.


Reddit has inked a $60 million licensing deal with a major AI company for its content. Key details include:

  1. Valuable Partnership: The deal, worth $60 million annually, grants AI access to Reddit’s vast user-generated content.
  2. Strategic Move: Aims to navigate legalities of AI training with web content, reflecting Reddit’s assertive negotiation stance.
  3. Public Offering Plans: Coincides with Reddit’s IPO ambitions, seeking a $5 billion valuation despite a recent market downturn.
Why It Matters

This agreement underscores the growing importance of user-generated data in AI development, marking a pivotal move for Reddit amidst its financial and strategic repositioning. It also highlights the platform’s leverage in the evolving digital and AI landscapes.

Until Next Week

And just like that, we’re at the end of another week in the world of AI. Not bad, am I right? Every week, AI is getting smarter, faster, and a bit more into our daily lives. Can’t wait to see what’s next. Catch you in the next update!

Last Week in AI: Episode 19 Read More »