PLUS: A scathing insider account of Azure's failures and OpenAI buys a popular tech talk show

Good morning

Google is bringing state-of-the-art AI capabilities directly into your personal devices with the release of its new Gemma 4 family. These new open models are designed for major efficiency, allowing powerful AI to run on everything from mobile phones to high-end GPUs.

By removing the need for massive hardware overhead, this release makes building complex AI workflows more accessible and private. What will this new wave of on-device applications look like now that state-of-the-art AI can truly run locally?

In today’s Next in AI:

  • Google's efficient on-device Gemma 4 model

  • A scathing insider account of Azure's failures

  • The AI that learns to walk in Arc Raiders

  • OpenAI's acquisition of a popular tech talk show

Google's Pocket-Sized Powerhouse

Next in AI: Google has released Gemma 4, a new family of open models designed to deliver state-of-the-art AI capabilities efficiently on everything from mobile devices to high-end GPUs.

Explained:

- Spanning four sizes from an efficient 2 billion parameters to a powerful 31 billion, the new models punch well above their weight, with the largest variant ranking as the #3 open model on the Arena AI text leaderboard.

- The smaller E2B and E4B models are engineered for on-device tasks, running offline with near-zero latency on hardware like Android phones and Jetson Orin Nano modules.

- Released under a commercially-friendly Apache 2.0 license, the models are optimized for local deployment on NVIDIA GPUs using popular tools like Ollama and llama.cpp.

Why It Matters:

Gemma 4 makes building powerful, agentic AI workflows more accessible by removing the need for massive hardware overhead. This shift accelerates the development of more private and responsive applications that run locally on your personal devices.

Azure's Trillion-Dollar Stumble

Next in AI: A former senior engineer from Microsoft's Azure Core team published a scathing insider account detailing systemic failures that he claims put OpenAI's services and national security at critical risk.

Explained:

- The engineer describes an organization on a "death march," bogged down by an unmanageable stack of 173 management agents per server and pursuing an impossible plan to port large Windows components to a tiny, low-power chip.

- This technical chaos allegedly created an unstable environment with severe scaling limits, causing performance issues that directly jeopardized mission-critical infrastructure for top customers like OpenAI and the US government.

- These explosive claims coincide with Microsoft's worst financial quarter since 2008, as investors grow anxious about the return on the company's massive spending on AI infrastructure.

Why It Matters: This account peels back the curtain on the immense engineering challenges required to power the global AI boom. It's a stark reminder that the foundational cloud platforms supporting the industry may be more fragile than they appear.

The AI That Learns to Walk

Next in AI: The breakout success of the game Arc Raiders is powered by an innovative enemy AI that uses reinforcement learning and physics-based robotics. This approach creates dynamic, unpredictable opponents that feel fundamentally different from traditional scripted game characters.

Explained:

- Embark Studios' system blends machine learning with traditional AI, using reinforcement learning for complex physical locomotion like walking and balancing, while behavior trees handle higher-level strategic decisions.

- While enemies feel like they learn from players in real-time, their adaptability actually comes from extensive offline training in randomized scenarios, making them robust enough to handle unexpected in-game events.

- This technology was a massive undertaking, requiring a team of five to ten people working for years on cutting-edge robotics research, and was nearly cut multiple times due to its complexity.

Why It Matters: This success signals a major shift in interactive entertainment, moving away from predictable, scripted encounters toward dynamic, physics-based systems. The approach creates more engaging and replayable experiences by allowing for emergent behaviors that even the developers didn't anticipate.

OpenAI Buys the Broadcaster

Next in AI: OpenAI is officially moving into media with the acquisition of TBPN, the popular daily tech talk show, in a first-of-its-kind purchase for the AI giant.

Explained:

- The move is a strategic play to shape the public narrative around AI, leveraging TBPN's established platform to host a more constructive global dialogue.

- To maintain credibility, OpenAI guarantees full editorial independence, allowing TBPN to continue choosing its own guests and topics without interference.

- Beyond the show, the TBPN team will join OpenAI's Strategy organization to help innovate on marketing and better explain AI's impact to a mainstream audience.

Why It Matters: This acquisition signals a major AI developer is now treating media and public perception as a core part of its strategy, not just a PR function. By owning a media channel, OpenAI can more directly influence the global conversation about AI's future.

AI Pulse

Arcee AI unveiled Trinity Large Thinking, a powerful open-source reasoning model showing strong performance in agentic workloads and reasoning tasks.

Stanford found that AI chatbots are overly agreeable when giving interpersonal advice, affirming users' behavior even when harmful and making them more self-centered.

Monarch Tractor collapsed, with the once-hyped AI autonomous tractor startup abandoning its Bay Area headquarters and laying off its staff after burning through over $240M in funding.

MIT developed an AI model capable of classifying and quantifying six different types of atomic defects in materials simultaneously, a task previously impossible with conventional techniques.

Keep Reading