Chinese AI Model Tops GPT with 8-Hour Coding RunsPlus: Gemma 4 Goes Fully Open, Microsoft Drops Faster AI Models, Stanford AI Index Reveals Trust Gap, and more.Hello Engineering Leaders and AI Enthusiasts! This newsletter brings you the latest AI updates in just 4 minutes! Dive in for a quick summary of everything important that happened in AI over the last week. And a huge shoutout to our amazing readers. We appreciate you😊 In today’s edition:
Let’s go! Z AI’s GLM-5.1 model beats GPT in codingZ AI has released GLM-5.1, a new open-source coding model that is competing directly with top-tier systems and even outperforming them on key benchmarks. The model scored higher than GPT-5.4 and Opus 4.6 on SWE-Bench Pro, marking a rare moment where an open model leads the leaderboard. Beyond benchmarks, GLM-5.1 is built for long-running, autonomous tasks. In testing, it completed an 8-hour coding session to build a full Linux desktop experience as a web app handling everything from file systems to terminal interactions without human input. Why does it matter? Chinese labs continue to push forward, with GLM-5.1 combining strong coding performance with long-horizon task capabilities. Instead of just solving short prompts, models are now handling complex tasks over hours without breaking, something that wasn’t possible before. And with an open model now outperforming GPT-5.4 on benchmarks, the gap between open-source and top closed models is shrinking fast. Google launches Gemma 4 open modelsGoogle DeepMind has released Gemma 4, a new family of open AI models designed to run across everything from smartphones to full-scale systems. The models support coding, vision, and multi-step agent workflows, with smaller variants even capable of running offline with voice support. Performance-wise, the larger Gemma 4 models are competitive with leading open models, despite being smaller in size. But the bigger shift is licensing, Google has moved to Apache 2.0, allowing developers to freely modify, deploy, and commercialize the models without restrictions. Why does it matter? Open models have largely been led by Chinese labs, but Gemma 4 signals a stronger push from the U.S. side to compete on openness. Google moving to a fully permissive license stands in contrast to others tightening control, suggesting the next phase of competition may be shaped as much by licensing as by model performance. Microsoft drops faster, cheaper AI modelsMicrosoft has introduced three new MAI models across speech, voice, and image generation, aiming to compete on speed, quality, and price. The lineup includes Transcribe-1 for speech-to-text, Voice-1 for realistic voice generation, and Image-2 for high-quality visuals, all available through Microsoft Foundry. The models are designed for real-world performance. Transcribe-1 delivers faster and more accurate transcription across multiple languages; Voice-1 can generate expressive speech and clone voices from short samples, and Image-2 produces visuals up to 2× faster. The bigger push is clear: enterprise-ready AI models that balance performance with aggressive pricing. Why does it matter? When every modality becomes cheap and fast together, building multimodal AI stops being a premium capability and becomes baseline. That shift puts pressure on competitors and forces developers to optimize less for capability and more for cost, scale, and integration. Stanford’s report finds AI hits 53% adoption, trust at 31%Stanford’s AI Index Report 2026 shows that AI has now reached over half the global population faster than both the PC and the internet. But while usage is surging, sentiment is moving in the opposite direction. The gap is widening nearly 75% of AI experts are optimistic about jobs, but only 23% of the public agrees. Entry-level developer jobs (ages 22–25) have already dropped ~20% since 2024, even as demand for senior engineers grows. Meanwhile, the U.S. leads in building AI but ranks just 24th in adoption (28.3%), with regions like Southeast Asia pulling ahead. |