Co-Founder
A wild weekend in AI — OpenAI pushed back GPT-5, Meta surprise-dropped three new Llama 4 models, and Microsoft decided it was time Copilot started remembering your dog’s birthday.
Let’s get into it.
Sam Altman says GPT-5 is harder than expected — OpenAI is holding off a bit longer to integrate all the performance gains they uncovered mid-build. But here’s the upside: they’re splitting out two models originally meant to be internal — o3 and o4-mini — and releasing them soon.
Bonus: o3 has reportedly hit top-tier coder status in internal testing.
Also, GPT’s Deep Research tool may be coming to free users, and OpenAI’s legal battle with the NYT just got more serious — a judge is letting the copyright claims move forward after calling OpenAI’s defense a “straw man.” Yikes.
Meta didn’t wait until Monday — they dropped Llama 4 over the weekend. Here’s the quick rundown:
All of them use MoE (mixture-of-experts) architecture, but there’s a bit of controversy around benchmark results vs. public releases — some say Meta optimized specifically for test environments.
Notably: EU devs are blocked from using Llama 4 due to AI Act uncertainty.
Microsoft isn’t sitting still either. Copilot is getting major personalization features:
It’s more like a personal assistant now — though it's still unclear if this makes users want to switch away from their current tools.
That’s it for today — but if I had to leave you with one takeaway?
This week proved that benchmarks ≠ real-world results.
The best models on paper might not be the best in your product. Stay curious, test often, and don’t get swept up in the leaderboard hype.
Talk soon,
David
P.S. Anyone else feel like "GPT-5 delayed" is the AI version of “album pushed back but it's gonna be fire”? 🎧