Episode Summary
ThursdAIβs Apr 2 episode is a high-signal mix of breaking AI news and hands-on operator analysis: the Claude Code leak fallout with Sigrid Jin, Gemma 4 coverage with Omar Sanseviero from Google DeepMind, and a rapid sweep of major lab moves. The panel unpacks what the leak actually revealed, where SessionGate friction is hurting real workflows, and how open-model momentum is reshaping tooling choices. They also cover OpenAIβs reported $122B funding round, Microsoftβs model push, and fresh benchmarks around agent harness performance. Itβs a practical episode for builders trying to separate hype from what is immediately usable.
In This Episode
- β‘ Intro
- π° TL;DR - This Week in AI
- π₯ Claude Code Leak
- π€ Claude Code Session Gate
- π₯ Gemma 4 Breaking News
- π° OpenAI $122B Funding Round
- π’ Microsoft AI Models
- π Gemma 4 with Omar Sanseviero
- π Gemma 4 - Google's Open Source Strategy
- π οΈ Gemma 4 - Agentic Capabilities & Local Models
- β‘ Gemma 4 - Recap & Community Features
- β‘ Gemma 4 - Community Reactions
- β‘ This Week's Buzz - Ralph Hackathon
- β‘ This weeks buzz - Wolfbench showes Hermes is better than OpenClaw
- π§ͺ Wolf Bench - Hermes Agent Results
- π§ͺ One-Bit Quantization (Prism ML)
- π Alibaba Qwen 3.6 & Wan 2.7
- π Fish Audio Speech-to-Text
- π₯ Google Veo 3.1 Light
- π€ Agent Harnesses & Open Claw
- π§ͺ Anthropic Emotion Vectors in Claude
- β‘ Outro
Hosts & Guests
By The Numbers
π₯ Breaking During The Show
β‘ Intro
The panel opens with a packed week in AI and quickly frames the two big arcs of the episode: the Claude Code leak drama and the wave of open model launches. The co-hosts set context for why this week felt unusually consequential for both builders and labs.
- Fast-moving week spanning model releases, leaks, and funding
- Set-up for deep dives on Claude Code and Gemma 4
π° TL;DR - This Week in AI
Alex runs a rapid-fire headline pass across major launches, funding news, and tool updates. It sets the backbone for the rest of the show before the conversation slows down into technical analysis.
- OpenAI, Anthropic, Google, Microsoft, and Alibaba all had major updates
- Episode structure moves from headlines to deeper operator discussion
π₯ Claude Code Leak
Sigrid Jin joins to unpack the clean-room Claude Code rewrite and the GitHub blow-up around the leak, including what was actually learned from the exposed package internals. The segment distinguishes technical facts from social-media exaggeration.
- Leak discourse separated from verifiable implementation details
- Sigrid explains why the clean-room rewrite mattered to developers
π€ Claude Code Session Gate
The hosts discuss SessionGate complaints, resume/caching quirks, and cost instability reports from power users. The thread focuses on reliability and trust when teams run long-lived coding sessions.
- Reported session-resume behavior can multiply costs
- Panel emphasizes observability and guardrails for agent workflows
π₯ Gemma 4 Breaking News
Breaking updates on Gemma 4 land mid-show, and the panel reacts in real time to where it sits against frontier proprietary models. The focus is practical: what can be run now, where, and for which workloads.
- Gemma 4 framed as a serious new open-model contender
- Immediate discussion of deployment and tradeoffs
π° OpenAI $122B Funding Round
Alex and co-hosts analyze OpenAIβs reported $122B raise and what that scale of capital implies for infra, product velocity, and competitive pressure on the rest of the market.
- $122B described as a historic financing event
- Discussion ties fundraising to model deployment economics
π’ Microsoft AI Models
The team reviews Microsoftβs in-house model push across transcription, image generation, and voice. They compare positioning versus specialist products and foundation-model APIs.
- Microsoft expanding first-party model stack
- Panel compares quality and differentiation across modalities
π Gemma 4 with Omar Sanseviero
Omar Sanseviero (Google DeepMind) joins to explain Gemma 4 from a builder and ecosystem perspective. Conversation covers model intent, community adoption, and practical entry points.
- Confirmed guest Omar provides launch context from Google DeepMind
- Focus on ecosystem and developer experience around Gemma 4
π Gemma 4 - Google's Open Source Strategy
The panel zooms out on Googleβs open-source strategy: why Gemma exists alongside closed systems and how that dual strategy may shape developer mindshare.
- Open vs closed strategy discussed as portfolio decision
- Community leverage and trust highlighted as differentiators
π οΈ Gemma 4 - Agentic Capabilities & Local Models
This section explores Gemma 4 in agentic setups and local workflows, including how smaller deployable models fit into multi-agent pipelines.
- Local-first and agentic uses highlighted
- Model size/perf tradeoffs discussed for real workloads
β‘ Gemma 4 - Recap & Community Features
A recap segment summarizes Gemma 4βs strongest points and where the community can contribute or extend the stack. The hosts emphasize experimentation over benchmark-chasing.
- Community contribution pathways discussed
- Recap connects product features to real usage
β‘ Gemma 4 - Community Reactions
The hosts review early reactions from builders and X/Twitter to gauge momentum and skepticism around the release claims.
- Early sentiment analysis from AI community
- Balance of excitement and skepticism captured
β‘ This Week's Buzz - Ralph Hackathon
The buzz segment highlights a hackathon demo and what it signals about agent UX and rapid prototyping culture around open tooling.
- Ralph hackathon mentioned as high-signal community experiment
- Demonstrates speed of current AI maker ecosystem
β‘ This weeks buzz - Wolfbench showes Hermes is better than OpenClaw
Wolfram brings WolfBench findings that stirred debate, especially claims around Hermes performance versus OpenClaw in specific harness conditions.
- WolfBench data used to challenge assumptions
- Benchmark methodology caveats called out
π§ͺ Wolf Bench - Hermes Agent Results
The panel spends additional time on Hermes agent results and interpretation quality, emphasizing reproducibility and fair harness configuration.
- Hermes/OpenClaw comparison dissected in detail
- Reproducible eval setup framed as essential
π§ͺ One-Bit Quantization (Prism ML)
A quick technical dive on Prism MLβs one-bit quantization ideas and where aggressive compression might unlock cheaper inference.
- One-bit quantization positioned as cost/performance lever
- Discussion focuses on practical deployment constraints
π Alibaba Qwen 3.6 & Wan 2.7
The open-source round-up covers Alibabaβs Qwen 3.6 and Wan 2.7 updates, with attention to multimodal capability and practical ranking versus other open models.
- Qwen family momentum remains strong
- Wan 2.7 noted in broader open ecosystem context
π Fish Audio Speech-to-Text
The team reviews Fish Audioβs speech stack progress and where it challenges incumbent speech providers for developers building voice workflows.
- Speech model quality and developer ergonomics discussed
- Voice tooling seen as rapidly commoditizing
π₯ Google Veo 3.1 Light
Googleβs Veo 3.1 Light segment focuses on video-gen quality/speed tradeoffs and likely creator workflows it unlocks.
- Veo 3.1 Light discussed as practical video model tier
- Quality-vs-latency tradeoff highlighted
π€ Agent Harnesses & Open Claw
The conversation broadens to agent harness architecture, safety boundaries, and why evaluation discipline matters more as capabilities accelerate.
- Harness design framed as central infra question
- Safety and control layers discussed with concrete examples
π§ͺ Anthropic Emotion Vectors in Claude
A late segment explores Anthropic emotion-vector work and what it might mean for steerability, user experience, and model behavior interpretability.
- Emotion-vector concept explained in practical terms
- Implications for controllability and UX debated
β‘ Outro
The show closes with follow-ups, next-week hooks, and a recap of the biggest takeaways: Claude Code leak fallout, Gemma 4 momentum, and funding-fueled competition.
- Strong close around builder-relevant takeaways
- Sets up next episode themes
Show Notes & Guests
Alex Volkov - AI Evangelist & Weights & Biases / CoreWeave (@altryne)
Co Hosts - @WolframRvnwlf @yampeleg @nisten @ldjconfirmed @ryancarson
Sigrid Jin (@realsigridjin) & Bellman (@bellman_ych) β creators of claw-code, fastest GitHub repo to 100K stars
Omar Sanseviero (@osanseviero) β DevEx at Google DeepMind, Gemma 4 launch
Ralphton Hackathon video (TikTok)
WolfBench.ai β agent harness benchmarking (Site)
Ryanβs Claw Chief open source setup (GitHub)
Big CO LLMs + APIs
Claude Codeβs entire 512K-line source code accidentally leaked via npm β revealing KAIROS daemon, Undercover Mode, Buddy System, anti-distillation protections, and unreleased model references (Alexβs thread, Fried_riceβs discovery, VentureBeat)
Anthropic SessionGate continues β cache bugs reverse-engineered, --resume flag causes 10-20x cost increase, silent OpusβSonnet fallback reported (Alexβs cache bug post, Alexβs quota post, Reddit investigation, GitHub analysis)
OpenAI closes $122 billion funding round β largest in history, $852B valuation, IPO incoming (X, Breakdown)
OpenAI acquires TBPN β live tech media show, rumored low hundreds of millions
Microsoft MAI drops 3 in-house models β #1 transcription (MAI-Transcribe-1), #3 image gen (MAI-Image-2), expressive voice (MAI-Voice-1) (Mustafa post, Transcribe blog, Image blog)
Alibaba Qwen3.6-Plus β near-Opus 4.5 agentic coding, 1M context (X, Blog)
Cursor 3 β agent-first rebuild, no longer VS Code fork, parallel cloud/local agents (X, Blog)
Anthropic publishes emotion vector research β desperate Claude cheats more, calm Claude cheats less (X, Alexβs reaction)
Open Source LLMs
Google Gemma 4 β Apache 2.0, 31B / 26B MOE / 8B / 5B, local-friendly, agentic tool use, 256K context (HF Collection, try in AI Studio)
PrismML Bonsai 1-bit models β 8B in 1.15 GB, 10x intelligence density, 34 years of research (X, HF, Site)
Liquid AI LFM2.5-350M β agentic tool calling at 350M params, under 500MB quantized (X, HF, Blog)
Alibaba Qwen3.5-Omni β native omni-modal (text, image, audio, video), 397B total / 17B active (X, Blog)
Tools & Agentic Engineering
Claw-code β Claude Code leak backup β clean room rewrite β fastest repo to 100K+ stars (GitHub)
WolfBench results: Hermes Agent outperforms Claude Code and OpenClaw on Terminal Bench 2.0 (WolfBench.ai)
Ryan Carson open sources Claw Chief β AI chief of staff with skills, crons, scheduling (GitHub)
Vision & Video
Voice & Audio
AI Art & Diffusion
This Weekβs Buzz
Ralphton hackathon at W&B SF β humans write specs, AI builds, touch your laptop = lobster of shame (Alexβs video, TikTok)
WolfBench update β Hermes Agent > Claude Code on most model combos