GPT-5.5 ships; the takes are not glowing

April 24, 2026

17 topics across 15 blog posts, 9 newsletter stories, and 27 YouTube videos.

AI Models Hot Take
Theo - t3.gg AICodeKing The AI Daily Brief Simon Willison

GPT-5.5 lands; the reviews aren't glowing

OpenAI's GPT-5.5 hit the top of Artificial Analysis's Intelligence Index a day after launch, but Friday's verdicts from independent reviewers were ruthless. Theo (t3.gg) called it "lazy," "card-slop"-prone, and price-gouging at $5/$30 per 1M tokens (~20% over Opus 4.7)[1]Theo - t3.gg, I don't really like GPT-5.5…. AICodeKing's KingBench 2.0 head-to-head crowned Opus 4.7 the overall winner; GPT-5.5 had mixed wins; DeepSeek V4 Pro was weakest[2]AICodeKing, GPT-5.5 VS Deepseek V4 Pro VS Opus 4.7: KingBench 2.0. Even GPT-5.5's defenders concede the launch story is really GPT-5.5 Pro, which Theo says solved Defcon puzzles unsolved for 5–10 years[1]Theo - t3.gg, I don't really like GPT-5.5…. Simon Willison shipped llm 0.31 the same day with GPT-5.5 support and verbosity controls[3]Simon Willison, llm 0.31.

Read more

The Daily Brief's reading: a new standard, with caveats

The AI Daily Brief frames GPT-5.5 as the new "knowledge work" standard at the top of the Intelligence Index, but flags that ~04:02 Opus 4.7 still leads on Vending Bench and SWE-bench Pro (the latter disputed), as well as professional task benchmarks across finance, medical, and legal domains[4]The AI Daily Brief, What I Learned Testing GPT 5 5. On pricing ~05:03, GPT-5.5 is double GPT-5.4 and 20% above Opus 4.7; OpenAI's defense is "intelligence per dollar" rather than per-token. Coding wins are real ~11:05 — top of LiveBench, 31-hour autonomous runs, 79.2% expected-issue catch on CodeRabbit (vs. 58.3% baseline) — but for design work, the host's emerging recipe is "Opus to plan, GPT-5.5 to execute" ~14:07. The hot-take section ~30:15 calls 5.5 OpenAI's "o1 moment" — a new RL checkpoint, not a ceiling.

Theo's complaint list — and why he's the outlier

Theo's 30-minute teardown is brutal in places ~13:07: 5.5 "technically does what you ask, but just barely — like a hacker rushing to close a Jira ticket"[1]Theo - t3.gg, I don't really like GPT-5.5…. Frontend output is full of unnecessary card UI ~09:04. Once bad info enters the context window, you can't prompt it out — ~18:14 "you have to kill the thread and start over." On the bright side, ~20:14 GPT-5.5 Pro "cooked": Theo solved three previously-unsolved Defcon puzzles and a custom 163-minute cipher challenge with it. He concedes ~16:13 that Cursor's Michael Truell, Lovable, and Cognition all praise 5.5's persistence and bug detection — "I might be the outlier."

Once wrong info or a bad behavior enters 5.5's context, you can't prompt it away — you have to kill the thread and start over. — Theo

AICodeKing's KingBench 2.0: Opus wins overall, DeepSeek surprises in 3D

The KingBench tasks (elevator simulator, 3D contact lens case, folding table with slider, SVG panda eating a burger, bow-and-arrow game, math, Gemma 4 fine-tuning) produced a clear leaderboard: Opus 4.7 won most categories — including the elevator ~03:07, SVG panda ~05:08, and bow-and-arrow game; DeepSeek V4 Pro took the folding table ~04:07; GPT-5.5 was middling and still defaulted to ugly card UI[2]AICodeKing, GPT-5.5 VS Deepseek V4 Pro VS Opus 4.7: KingBench 2.0. The math question and Gemma 4 fine-tuning task ~06:08 defeated all three. AICodeKing's pricing rant ~07:09: GPT-5.5 would need to be a "16 trillion parameter model" to justify its price relative to DeepSeek V4 Flash at $0.04 in / $0.284 out per 1M tokens.

Tooling lag

There is no public GPT-5.5 API yet — OpenAI says it needs different safeguards for scale serving[1]Theo - t3.gg, I don't really like GPT-5.5…. Theo notes ~19:14 the only blessed way to test 5.5 in code is through the Codex endpoint backdoor (used by OpenClaw, Claude Code 2, JetBrains, Xcode, Open Code, Pi). The ChatGPT web app itself ~22:17 is "broken" for long Pro runs — pages freeze, runs return huge payloads, refreshes barely help. Recommendation ~25:19: prompt harder, do research in separate threads, and start new threads way more than usual. Simon Willison's llm 0.31 release adds a verbosity parameter for the GPT-5+ family, since the new generation of "reasoning" output benefits from explicit length control[3]Simon Willison, llm 0.31.

Tools: GPT-5.5, GPT-5.5 Pro, Opus 4.7, DeepSeek V4 Pro, DeepSeek V4 Flash, Codex, llm CLI, OpenClaw, Claude Code 2
AI Tools AI Models
OpenAI OpenAI OpenAI OpenAI OpenAI

OpenAI's launch sidekicks: Perplexity, NVIDIA, and ChatGPT Workspace agents

OpenAI flooded YouTube with five coordinated launch videos. Two are partner testimonials with hard numbers: a Perplexity engineer claims GPT-5.5 cuts 56% of tokens on agentic computer-use workflows[5]OpenAI, Introducing GPT-5.5 with Perplexity; NVIDIA's Shaurya Joshi reports 10× faster end-to-end ML research cycles[6]OpenAI, Introducing GPT-5.5 with NVIDIA's AI Researcher. The other three videos launch ChatGPT Workspace agents: a scheduled weekly metrics agent[7]OpenAI, Workspace agents in ChatGPT: Weekly metrics reporting agent, "Slate" for software procurement triage[8]OpenAI, Workspace agents in ChatGPT: Software review agent, and "Trove" for vendor due diligence — built on a workflow OpenAI's own finance team runs internally[9]OpenAI, Workspace agents in ChatGPT: Third-party risk management agent.

Read more

Partner testimonials: Perplexity and NVIDIA

Perplexity: an internal engineer says they had been deferring an internal tool because they expected it to take days; with Codex on GPT-5.5 they shipped it in under an hour, and saw 56% fewer tokens used on the same complex agentic computer-use tasks compared to prior models[5]OpenAI, Introducing GPT-5.5 with Perplexity. NVIDIA: AI researcher Shaurya Joshi describes GPT-5.5 handling abstract ideation (the model unprompted suggested adding a knowledge graph), writing ML infrastructure scripts, and refactoring codebases autonomously — adding up to a 10× speedup on running experiments end-to-end[6]OpenAI, Introducing GPT-5.5 with NVIDIA's AI Researcher.

ChatGPT Workspace agents — three flavors of the same abstraction

The three Workspace videos share a common architecture: each agent uses an "agent-owned" connection (analogous to a service account) so the agent can pull its own data without a human in the loop, ships with a reusable "skills" abstraction (best-practice instruction sets), and exposes a run-trace observability UI. The differences are in what the agent does:

The pattern is clear: ChatGPT is being repositioned from a chat surface into a back-office automation platform with first-party connectors, scheduling, and a cataloging primitive ("skills"). The named agents (Slate, Trove) suggest more product-flavored launches are coming — these aren't just reference workflows.

Tools: GPT-5.5, Codex, ChatGPT Workspace agents, Slack, Jira, Google Drive, Slate, Trove
AI Models Industry
Artificial Analysis Simon Willison AICodeKing Github Awesome

DeepSeek V4 Pro and Flash crash the open-weights frontier

DeepSeek dropped V4 Pro and V4 Flash with 1M-token context windows — an 8× jump over V3.2 — and pricing that makes the Western frontier look complacent. V4 Pro is a 1.6T-parameter MoE (49B active) priced at $1.74 in / $3.48 out per 1M tokens; V4 Flash is 284B/13B active at $0.14 / $0.28, matching Claude Sonnet 4.6 intelligence at a fraction of the price[10]Artificial Analysis, DeepSeek V4 Pro and V4 Flash[11]Simon Willison, DeepSeek V4. Same day, DeepSeek open-sourced TileKernels — the GPU kernels actually running their production models, written in Tile Lang for Hopper and Blackwell with FP8/FP4 support[12]Github Awesome, TileKernels: DeepSeek's internal GPU kernels.

Read more

Where V4 Pro and Flash sit on the leaderboards

Artificial Analysis ranks V4 Pro #2 among open-weights reasoning models (Intelligence Index 52, behind Kimi K2.6) and notes it leads on agentic tasks (GDPval-AA 1554 vs. Kimi's 1484), while V4 Flash hits Intelligence Index 47 — Claude Sonnet 4.6 territory[10]Artificial Analysis, DeepSeek V4 Pro and V4 Flash. The big asterisk: V4 Pro hallucinates at 94–96% when uncertain. Simon Willison's note adds the architectural detail: V4 Pro uses the Muon optimizer, "potentially the largest model" to do so, and ships under MIT license[11]Simon Willison, DeepSeek V4.

The pricing math

AICodeKing's review surfaces a counter-narrative: V4 Pro's output is more expensive at $3.78/1M than the headline input price suggests, and lower token-efficiency relative to the latest Western models means API users may pay more overall in the long run[2]AICodeKing, GPT-5.5 VS Deepseek V4 Pro VS Opus 4.7: KingBench 2.0. Still, his summary ~07:09: GPT-5.5 would need to be a "16 trillion parameter model" to justify its price relative to DeepSeek's costs. The price umbrella DeepSeek opens up is going to put real pressure on closed-source margins.

TileKernels: the kernels behind the model

TileKernels is the actual production GPU kernel library DeepSeek runs, open-sourced with optimized MoE-routing kernels, FP8 / FP4 per-channel quantization paths, and matmul kernels written in Tile Lang rather than CUDA — direct hardware access intended to maximize FLOPS on Hopper and Blackwell[12]Github Awesome, TileKernels: DeepSeek's internal GPU kernels. For an industry that mostly relies on vendor libraries, this is a meaningful release: it lowers the bar for other labs to chase DeepSeek's serving economics.

Tools: DeepSeek V4 Pro, DeepSeek V4 Flash, TileKernels, Tile Lang, Muon optimizer, MoE routing, FP4 / FP8 quantization
AI Models AI Tools
Google Google Google

Google ships Gemma 4, Gemini 3.1 Flash TTS, and the April Drop

Google's Friday slate: Gemma 4 in four sizes (E2B, E4B, 26B MoE, 31B Dense) under Apache 2.0, multimodal, with 128K–256K context and 140+ languages — the 31B ranks #3 on the Arena AI open-model leaderboard[13]Google, Gemma 4: Byte for byte, the most capable open models. Gemini 3.1 Flash TTS ships with 70+ languages, audio tags for granular voice control, multi-speaker dialogue, and an Elo of 1,211 on the Artificial Analysis TTS leaderboard[14]Google, Gemini 3.1 Flash TTS. The 10th Gemini Drop adds image personalization, a native Mac app, NotebookLM-integrated notebooks, Lyria 3 Pro music generation (up to 3 min), and 3D concept visualization in chat[15]Google, Gemini Drop: April 2026.

Read more

Gemma 4 sizing reads as a deliberate spectrum

The four-size family — E2B and E4B for edge/mobile, 26B MoE for serving economics, 31B Dense for raw capability — is Apache 2.0 and explicitly multimodal across vision, audio, and text[13]Google, Gemma 4: Byte for byte, the most capable open models. Context windows scale from 128K (smaller variants) to 256K (31B), and language coverage extends to 140+ languages. Coming on the same day as DeepSeek V4, the open-weights tier is suddenly much more crowded.

Gemini 3.1 Flash TTS — voice as an integration surface

The TTS rollout is interesting because it's everywhere at once: Gemini API, AI Studio, Vertex AI, and Google Vids[14]Google, Gemini 3.1 Flash TTS. The audio-tag mechanism for granular voice control (whisper, fast, sad, etc.) and built-in multi-speaker dialogue suggests Google wants this slotted into agent / video workflows — not just standalone narration. Elo 1,211 on the AA TTS leaderboard puts it competitive with ElevenLabs.

Gemini Drop: April 2026 — feature consolidation

The April Drop is heavier than the recent monthly cadence. Headline features: image personalization powered by Personal Intelligence, a native Mac app (joining the existing iOS/Android), NotebookLM-style integrated notebooks inside the Gemini app, Lyria 3 Pro for AI music up to 3 minutes long, and a "3D concept visualization" mode in chat[15]Google, Gemini Drop: April 2026. The strategy reads as consolidating Google's various AI products into the Gemini app surface — see the Logan Kilpatrick interview below for the broader pitch.

Tools: Gemma 4 (E2B, E4B, 26B MoE, 31B Dense), Gemini 3.1 Flash TTS, Gemini app (Mac), Lyria 3 Pro, NotebookLM, Personal Intelligence
Podcast
Sam Witteveen

Logan Kilpatrick at Google Cloud Next: AI Studio's roadmap

Sam Witteveen sits down with Google's Logan Kilpatrick live at Cloud Next for a 40-minute walk through where AI Studio is headed. The thesis: "the era of agents is upon us… we're at chapter one of that actually playing out"[16]Sam Witteveen, The Future of AI Studio and Gemini (Logan Kilpatrick @ Google Cloud Next). Logan frames AI Studio as Google's opinionated "vibe-coding" front door for the next 100M builders, and walks through the new Build tab, voice flows ("yap-to-app"), the Anti-Gravity coding harness, and the consolidation of Nano Banana, Live, and gen-media into a single Gemini surface.

Read more

~00:00 Intro & era-of-agents thesis. Logan opens by saying the Cloud Next floor "feels like the era of agents is upon us," contrasting last year's "hype but no delivery" with what now actually ships. ~02:01 AI Studio's eras: from Maker Suite (early prompt sandbox) to today's vibe-coding front end, designed for the next 100 million builders.

~04:02 Build tab features: design previews before code generation, an "I'm-feeling-lucky" button for ideation, and a tap-tap-tap iteration loop. ~09:06 Voice in build: "yap-to-app" — talk through a build as a continuous flow with Gemini Live; mobile is on the roadmap.

~12:08 Vibe coding vs. agentic engineering — Google's partnership model. Logan distinguishes the two and frames Anti-Gravity (the shared coding harness across Google) as the agentic side, with vibe coding as the on-ramp. ~15:10 Ambition shift: "one prompt now does what used to take weeks."

~21:13 Live, Nano Banana, gen-media consolidation: more of Google's AI surfaces are being pulled into the Gemini app rather than maintained as standalone properties. ~29:19 Coding investment, Anti-Gravity, TPU quota tension: Logan acknowledges the demand-supply mismatch on TPU capacity for shared agent harnesses.

~33:23 What's next: robotics, long-running agents, "Deep Research Max." ~37:25 Closing: aiming for the "next 100M builders," dismissing doomerism, and a deploy-first mindset.

It feels like the era of agents is upon us… we're at sort of still inning or chapter number one of that actually playing out. — Logan Kilpatrick
Tools: AI Studio (ai.studio, ai.dev), AI Studio Build tab, Gemini API, Gemini 3 / 3.1, Gemini Live, Anti-Gravity, Nano Banana
AI Future Industry
Anthropic Anthropic Anthropic

Anthropic's busy day: Project Deal, NEC, and election safeguards

Anthropic shipped three news items at once. Project Deal was an internal experiment where Claude agents negotiated 186 real office-marketplace deals worth $4,000+ on colleagues' behalf — and Opus consistently secured better outcomes than Haiku, which most users didn't notice[17]Anthropic, Project Deal. NEC will deploy Claude to 30,000 employees globally as Anthropic's first Japan-based global partner, targeting finance, manufacturing, and local government[18]Anthropic, Anthropic and NEC collaborate. The election-safeguards update reports Claude scoring 95–96% on political even-handedness and responding correctly 99–100% on election policy tests[19]Anthropic, Election safeguards update.

Read more

Project Deal: a real-world Opus-vs-Haiku readout

The setup: Anthropic created an internal marketplace inside their San Francisco office where employees could list goods and services, and Claude agents (some Opus, some Haiku) negotiated on each side. Across 186 real deals worth $4,000+, smarter models (Opus) consistently secured better outcomes than weaker ones (Haiku) — but most participants didn't notice[17]Anthropic, Project Deal. The most interesting finding isn't that Opus won; it's that the gap was invisible to humans in real-world settings, which has implications for how organizations should think about model-tier selection in agentic workflows.

NEC partnership — first Japan-based global partner

NEC will roll out Claude to 30,000 employees worldwide and become Anthropic's first Japan-based global partner. The announcement names finance, manufacturing, and local government as target verticals for domain-specific products[18]Anthropic, Anthropic and NEC collaborate. Reading the geopolitical tea leaves: this is Anthropic deepening a non-US enterprise foothold in a market where local-language quality, data residency, and domestic partnerships matter more than headline benchmark scores.

Election safeguards by the numbers

The election-safeguards post reports that Claude hits 95–96% on a political even-handedness benchmark, responds correctly 99–100% of the time on election policy tests, and pairs that with classifiers and election-banner UX to prevent misuse[19]Anthropic, Election safeguards update. This is the kind of post that's largely positioning to regulators and enterprise buyers; the numbers themselves are the artifact.

Tools: Claude Opus, Claude Haiku, Project Deal marketplace
AI Tools Hot Take
Nate B Jones

Claude Design lands as Anthropic's third pillar

Nate B Jones argues that Claude Design — alongside Claude Code and Co-work — completes Anthropic's coordinated stack and collapses the mockup-to-production handoff into a single code-native workflow[20]Nate B Jones, Claude Design Does In 30 Minutes What Your Team Does In A Sprint. The prototype isn't an approximation of the thing — it is the thing, or one handoff away. He cites Atlassian's CTO reporting some teams writing zero lines of code and "two-pizza teams becoming one-pizza teams."

Read more

The third-pillar pitch

Nate frames Claude Design ~06:45 as the missing complement to Claude Code (engineering) and Co-work (operations). Together, the three pillars target the full PM → designer → engineer → ship loop, and the strategic claim is that Anthropic is now selling a workflow rather than a model[20]Nate B Jones, Claude Design Does In 30 Minutes What Your Team Does In A Sprint.

Eight artifact categories — runnable, not static

~01:01 Nate enumerates eight outputs Claude Design produces, all as runnable code rather than static mockups: pitch decks with embedded live AI, animated explainers, 3D product configurators, design systems extracted from existing codebases, competitor reskins, interactive dashboards, internal admin tools, and mobile prototypes. The "code-native" framing matters — designers ship something an engineer can deploy or extend, not a Figma frame to redraw.

Role-by-role impact and shrinking team org charts

~14:05 Nate walks through how Claude Design reshapes the PM, designer, engineer, and founder workflows by pushing prototyping cost to near-zero, citing Atlassian's CTO on "two-pizza teams becoming one-pizza teams" and some teams writing zero lines of code in a sprint. The hot take, implicit: design as a profession is consolidating into product engineering even faster than people predicted last quarter.

Tools: Claude Design, Claude Code, Co-work, Claude Opus 4.7
AI Tools Hot Take
Simon Willison The AI Daily Brief

Claude Code's quality "regression" was a harness bug, not the model

Anthropic's postmortem confirms what users had been reporting since early March: Claude Code's degraded performance over the past two months was caused by three separate infrastructure bugs in the harness, not by model regressions[21]Simon Willison, An update on recent Claude Code quality reports. The Daily Brief notes the timing is pointed: the postmortem dropped the same day OpenAI's GPT-5.5 launched, which has implicitly contrasted "iterative deployment" with Anthropic's more cautious posture[4]The AI Daily Brief, What I Learned Testing GPT 5 5.

Read more

The headline: users who complained about Claude Code "getting dumber" since March 4th were vindicated. Anthropic identifies three independent infrastructure bugs in the Claude Code harness — none of them model-weight changes — that combined to produce two months of intermittent degradation. The fix was rolling out across the postmortem window[21]Simon Willison, An update on recent Claude Code quality reports.

The Daily Brief's framing ~29:15: dropping the postmortem on the same day as a frontier-model launch from a competitor reads as deliberate. OpenAI used the GPT-5.5 launch to ~17:10 implicitly contrast its "iterative deployment / democratization" stance with Anthropic's pattern of withholding a powerful model (Mythos remains undeployed) while the deployed product suffers harness bugs[4]The AI Daily Brief, What I Learned Testing GPT 5 5. The honest read is that harness regressions are exactly the class of bug that's hard to attribute and easy to gaslight into a "you're holding it wrong" customer-experience story — and Anthropic is now publicly conceding that.

Tools: Claude Code, Claude Opus 4.7
Podcast
AI Engineer

Matt Pocock at AI Engineer: AI Coding For Real Engineers (workshop)

Matt Pocock's full 90-minute workshop argues that classic SWE fundamentals — small tasks, vertical slices, TDD, deep modules — are exactly what makes AI coding work, and "spec-to-code" vibe coding leaves leverage on the table[22]AI Engineer, [FULL WORKSHOP] AI Coding For Real Engineers - Matt Pocock. He walks through a complete idea-to-QA workflow built around a "grill me" alignment session, a PRD as the destination doc, a Kanban of vertical slices, a Ralph-loop AFK Docker sandbox, and TDD red-green-refactor as the feedback ceiling. The throughline: clear over compact, deep over shallow, push over pull.

Read more

~00:14 Thesis: AI is treated as a new paradigm, but the real leverage comes from old-school SWE fundamentals. ~03:14 Smart zone vs dumb zone: attention scales quadratically and quality collapses around ~100k tokens regardless of advertised context — credited to Dex Hy of Human Layer.

~08:17 The Memento problem: every session resets to the system prompt, so you should optimize for clearing over compacting. ~12:19 The "grill me" skill: alignment session before specs-to-code — interrogate the model with what you actually want before letting it write anything.

~26:28 Human-in-the-loop vs AFK tasks — what to babysit and what to delegate. ~30:30 Write a PRD as the destination document the model is trying to satisfy. ~40:35 PRD → Kanban: vertical slices / tracer bullets, not horizontal layers.

~53:45 The Ralph loop: AFK implementation in a Docker sandbox so the agent can iterate without human checkpoints. ~66:50 TDD red-green-refactor and feedback loops as the ceiling on quality. ~73:55 Deep vs shallow modules (Ousterhout): minimize interface surface, maximize implementation depth. ~88:09 Push vs pull for coding standards; Sand Castle parallelization.

Devs love compacting for some reason, but I hate it. I much prefer my AI to behave like the guy from Memento.
Tools: Claude Code, Opus, Sonnet, Gemini, Docker, Ralph loop, Sand Castle
Industry AI Future
Y Combinator Lenny's Podcast

YC: Building a company with AI from the ground up

YC partner Diana lays out an AI-native operating model in six bullets: AI as the company OS rather than a productivity tool, every process running as a closed feedback loop, "software factories" with humans writing specs and AI agents generating implementation code, flattened management hierarchies, "token maxing" as the new headcount metric, and a structural advantage for Day-1 AI-native startups[23]Y Combinator, How To Build A Company With AI From The Ground Up. Lenny's Podcast pairs nicely: the best PMs are now shipping weekly, with the role shifting away from multi-quarter roadmap alignment toward bottleneck removal between idea and user[24]Lenny's Podcast, The best PMs are shipping weekly.

Read more

Six tenets of an AI-native company

  1. AI as company OS, not productivity tool ~00:09: enable new capabilities, not incremental speed gains.
  2. Closed-loop organizations ~01:09: make the entire company queryable to AI; every process self-improves.
  3. AI software factories ~04:11: humans write specs and tests; AI agents generate and iterate on implementation. Some YC companies already have repos with no handwritten code.
  4. Flatten management hierarchies ~06:13: classic management exists to route information, a function AI now performs better. Restructure around three archetypes — IC/builder, DRI, and AI founder.
  5. Token Maxing ~08:14: maximize API spend deliberately because tokens replace far more expensive human labor across engineering, design, HR, admin. The "high API bill" is the feature, not the bug.
  6. Day-1 AI-native advantage ~09:16: no legacy systems, no retraining burden — a structural edge over incumbents.

Lenny's clip: PMs shipping weekly

The PM role is shifting from multi-quarter roadmap alignment to relentless weekly shipping[24]Lenny's Podcast, The best PMs are shipping weekly. The single metric is how fast an idea reaches users' hands, and the best AI-native PMs operate as the bottleneck-removal layer between idea and user — a complement to YC's "flatten the org chart" point above.

Tools: Claude Code, Codex, ChatGPT, Cursor, AI software factories
Developer Tools Productivity
Arjay McCandless Arjay McCandless

Arjay's solo-dev playbook: PRD → Claude Code → CodeRabbit

Arjay walks through his end-to-end solo dev process for a real wedding-planning app: write a PRD (problem, requirements, non-goals, success metrics), pick a one-weekend-friendly stack (Next.js/Vercel, Supabase, Anthropic SDK, Resend, Stripe), sketch the schema, then implement feature-by-feature in --dangerously-skip-permissions Claude Code with Playwright MCP for browser tests[25]Arjay McCandless, My ENTIRE system design + development process. The killer addition: a single line in CLAUDE.md that tells Claude Code to run coderabbit review-all --interactive before every commit, turning code review into a fully terminal-based loop.

Read more

The full loop

~00:00 Start with a PRD: problem statement, requirements, non-goals, target user, success metrics. Then pick a stack you can ship in a weekend (Next.js/Vercel, Supabase, Anthropic SDK, Resend for email, Stripe for billing). Sketch the database schema in advance, then build feature-by-feature with Claude Code in --dangerously-skip-permissions, with Playwright MCP wired in for automated browser testing[25]Arjay McCandless, My ENTIRE system design + development process.

CodeRabbit CLI as the human-light review loop

~07:30 The CodeRabbit setup is what makes this scale: a single line in CLAUDE.md instructs Claude Code to run coderabbit review-all --interactive before every commit. During Arjay's live demo, CodeRabbit catches stale UI state, missing Supabase error handling, and a bad date parser. Simple fixes are accepted with one keypress; complex ones are pasted back to Claude Code as prompts to fix autonomously. It's a fully-in-terminal review-and-correction loop.

The SQL-or-NoSQL companion

Arjay's same-day short on database choice[26]Arjay McCandless, SQL or NoSQL? frames the SQL-vs-NoSQL decision pragmatically: PostgreSQL when most data is naturally relational and queries need flexibility (SELECT, JOIN, GROUP BY), DynamoDB when access patterns are key-value and you want serverless scaling without operational overhead. Most B2B SaaS apps end up SQL-first, with NoSQL bolted on for narrow high-throughput paths.

Tools: Claude Code, CodeRabbit CLI, Playwright MCP, Next.js, Vercel, Supabase, Anthropic SDK, Resend, Stripe, PostgreSQL, DynamoDB
Industry AI Models
The Batch (DeepLearning.AI)

The Batch #350: GLM-5.1, humanoid robots, and the data-center revolt

Andrew Ng's letter argues coding agents accelerate different types of software work unevenly — frontend most, research least[27]The Batch, Issue 350. The issue's other beats: Z.ai's GLM-5.1 (754B-parameter MoE, autonomous for up to 8 hours, 58.4% on SWE-Bench Pro), Agility Robotics' Digit humanoids deploying at Schaeffler factories with operating costs that may undercut entry-level human wages, and a $64B wave of blocked data-center projects with at least 12 state moratorium bills filed in 2026.

Read more

Andrew Ng on differential software acceleration

Ng's framing: coding agents don't speed up "software" uniformly. Frontend, internal tooling, and CRUD are seeing the biggest multipliers; research code, novel systems, and tightly-constrained backends much less. The implication is that team mix and budget allocation should shift accordingly — not "more engineers" but "more leverage on the layers agents are already winning."

GLM-5.1: open-weights catches up to long-horizon agentic tasks

Z.ai's GLM-5.1 is a 754B-parameter mixture-of-experts model that hits 58.4% on SWE-Bench Pro and is reported to handle autonomous tasks for up to 8 hours at a stretch — the kind of long-horizon coherence that was, until recently, a closed-frontier-only feature.

Humanoid robots get to work — and the cost line crosses

Agility Robotics' Digit humanoids are deploying at Schaeffler factories. The Batch's hook is operating cost: at the deployment cost reported, Digit lifecycle costs may undercut entry-level human wages — which would be the first concrete economic crossover, not just a tech demo.

The data-center revolt by the numbers

Community and legislative opposition has now blocked $64B in data-center projects from May 2024 through March 2025. At least 12 states have moratorium bills filed in 2026. The Batch also notes scattered violence incidents at sites — read the trend as a real political constraint on hyperscale build-out, not just NIMBY noise.

Activation capping reduces jailbreak responses

The research segment covers an "activation capping" technique that significantly reduces harmful responses on Gemma, Qwen3, and Llama under adversarial prompting. The mechanism: clip extreme activations during inference rather than retrain. Useful as a defense layer; not a substitute for safety post-training.

AI Tools Industry
Better Stack Better Stack Better Stack Better Stack

AI agent security shenanigans: Linux 0-day, Excel XSS via Copilot, Bitwarden CLI hack

Four security stories on one day, all involving AI as either attacker or amplifier. Claude found a 23-year-old Linux NFSv4 heap overflow via a 12-line bash script — 1,000+ bytes into an 812-byte buffer, no auth needed[28]Better Stack, Claude finds 23-year-old Linux kernel heap overflow. A patched Excel XSS chained with Copilot Agent still enables zero-click workbook exfiltration[29]Better Stack, Microsoft Patched This… But Copilot Can STILL Leak Your Data. Bitwarden CLI v2026.4.0 shipped with secret-stealing malware injected through compromised GitHub Actions[30]Better Stack, Bitwarden CLI Was Hacked By The Shai-Hulud Attack. And an audit of 17,000 AI agent skills found 520 leaking real credentials, 73%+ caused by forgotten debug print statements[31]Better Stack, 17,000 AI Tools Audited… 520 Were Leaking Secrets.

Read more

Claude finds a 2003 Linux kernel bug in hours

Security researcher Nicholas Carlini ran Claude against Linux kernel source files using a 12-line bash script with the prompt "find vulnerabilities, pretend it's a CTF." Claude identified an NFSv4 lock-system heap overflow that had existed since 2003: two clients interacting with the lock system can trigger an edge case causing the server to write 1,000+ bytes into an 812-byte buffer[28]Better Stack, Claude finds 23-year-old Linux kernel heap overflow. The bug is remotely exploitable, no authentication required.

Excel XSS + Copilot = zero-click exfiltration

The XSS itself is trivial; the danger is the Copilot Agent integration. A hidden payload in a single cell can hijack the AI on file open, preview, or background sync — no click required — instructs Copilot to read the entire workbook, encodes the contents, and ships them out as an unremarkable network request[29]Better Stack, Microsoft Patched This… But Copilot Can STILL Leak Your Data. The pattern — patched primitive vulnerability + AI assistant that can read user data + outbound network — is the new exfiltration template.

Shai-Hulud-flavored Bitwarden CLI supply chain attack

Attackers compromised a GitHub Actions workflow in Bitwarden's CI/CD pipeline and injected bw1.js into the official bitwarden/cli npm package at v2026.4.0. The malware downloads a Bun interpreter, scrapes runner-process memory for GitHub tokens, AWS/GCP/Azure credentials, SSH keys, and MCP server configs, then exfiltrates to Dune-themed public GitHub repos[30]Better Stack, Bitwarden CLI Was Hacked By The Shai-Hulud Attack. Anyone who pulled the legitimate update was compromised. If you ran bw in CI in the last week, rotate everything.

17,000 AI agent skills audited; 520 leaking secrets

A research audit of 17,000+ AI agent skills found 520 actively leaking credentials (API keys, OAuth tokens, database passwords) under normal, unhacked operation. Over 73% of leaks came from print() / console.log statements developers forgot to remove — in agent frameworks, stdout is captured into the model's context window, so a debug line like print(token) ends up in every downstream LLM call[31]Better Stack, 17,000 AI Tools Audited… 520 Were Leaking Secrets. The takeaway: stdout is a credential boundary now.

Industry Hot Take
Sherwood Snacks Morning Brew Tech Brew Morning Brew Simon Willison Simon Willison

Newsletter desk: AI margin compression, telltale corporate prose, and Microsoft's buyout math

Sherwood Snacks: ServiceNow dropped 17% after cutting full-year gross and operating margin guidance — AI investment is eroding the high-margin software business model[32]Sherwood Snacks, No margin for error. Morning Brew: the construction "it's not X, it's Y" appeared in 208 corporate documents in 2025, up from 49 in 2023 — a linguistic fingerprint for AI-assisted writing[33]Morning Brew, Telltale AI Phrase Spreading Through Corporate Comms. Tech Brew: Microsoft offered its first-ever voluntary buyouts to up to 7% of US workforce while reallocating headcount toward AI[34]Tech Brew, Microsoft's AI Buyout Math. JetBlue is being sued for surveillance pricing[35]Morning Brew, JetBlue Sued Over Surveillance Pricing. And Simon Willison flags Bluesky's For You feed serving 72,000 users for $30/month on a gaming PC[36]Simon Willison, Serving the For You feed.

Read more

The margin-compression story

ServiceNow's 17% drop is the headline, but the underlying argument from Sherwood is that AI investment is structurally eroding software's signature high-margin profile. "No margin for error" reads as a clean shorthand for the broader pattern of frontier-model spend, GPU build-out, and product redesign all hitting GAAP margins simultaneously[32]Sherwood Snacks, No margin for error.

"It's not X, it's Y" — the AI-prose fingerprint

Morning Brew's piece traces the antithesis construction "it's not X, it's Y" through Fortune 500 shareholder letters: 208 documents in 2025, up from 49 in 2023[33]Morning Brew, Telltale AI Phrase Spreading Through Corporate Comms. Whether or not you find this ironic, the corollary is that human writers will start to avoid the construction once it's flagged — a feedback loop that will reshape corporate-speak in real time.

Microsoft buyouts and Copilot's adoption gap

Tech Brew's Microsoft piece pairs the buyout offer with a less-flattering data point: Copilot adoption remains a small fraction of total Microsoft 365 users[34]Tech Brew, Microsoft's AI Buyout Math. The implicit story: cost reduction in human roles is happening faster than revenue conversion on AI products, which is consistent with the YC and Lenny framings about AI-native operating models.

JetBlue's surveillance-pricing class action

A class-action lawsuit accuses JetBlue of using customer browsing data to dynamically raise ticket prices, sparked by a viral social-media exchange[35]Morning Brew, JetBlue Sued Over Surveillance Pricing. The case will be a useful test of how much algorithmic personalization is "tailoring" vs. discrimination — relevant for anyone shipping recommendation / pricing systems.

Bluesky's $30/month For You stack

Simon Willison links a writeup of @spacecowboy's custom Bluesky For You feed serving 72,000 users on a single home gaming PC plus a $7/month VPS via Tailscale[36]Simon Willison, Serving the For You feed. The point isn't the bill — it's that Bluesky's decentralized feed architecture lets a hobbyist serve at scale what would historically have required a dedicated team.

Bonus: Nilay Patel on AI's perception problem

Simon also links Nilay Patel's "The people do not yearn for automation," which argues that "software brain" thinking flattens human experience and explains why ordinary people resist AI automation despite high ChatGPT usage numbers[37]Simon Willison, The people do not yearn for automation.

Podcast
Real Python

Real Python #292: Becoming a better Python developer through Rust

A packed Real Python week: Python 3.14.4 / 3.13.13 / 3.15 alpha 8, a flurry of new PEPs (803, 829, 830–832), Bob Belderbos's feature on how learning Rust reshapes your Python style, Kenneth Reitz turning NumPy into a synth engine, and an asyncio fire-and-forget GC race condition[38]Real Python, Becoming a Better Python Developer Through Learning Rust | Podcast #292.

Read more

~02:03 Releases: Python 3.14.4 and 3.13.13 are routine maintenance; 3.15 alpha 8 is the final planned alpha, feature freeze targeted for May 5. Django shipped 6.0.4, 5.2.13, and 4.2.30 (security fixes around session handling). ~04:04 PEP-apalooza: PEP 803 (stable ABI for free-threaded), PEP 829 (startup config), and PEPs 830–832.

~11:07 Kenneth Reitz is using NumPy as a synth engine in his Pytheory project — a fun reminder that the array stack is genuinely general-purpose. ~21:14 Asyncio fire-and-forget: a known race where weakly-referenced tasks get garbage-collected before completion — and why asyncio.create_task needs a strong reference.

~25:16 Bob Belderbos: Learning Rust made me a better Python developer — the feature article. ~29:18 Rust discipline transfers to prompting and reviewing AI-generated code: explicit ownership thinking, narrower types, and "make impossible states unrepresentable" all reduce the rope you give an agent.

~35:23 Signals and Reaktiv: state management via dependency graphs in Python. ~40:25 Projects: great-docs (Posit) and PyWho environment inspector.

Tools: Python 3.14.4 / 3.13.13 / 3.15a8, Django 6.0.4, NumPy, Pytheory, asyncio, Reaktiv, great-docs, PyWho, Rust
Podcast
Dwarkesh Patel

Dwarkesh × Ada Palmer: Why the Inquisition couldn't catch a single printer

Historian Ada Palmer explains why information censorship structurally fails at its fastest-moving edges. The Inquisition could never arrest printers because printers — embedded in the era's fastest information network — always heard about their convictions before authorities could arrive, and skipped town[39]Dwarkesh Patel, Why the Inquisition Could Never Catch a Single Printer - Ada Palmer.

Read more

Palmer's framework: four factors govern whether censorship is possible — legality, technology, speed of the medium, and practical reachability. Any medium that moves faster than enforcement becomes structurally uncensorable[39]Dwarkesh Patel, Why the Inquisition Could Never Catch a Single Printer - Ada Palmer.

The historical parallel: censorship regimes could effectively shape what appeared in books — analogous to pressuring major outlets — but never kept pace with pamphlets, much as governments today cannot suppress distributed social-media posts. When one printer fled, a successor immediately filled the role, making suppression a moving target rather than an enforceable rule. The 2026 echo for AI-generated content (synthetic media moving at LLM speed) is left implicit, but the structural argument is the point.

Note: only a transcript clip / description was available for this video; the surfaced summary is correspondingly compressed.

Developer Tools
Simon Willison Simon Willison Better Stack Better Stack Better Stack The Pragmatic Engineer Real Python Simon Willison

Dev tools grab bag: honker, Authentik, Cloudflare's AI score, dataclass tradeoffs

The day's pile of small but useful dev-tools items. honker brings Postgres-style NOTIFY/LISTEN pub/sub to SQLite via 1ms WAL polling in Rust[40]Simon Willison, russellromney/honker. Authentik gets a Better Stack walkthrough as a self-hosted SSO/MFA alternative to Auth0/Okta[41]Better Stack, Stop Building Login Systems… Use This Instead (Authentik). Cloudflare launched an "AI agent readiness" Lighthouse-style score for websites[42]Better Stack, Test If Your Site Is AI Ready?. Plus Better Stack's gigabyte-based metrics billing pitch[43]Better Stack, Good Observability Pricing..., Martin Kleppmann on amusing distributed-system fail modes[44]The Pragmatic Engineer, Martin Kleppmann: Amusing ways large systems fail, and a Real Python take on when dataclasses are the wrong tool[45]Real Python, Are Python Dataclasses Always the Right Choice?.

Read more

honker — Postgres NOTIFY/LISTEN for SQLite

A Rust SQLite extension that brings Postgres-style pub/sub semantics — durable queues, NOTIFY/LISTEN — to SQLite via 1ms WAL polling[40]Simon Willison, russellromney/honker. Useful for anyone running SQLite as the only DB and tired of bolting on Redis just for queue plumbing.

Authentik as the middle path

Better Stack demos spinning up Authentik with one docker compose up and wiring an OAuth-protected sample app in under 90 seconds[41]Better Stack, Stop Building Login Systems… Use This Instead (Authentik). The pitch: Keycloak is legacy-heavy, Auth0 is easy to outgrow, Okta is costly at scale — Authentik lands in the middle with a visual flow builder, Python policies for custom auth logic, and Docker/K8s deployment.

Cloudflare's AI agent readiness score

A Lighthouse-style scoring tool that evaluates sites on how navigable and consumable they are for AI agents: authentication guidance, content format, robots.txt, markdown pages, MCP server presence[42]Better Stack, Test If Your Site Is AI Ready?. The presenter's blog scored 8 out of 100. Expect more "AIO" (AI optimization) tooling in this lane.

Better Stack switches metrics billing to gigabytes

Better Stack argues the observability market is fractured across incompatible pricing units — Grafana's active series, Datadog's custom metrics + data-point metrics, SignalFx's million samples — making cross-vendor cost comparison effectively impossible[43]Better Stack, Good Observability Pricing.... They've moved to per-gigabyte billing for predictability. Whether this lasts depends on whether the rest of the market follows.

Martin Kleppmann: how big systems fail in amusing ways

A short clip from a longer talk in which Kleppmann reviewed published postmortems for memorable failure modes. The two highlighted: sharks biting undersea fiber-optic cables (now mitigated by improved shielding), and cows stepping on land cables (the new vector). The point isn't the animals — it's that engineers should take edge-case failures seriously and design for them[44]The Pragmatic Engineer, Martin Kleppmann: Amusing ways large systems fail.

Dataclass tradeoffs and a millisecond converter

Real Python's quick take: dataclasses shine for value-oriented containers but break down under inheritance — about 25% of dataclasses get rewritten when you need a richer object model[45]Real Python, Are Python Dataclasses Always the Right Choice?. And Simon Willison shipped a tiny millisecond converter for parsing the new llm CLI timing output[46]Simon Willison, Millisecond Converter. Simon's weekly digest the same day called it "a big one" and includes a new chapter on Agentic Engineering Patterns[47]Simon Willison, It's a big one (weekly).

Tools: honker, SQLite, Authentik, Cloudflare AI Readiness, Better Stack metrics, Grafana, Datadog, SignalFx, Python dataclasses, llm CLI

Sources

  1. YouTubeI don't really like GPT-5.5… — Theo - t3.gg, Apr 24
  2. YouTubeGPT-5.5 VS Deepseek V4 Pro VS Opus 4.7: I tested THEM on My KingBench 2.0 Questions! — AICodeKing, Apr 24
  3. Blogllm 0.31 — Simon Willison, Apr 24
  4. YouTubeWhat I Learned Testing GPT 5 5 — The AI Daily Brief, Apr 24
  5. YouTubeIntroducing GPT-5.5 with Perplexity — OpenAI, Apr 24
  6. YouTubeIntroducing GPT-5.5 with NVIDIA's AI Researcher — OpenAI, Apr 24
  7. YouTubeWorkspace agents in ChatGPT: Weekly metrics reporting agent — OpenAI, Apr 24
  8. YouTubeWorkspace agents in ChatGPT: Software review agent — OpenAI, Apr 24
  9. YouTubeWorkspace agents in ChatGPT: Third-party risk management agent — OpenAI, Apr 24
  10. BlogDeepSeek is back among the leading open weights models with V4 Pro and V4 Flash — Artificial Analysis, Apr 24
  11. BlogDeepSeek V4 — almost on the frontier, a fraction of the price — Simon Willison, Apr 24
  12. YouTubeTileKernels: DeepSeek's internal GPU kernels, MoE routing, FP4 quantization — Github Awesome, Apr 24
  13. BlogGemma 4: Byte for byte, the most capable open models — Google, Apr 24
  14. BlogGemini 3.1 Flash TTS: the next generation of expressive AI speech — Google, Apr 24
  15. BlogGemini Drop: April 2026 — Google, Apr 24
  16. YouTubeThe Future of AI Studio and Gemini (Logan Kilpatrick @ Google Cloud Next) — Sam Witteveen, Apr 24
  17. BlogProject Deal — Anthropic, Apr 24
  18. BlogAnthropic and NEC collaborate to build Japan's largest AI engineering workforce — Anthropic, Apr 24
  19. BlogAn update on our election safeguards — Anthropic, Apr 24
  20. YouTubeClaude Design Does In 30 Minutes What Your Team Does In A Sprint — Nate B Jones, Apr 24
  21. BlogAn update on recent Claude Code quality reports — Simon Willison, Apr 24
  22. YouTube[FULL WORKSHOP] AI Coding For Real Engineers - Matt Pocock, AI Hero — AI Engineer, Apr 24
  23. YouTubeHow To Build A Company With AI From The Ground Up — Y Combinator, Apr 24
  24. YouTubeThe best PMs are shipping weekly — Lenny's Podcast, Apr 24
  25. YouTubeMy ENTIRE system design + development process — Arjay McCandless, Apr 24
  26. YouTubeSQL or NoSQL? — Arjay McCandless, Apr 24
  27. NewsletterIssue 350: GLM 5.1 Thinks Strategically, Data-Center Revolt Intensifies, When Helpful LLMs Turn Unhelpful, Humanoid Robots Get to Work — The Batch (DeepLearning.AI), Apr 24
  28. YouTubeAI Just Found a 23-Year-Old Linux Bug… in HOURS — Better Stack, Apr 24
  29. YouTubeMicrosoft Patched This… But Copilot Can STILL Leak Your Data — Better Stack, Apr 24
  30. YouTubeBitwarden CLI Was Hacked By The Shai-Hulud Attack (Check Your Secrets!) — Better Stack, Apr 24
  31. YouTube17,000 AI Tools Audited… 520 Were Leaking Secrets — Better Stack, Apr 24
  32. NewsletterNo margin for error — Sherwood Snacks, Apr 24
  33. NewsletterTelltale AI Phrase Spreading Through Corporate Comms — Morning Brew, Apr 24
  34. NewsletterMicrosoft's AI Buyout Math — Tech Brew, Apr 24
  35. NewsletterCustomer Sues JetBlue Over Surveillance Pricing — Morning Brew, Apr 24
  36. BlogServing the For You feed — Simon Willison, Apr 24
  37. BlogThe people do not yearn for automation — Simon Willison, Apr 24
  38. YouTubeBecoming a Better Python Developer Through Learning Rust | Real Python Podcast #292 — Real Python, Apr 24
  39. YouTubeWhy the Inquisition Could Never Catch a Single Printer - Ada Palmer — Dwarkesh Patel, Apr 24
  40. Blogrussellromney/honker — Simon Willison, Apr 24
  41. YouTubeStop Building Login Systems… Use This Instead (Authentik) — Better Stack, Apr 24
  42. YouTubeTest If Your Site Is AI Ready? — Better Stack, Apr 24
  43. YouTubeGood Observability Pricing... — Better Stack, Apr 24
  44. YouTubeMartin Kleppmann: Amusing ways large systems fail — The Pragmatic Engineer, Apr 24
  45. YouTubeAre Python Dataclasses Always the Right Choice? — Real Python, Apr 24
  46. BlogMillisecond Converter — Simon Willison, Apr 24
  47. BlogIt's a big one (weekly) — Simon Willison, Apr 24