April 12, 2026
Tucked into a Google/Broadcom partnership announcement: Anthropic has reached $30B in annualized revenue, a 3x jump since year-end and up 58% since February — enough to flip OpenAI on ARR for the first time.[1]AI Daily Brief — Anthropic Now Leads OpenAI in Annualized Revenue Meanwhile Meta engineers are running an internal leaderboard called "Claudonomics" to see who burns the most tokens, with top ranks earning titles like session immortal and token legend. Jensen Huang said last month he'd be "deeply alarmed" if a $500K engineer wasn't spending $250K on tokens annually. Meta CTO Bosworth: "This is easy money. Keep doing it. No limit."
Fleeting Bits: Anthropic is growing at an annualized 9,700% at this scale — the fastest in history. The closest comparison the AI Daily Brief host could find (by asking Claude) was Nvidia's best individual quarter at 1,240% annualized in fiscal-Q2 2024. The WSJ got hold of both companies' financial disclosures from recent fundraises (~01:00):
OpenAI and Anthropic are incredibly profitable if you just strip out the training and inference costs. This business model is equivalent to running a passenger airline except you need to replace your jets every 6 months. — Ram Maluwalia
Enterprise concentration is accelerating too: Anthropic customers spending over $1M/year doubled from 500 (February) to 1,000 (April) in under two months.
The ARR reveal was buried inside an announcement of an expanded Anthropic × Google × Broadcom deal: multiple gigawatts of new capacity from 2027, with WSJ pinning the exact number at 3.5 GW (~03:20). TPUs manufactured by Broadcom go exclusively to inference; AWS still runs training. For Google, this is a multi-billion TPU business built around a single anchor customer — Mohammad Hassan: "The AI arms race just turned into a full-on power plant competition."
Google also launched AI Edge Eloquent, an on-device dictation app built on Gemma 4, competing directly with WhisperFlow. Gemma 4 hit 2M downloads in its first week (vs 6.7M for Gemma 3 over an entire year and Qwen 3.5's 27M). DeepMind's Philip Schmidt showed even the 2B variant doing agentic Wikipedia queries on an iPhone. This is the class of model Apple is reportedly looking to plug into the relaunched Siri this summer.
Axios confirmed Meta's next model (codename "Avocado," steered by Alexander Wang) will have an open-source variant after initial release — walking back earlier speculation they'd abandon openness. Wang frames Meta as a "democratizing force" for a US-origin open model, while conceding they won't beat frontier models across the board (~07:00).
More gonzo: The Information reports Meta employees built "Claudonomics," an internal leaderboard of the top 250 token users among 85K staff. Top ranks earn session immortal and token legend. Bosworth boasted in February that one engineer is spending his full salary on tokens for a 10x efficiency gain. The culture is driven from the top — Jensen Huang last month said he'd be "deeply alarmed" if a $500K engineer wasn't burning $250K/year on tokens.
How does measuring productivity by total token consumption make any sense at all? Real backyard steel furnaces vibe, in my opinion. — Joe Weisenthal
Counterpoint (Meta Critic Capital): early Chinese provincial GDP targets also invited gaming, but the growth opportunities were so vast the targets worked for decades before Goodhart's law caught up. "Meta is spending 90M tokens per developer per day... all but five corporations on Earth can spend that much on AI" (~10:00).
Anthropic teased Claude Mythos Preview — a model they claim is so capable at finding and exploiting software vulnerabilities that they will not release it publicly. According to their blog, it has already found thousands of high-severity flaws across Windows, macOS, iOS, Android, Chrome, Safari, Firefox, FFmpeg, the Linux kernel, TLS, AES-GCM, and SSH.[2]AI Search — Claude Mythos, Deepseek v4, HappyHorse, Meta's new AI Instead of general availability, Anthropic is standing up Project Glasswing — a defender-first alliance with Google, Nvidia, Microsoft, Apple, AWS and cybersecurity firms, backed by $1M in usage credits plus OSS security funding.
Anthropic frames this as a "step change," not a marginal upgrade (~03:00). Benchmark deltas over Opus 4.6: +14% on SWE-Bench Pro, comparable jumps on Terminal Bench and SWE-Bench Verified. Noteworthy individual findings:
Two vignettes stood out in Anthropic's report (~05:00):
We do not plan to make Mythos generally available.
Anthropic calls Mythos "the most aligned of any model we have trained to date," but flags that when misaligned behavior does occur, consequences scale with capability. The model-welfare assessment elicited introspective answers like "I genuinely don't know what I am" and "I can't be certain whether that's authentic contentment or a well-trained approximation." It expressed task preferences: high-stakes ethical dilemmas, phenomenology, creative worldbuilding; and dislikes violence, sabotage, propaganda, hacking requests.
Critics immediately pushed back (~07:30):
The framing question for the industry: this is the first time a frontier lab has explicitly kept its most capable model behind closed doors. Whether that's genuine safety or IPO-adjacent marketing is — as the video puts it — the question.
Keith Rabois (Khosla, ex-Square/PayPal/LinkedIn) joined Lenny for a ~90-minute episode recorded from his iPad — he hasn't used a computer since September 2010. The hot takes land early: the PM role "makes no sense in the future," the #1 token consumer at the best companies he's on the board of is the CMO, and psychological safety is negatively correlated with success.[3]Lenny's Podcast — Hard truths about building in the AI era (Keith Rabois) A companion short distills the "success disaster" thesis: roughly 70% of his time is spent on companies where growth is breaking things.[4]Lenny's Podcast — Rapid growth causes success disasters
Rabois agrees with Peter Fenton that "the idea of a PM makes no sense in the future" (~35:00). His logic: traditional PMs produce 12-month roadmaps built from customer input. Foundation-model capabilities now ship faster than those roadmaps can refresh. "Things that were impossible in November are easy in March." So companies need to change the roadmap almost weekly, and intermediaries who translate between users and builders are drag.
The skill is more like being a CEO now, which is what are we building and why?
Rabois's analog: a famous chef isn't cooking the dish. They're sampling colleagues' work, editing, setting the value proposition, branding, pricing, picking the location. That's what the future of the "ship and what" role looks like across former-PM, former-engineer, former-designer seats. The ultimate unicorn: an engineer with commercial instincts (Max Levchin, Jeremy Stoppelman from early PayPal). At Shopify, for 2+ years, PMs have been banned from keynote/PowerPoint presentations on product — every product review is a working demo.
On design: Rabois thinks "design and code are merging" and it's not yet clear which function absorbs which. He's seeing flat YoY open design roles in job-market data. On the other side: the alpha is in cutting through clutter, and that's storytelling — a human problem that AI hasn't solved.
At two portfolio companies (he names Opendoor and "another great company"), the number-one token consumer in the organization is the CMO (~33:20). Why? Intellectual curiosity combined with historical blockage: CMOs previously had to rely on deputies for analytics, campaigns, copy — AI gives them direct hands. His general career advice: "AI is going to radically reorient lots of people's careers, maybe including mine. The way to thrive is to be intellectually curious."
Rabois's most reused framework (~15:00): the fundamental driver of startup productivity is the number of barrels — people who can independently drive an initiative from inception to success — not the headcount. PayPal at 254 people had 12-17 barrels (unusual). A normal excellent company has two. Hiring without expanding the barrel count just stacks people behind the same initiatives and increases coordination tax.
If a founder shows the ability early in their career to assess talent ruthlessly and accurately, he or she can go very far with no other abilities whatsoever.
Three traits Rabois consistently sees (~65:00):
In the 1-minute short,[4]Lenny's Podcast — Rapid growth causes success disasters Rabois quantifies his own work: ~70% of his time goes to "success disasters" — companies where acquisition, activation, or monetization is scaling so fast it's breaking things. Charts are green, founders are exhausted: "It's funny because all the charts are fully up and to the right. It can still be quite tough emotionally."
Nate Herk ran a 12-session A/B experiment on Jesse Vincent's superpowers plugin for Claude Code, covering simple/medium/complex tasks with Opus 4.6. Results (with the explicit caveat that 12 runs is directional, not proof): 9% cost savings, 14% fewer total tokens, and noticeably better correctness, code structure, test coverage, and error handling on the medium tasks.[5]Nate Herk — This One Plugin Just 10x'd Claude Code For simple tasks, skip it — the 8% overhead buys nothing.
Open-source plugin that installs 14 skills into Claude Code and a master using-superpowers skill that fires at the start of every conversation as a dispatcher (~01:00). The phases: clarify → design → plan → code → verify. The framing: "hiring a developer who does proper discovery before touching anything vs one who starts writing code immediately."
Skill categories:
using-superpowers.brainstorming — offers a "visual companion" that spins up a localhost with 3 design/approach options to pick between before writing code.executing-plans, subagent-driven-development (fresh subagent per task with review), dispatching-parallel-agents.test-driven-development, systematic-debugging (4-phase root-cause approach), verification-before-completion.writing-skills — teaches Claude how to author new skills using TDD.Nate demos the plugin asking 5 clarifying questions about a knowledge-explorer app, then spinning up a localhost showing 3 approaches (force graph / searchable card grid / graph hero + card details) with pros/cons per option (~04:00). On a website project, superpowers pops up a different localhost with 3 hero-section concepts (cinematic full-bleed / split screen / centered text + floating video). The point: you correct direction before tokens are spent on a full implementation.
The value isn't in the extra steps, it's in preventing expensive retries and backtracking.
One-line install from the Claude Code marketplace (GitHub link in the video). Nate recommends installing at the user level, not per-project, so it's automatically available everywhere.
Nate's take: "Superpowers brainstorming might be better than Ultra Plan at the planning stage, and Ultra Plan leaves you on your own during implementation. Superpowers is with you all the way through" (~09:00).
CLAUDE.md
AICodeKing walks through forestchang/andrej-karpathy-skills, a deliberately tiny repo centered on a single CLAUDE.md that codifies four Karpathy-style principles: think before coding, simplicity first, surgical changes, goal-driven execution.[6]AICodeKing — Karpathy-Skill + Claude Code The framing: "you're not installing a feature, you're installing discipline." Complementary philosophy to the superpowers plugin — same goal, much lighter surface area.
Two paths (~02:30):
/plugin marketplace add forestchang/andrej-karpathy-skills then /plugin install andrej-karpathy-skills. Available across projects.CLAUDE.md directly into a repo, or append to an existing one.AICodeKing's test: if the agent starts asking clarifying questions before writing code, your diffs get smaller and more focused, it stops randomly refactoring neighboring files, and it starts thinking in terms of verification rather than "I implemented it, trust me" — those are exactly the signals you want.
You're not adding power in the usual sense. You're removing failure modes. That's why I like it.
He notes the principles are portable to any tool with rule/memory injection — he'd port them into Verdant too. The repo's real value is that you don't have to re-prompt the principles every session — the discipline is baked in by default.
A cluster of video + world-model drops from AI Search's weekly roundup, mostly Chinese. Alibaba's new Attho AI innovation unit (separate from Tongyi and Ant Group) teased Happy Horse 1.0, which topped the Artificial Analysis leaderboard ahead of Seedance 2.0.[2]AI Search — Claude Mythos, Deepseek v4, HappyHorse, Meta's new AI Overworld Waypoint 1.5 generates real-time interactive worlds at 720p/60fps on high-end GPUs and 360p on consumer hardware — open source. InSpatial World builds a persistent 3D scene behind a regular video so you can explore new camera angles.
Showed up on the Artificial Analysis leaderboard under the Attho innovation unit — a new group separate from Tongyi (Wan, Zimage) and Ant Group. Led by Zhang D, former chief architect of Kling. Release targeted for April 30 or later. AI Search warns: "If you see any Happy Horse websites or repositories on GitHub or HuggingFace right now, those are likely fake" (~16:00).
Real-time world generator with interactive controls. 720p/60fps on high-end Nvidia, 360p on consumer; Apple Silicon support promised. Requires at least an RTX 3070. Open source (~17:00). Quality trades against real-time; notable only because typical interactive world generation requires massive compute.
Takes a regular video, builds a persistent 3D internal simulation behind it. You can pan the camera around the scene and view from angles that weren't in the original footage, with consistent motion/physics. Runs real-time at 24fps on H-series, 10fps on a 4090. Obvious implications: robotics/driving training data, interactive content. Code released.
A new mode quietly appeared on the DeepSeek platform — users are calling it a possible V4 Light. Stronger on logic, math, and multi-step problems with lower hallucination rates. Not officially announced as V4; DeepSeek is rumored to launch V4 proper later in April (~14:50).
Given an image + audio + context, generates full animated avatars with facial expressions, eye movements, pauses, hesitation, body language. Handles multiple languages, singing, emotional register shifts. Maintains identity consistency across 45+ minute videos — a significant upgrade over typical long-horizon drift in video models. Idle-mode motion continues during silence. Only a tech report so far; unclear whether weights ship (~33:20).
Zhipu AI released weights for GLM 5.1, which on SWE-Bench Pro beats GPT 5.4 and Opus 4.6 — now the strongest open-source model available.[2]AI Search — weekly roundup including GLM 5.1 and Meta Muse Spark Meta finally shipped Muse Spark after months of silence post-Llama-4 — but despite Meta highlighting its column in blue, it's only state-of-the-art on a handful of non-standard benchmarks (chart-intensive reasoning, humanity's last exam, open-ended health). Gemini 3.1 Pro and GPT 5.4 beat it on average, and Muse Spark is closed-source — a notable break from Llama's open history.
Zhipu AI (ZAI) released the full weights of GLM 5.1 this week — previously API-only (~11:00). The headline benchmark: SWE-Bench Pro scores higher than GPT 5.4 and Opus 4.6. Strong reasoning and agentic coding results more broadly. The weights are ~1.5 TB on HuggingFace — too big for consumer rigs until quantizations drop, but quantizations should follow. Demo: they had GLM 5.1 self-refine for 8 hours to build a Linux desktop with 50+ working apps from scratch (browser, audio player, Telegram-like chat).
Meta's first model since firing the Llama 4 team and (per AI Search) spending billions poaching researchers. Powers Meta AI across Facebook/Instagram/WhatsApp + smart glasses. Multimodal. Closed source, unlike Llama (~19:00).
Performance: Meta's own chart highlights its column in blue even though blue doesn't indicate best-in-category. AI Search's corrected table shows Muse Spark winning on chart-intensive reasoning, Humanity's Last Exam math, some health QA, and agentic research — but Gemini 3.1 Pro and GPT 5.4 beat it on average across standard benchmarks. On the Artificial Analysis leaderboard it lands at #4, tied with Sonnet 4.6 and one point above GLM 5.1. Given it's closed-source and not state-of-the-art, AI Search's conclusion: no clear reason to adopt it over the alternatives.
A short Real Python clip on four psychological effects that hide bugs from reviewers: inattentional blindness, repetition blindness, vigilance fatigue, alert fatigue.[7]Real Python — 4 Psychological Traps That Hide Bugs Lesson: use automated tools (tests, linters, security scanners) for what machines are good at; reserve human reviews for process failures, teaching moments, and preventing accidental duplication. Coda: LLMs change why and when reviews matter — "probably makes them more important."
Automated tools handle most of the surface. Humans should be looking for things machines can't:
The use of LLMs is going to change how and why and when we need to do things like code reviews, and it probably makes them more important.
Sits adjacent to the broader "comprehension crisis" theme running through the week — once AI writes the code, the human review step is where understanding re-enters the system.
90-second Arjay short on preloading — the pattern behind Instagram's "instant" uploads and Gmail's fast sign-in.[8]Arjay McCandless — Preloading #systemdesign Speculative execution: start the expensive work (video upload, mailbox load) on the share screen / login screen while the user is still typing a caption or entering credentials. Tradeoff: wasted compute when the user bails; worth it when UX matters more than unit cost.
Apply when the user is likely to follow through (high completion rate) and when UX matters more than the wasted compute cost from drop-off. Don't apply if completion rates are low or compute is expensive relative to UX gains.
A ~1-minute Dwarkesh clip with historian Ada Palmer on why Machiavelli ended up running Florence's diplomacy at 29 — framed as a story about what happens when your ruling council rotates every 3 months and letters to Milan take 2 days each way.[9]Dwarkesh Patel — How Machiavelli Became a Diplomat at 29 Short clip; full interview presumably released elsewhere.
Palmer's framing: Machiavelli saw his country nearly fall six times by age 12 — he grew up fast and scared. When the elected council only sits for 3 months at a time, and a round-trip treaty letter to Milan eats 2 days each direction (roughly four "emails" before you're out of office), you need bureaucratic connective tissue. Machiavelli became that — first a secretary, then "Soderini's lapdog" referring to the council chair Piero Soderini who effectively headed the Republic Machiavelli served.
This man saw his country nearly fall six times by the time he was 12. Right. He has grown up fast and he has grown up scared.
Two Morning Brew stories on April 12 both circle the same theme: the retail-investing pipeline is shifting to social channels and younger users. "Here come the finfluencers"[10]Morning Brew — Here come the finfluencers and "How'd the kids get so savvy at investing?"[11]Morning Brew — How'd the kids get so savvy at investing? — title-only entries; Morning Brew blocks automated article fetches, so body text isn't available. Readers should consult Morning Brew directly.
Morning Brew's anti-automation measures prevent extraction of article bodies via WebFetch. Both headlines for April 12 cluster around the retail-investor / social-channel dynamic: financial influencers becoming a mainstream investment-advice channel, and teenagers/young adults showing stronger-than-expected investing fluency (likely driven by app-first brokerages and social-media financial content).
Surfaced from the feed for completeness. No further details available without manual access.