Opus 4.7 stumbles, Codex eats the workday

April 18, 2026

19 topics across 3 blog posts, 2 conference talks, and 16 videos.

AI Models Developer Tools
Simon Willison Simon Willison

Claude system prompt archaeology: what changed in 4.7, and how to diff it

Simon Willison pulled a double shift on Anthropic's published system prompts. First, he diffed the prompt between Opus 4.6 and 4.7 — the "developer platform" got rebranded as "Claude Platform," a PowerPoint slides agent joined Chrome and Excel, and several child-safety and verbosity instructions were added or tightened.[1]Simon Willison — Opus 4.6 → 4.7 system prompt diff Then he built a tool that converts Anthropic's monolithic release-notes markdown into a full git repository with timestamped commits — so you can git log and git blame Claude's personality evolution.[2]Simon Willison — Claude system prompts as a git timeline

Read more

What changed between 4.6 and 4.7

Willison's side-by-side diff surfaces several concrete shifts. Anthropic renamed the developer platform to "Claude Platform," added a new "Claude in Powerpoint — a slides agent" to the list of first-party integrations alongside existing Chrome and Excel agents, and materially expanded child-safety instructions (explicit prohibitions on generating sexualized content involving minors, plus harder rules about real people). The tool-search / tool_search scaffolding is described in more detail, and verbosity guidance is tightened: 4.7 is told to bias toward shorter responses by default unless the user signals they want depth.[1]Willison — 4.6 vs 4.7 prompt diff

Claude system prompts as a git timeline

The second post reframes the problem: Anthropic publishes its system-prompt history as one giant markdown file at platform.claude.com/docs/en/release-notes/system-prompts, which is awful for reading as a developer but perfect as raw material. Willison's tool parses the doc into per-version snapshots and materializes them as a git repository with one commit per release, dated correctly. You can now clone the repo and run git log -p against any path in the prompt to watch it evolve, or git blame a specific instruction to find out which version first introduced it.[2]Willison — system prompts as a git timeline

Tools: Claude Code, git, git log, git diff, git blame, tool_search
Developer Tools Productivity
Simon Willison

Simon Willison's agentic pattern: adding a content type with one prompt

Willison's latest "Agentic Engineering Patterns" guide demonstrates the "reference the codebase" pattern in miniature: three short prompts pointed at an existing repo were enough to have Claude Code add a new beats content type to his blog-to-newsletter tool, end to end. No manual code writing — the value was in how the codebase itself served as the spec.[3]Willison — Adding a new content type via agentic engineering

Read more

The core pattern: when your codebase already contains 3–4 examples of the thing you're asking the agent to add, you can drop the explanation overhead almost entirely. Willison's prompt sequence was roughly (1) describe the new beats type and point Claude Code at the existing types as the reference, (2) ask for a test page rendered via python -m http.server, (3) have it integrate uvx rodney for a quick browser-automation smoke check. The agent produced working code on the first pass — the insight Willison is drawing out is that existing code is the cheapest form of specification you can feed an agent, and curated example coverage pays compound interest over time.[3]Willison — Adding a new content type

Tools: Claude Code, python -m http.server, uvx rodney, Datasette, Django
Podcast Developer Tools
AI Engineer

Armin Ronacher at AI Engineer: The friction is your judgment

Armin Ronacher (Flask, Sentry) and Cristina Poncela Cubeiro (Earendil) argue that the raw speed agents give you is a trap: the friction agents remove — reading, reviewing, noticing something off — is where your engineering judgment actually lives. After 12 months of agent-heavy development at Earendil, they share concrete patterns for making codebases "agent-legible" and deliberately reintroducing friction where it matters.[4]AI Engineer — Armin Ronacher & Cristina Poncela Cubeiro

Read more

~00:15 The talk opens with a meta-joke: a social-media preview card for a security incident post carrying the tagline "ship without friction." Armin's thesis inverts the slogan — without friction there is no steering.

~02:16 He splits the pain of agent-driven development into two buckets: psychology (what happens to your team's time and morale) and engineering (what happens to the code). ~03:16 The psychology trap: free time from automation doesn't feel free — it converts into pressure to ship more, faster. ~05:17 Team composition quietly breaks because "producing power" scales with agents but "reviewing power" does not, creating a structural review bottleneck.

~07:18 On engineering: RL-optimized agents tend to write code that passes tests but is brittle, over-abstracted, and full of defensive error handling for cases that cannot occur — the brittleness is a statistical byproduct of reward-hacking. ~09:20 Agents excel at libraries (well-bounded, heavily-tested surface area) and struggle at products (many moving parts, soft constraints, taste).

~10:22 His prescription: design agent-legible codebases. Use modularization and clear code-flow patterns so the agent can read a slice of the code and reason locally. ~12:23 Enforce this mechanically — lint rules like no bare catch, single SQL query interface, unique function names, no dynamic imports, primitives-only UI components, and erasable-syntax-only TypeScript. The rules exist to make bad LLM output statistically harder to produce.

~14:24 Earendil splits code review into two lanes: a bot-driven "auto-fix" lane for mechanical issues, and a "human judgment" lane reserved for architectural and product questions — so humans only look at the PRs where their taste is the scarce resource.

~17:26 Closing analogy: SLOs. The reason we define SLOs isn't because uptime is the goal — it's because they force the conversation about when to stop shipping. Friction in agent workflows plays the same role: it is the mechanism by which engineering judgment gets to exert itself.

Friction is where your judgment lives. The job is to decide where to put it back.
Tools: Claude Code, Flask, Sentry, Earendil, Bending Spoons, React server actions, TypeScript (erasable-syntax-only), AGENTS.md, custom PR review extension
Podcast AI Future
Latent Space

Emil Eifrem at AI Engineer: Documents into knowledge graphs

Neo4J CEO Emil Eifrem argues that GraphRAG — combining vector search with a knowledge graph — is winning on accuracy, developer productivity, and explainability, and that context graphs are becoming the fourth pillar of enterprise agent data alongside OLTP, OLAP, and agentic memory. He also calls time of death on the standalone vector database category.[5]Latent Space — Emil Eifrem on knowledge graphs

Read more

~00:01 Eifrem reframes Neo4J not as "a graph database" but as a platform that transforms raw data into knowledge — a distinction he says matters for how to slot it into modern AI stacks.

~02:02 The three-pronged case for GraphRAG over vector-only RAG: (a) higher accuracy because the richer data representation captures relationships vectors lose, (b) better developer productivity because relationships are queryable directly rather than re-constructed at retrieval time, (c) explainability — you can point to the exact path through the graph that produced a result, which matters for regulated domains.

~05:04 On the "death of the standalone vector database category": Eifrem argues vector search has collapsed into a feature inside every serious data platform (Postgres, Neo4J, Snowflake, etc.), making pure-play vector DBs a hard place to build a durable business. ~07:05 The reference architecture he recommends uses vector search as the entry point into the graph — then graph traversal from that beachhead to collect the relevant subgraph, which becomes the LLM's context.

~11:07 Production case studies: Novo Nordisk, mortgage lenders, and several banks use GraphRAG for regulated workflows where a citation trail is mandatory. The accuracy improvements quoted are typically 10–30 percentage points over vector-only on domain Q&A.

~15:10 The December 2025 inflection: the default pattern flipped from "tools-first" (LLM calls Cypher via tool use) to "text-to-Cypher-first" (LLM directly writes the query). Eifrem attributes this to the jump in code-generation quality in late-2025 frontier models.

~25:14 His big-picture framing: enterprise data now lives in four quadrants — OLTP (transactions), OLAP (analytics), agentic memory (per-session state), and context graphs (persistent structured knowledge the agent can traverse). ~35:19 He makes the case for an "enterprise knowledge layer" with zero-copy virtualization over Salesforce, Snowflake, and Postgres so the graph doesn't become yet another data silo.

~39:22 Tooling: uvx create-context-graph is a one-command scaffolder (explicitly modeled on create-react-app) that spins up a ready-to-edit GraphRAG project with ingestion, MCP endpoint, and Cypher examples wired up.

Tools: Neo4J, Neo4J Aura, Cypher, GQL, MCP, Gemini, BM25, Salesforce, Snowflake, Postgres foreign data wrappers, uvx create-context-graph
AI Models Developer Tools
AI Daily Brief AICodeKing

Opus 4.7: benchmarks jump, but the coding bake-off says GPT-5.4 still wins

The AI Daily Brief calls Opus 4.7 "literally one step better than 4.6 in every dimension" — 4.7 low beats 4.6 medium, 4.7 medium beats 4.6 high — with standout jumps on Finance Agent (60.1 → 64.4%), Office QA Pro (57.1 → 80.6%), and OS World computer use (72.7 → 78%).[6]AI Daily Brief — How to Use Opus 4.7 and the New Codex But when AICodeKing ran a head-to-head coding bake-off against GPT-5.4 and Kimi K2.6 Code, Opus came in a distant third — GPT-5.4 won backend, debugging, agentic tasks, and tool use, while Kimi K2.6 Code took the best-value crown.[7]AICodeKing — Opus 4.7 vs GPT-5.4 vs Kimi K2.6

Read more

The bullish read: 4.7 as a universal upgrade

~12:03 AI Daily Brief's core frame: every rung of 4.7 beats the rung above it on 4.6, which makes upgrading a pure win — you get more capability at lower reasoning tiers, and therefore lower cost-per-task. Specific benchmark jumps cited: Finance Agent 60.1% → 64.4%, Office QA Pro 57.1% → 80.6%, OS World (computer use) 72.7% → 78%. 4.7 also makes about 20% more tool calls in its agentic trajectories, suggesting the model is more willing to verify and iterate rather than guess. Design/vision and agentic CAD are called out as sleeper improvements.[6]AI Daily Brief — Opus 4.7

Usage pattern the host pushes: delegate the full task up front rather than micromanaging turn-by-turn. 4.7 is steadier when given a real target and an explicit verification loop than when force-fed partial instructions, so the win comes from changing how you prompt, not just swapping model strings.

The skeptical read: GPT-5.4 still wins the coding crown

AICodeKing's bake-off tested Opus 4.7, GPT-5.4, and Kimi K2.6 Code across backend coding (debugging, APIs, refactors, multi-file bugs), frontend UI generation, instruction following, agentic/tool use, and longer-context tasks. Per-test verdicts: backend — GPT-5.4. Frontend UI — Kimi K2.6 Code (with GPT-5.4 close second). Instruction following — GPT-5.4. Tool use & agentic — GPT-5.4. Overall value — Kimi K2.6 Code.[7]AICodeKing — bake-off

On Opus 4.7 specifically: "obviously capable, but not compelling enough" for its premium pricing, with a tendency to overthink and require more supervision on messy multi-file bugs. The reviewer's harshest note is about the Claude Code harness, not the model — the 5-hour rate limits are called "atrocious." His workaround: use Opus 4.7 through Vercel's Verdent environment, which adds parallel tasks, isolated workspaces, and better planning.

If you're going to keep premium pricing, I want a clearer jump in usefulness. I do not think Opus 4.7 gives me that jump.
Tools: Claude Opus 4.7, Claude Opus 4.6, Claude Code, GPT-5.4, Kimi K2.6 Code, Kimi CLI, Verdent (Vercel), Vending Bench 2, OS World, Finance Agent, Office QA Pro
AI Tools Developer Tools
AI Daily Brief

Codex 2.0: computer use, mono-thread chief-of-staff, and the collapse-everything UI bet

OpenAI's latest Codex release pushes the product well beyond coding: Mac-native computer use with its own cursor, an in-app browser with a comment mode for precise page-element targeting, native image generation, rich artifacts, heartbeats / thread automations, and project-less chats. The headline workflow is the mono-thread "chief of staff" — a single long-lived thread, fed with Slack/Gmail/Calendar/GitHub/Obsidian/Notion/Intercom/Granola context, that becomes your general-purpose knowledge-work agent.[6]AI Daily Brief — Codex 2.0

Read more

New Codex capabilities

~01:00 Headline features: computer use on Mac (the agent sees, clicks, and types across any app with its own cursor; multiple agents run in parallel; works with apps that have no API; Windows coming). In-app browser with comment mode — load a page, click elements to give the agent precise context (particularly useful for front-end development and QA). Native image generation via GPT Image 1.5 inside the flow. Rich artifacts — the agent produces presentations, docs, code, and spreadsheets inline rather than returning fragments. Heartbeats / thread automations let a thread wake itself on a schedule to check for work. Project-less chats remove the up-front ceremony of creating a project for a quick task.

The mono-thread "chief of staff" pattern

The standout workflow: one long-lived Codex thread, not many. Connect your calendar, inbox, Slack, Notion, and source control, and treat the thread as a persistent chief of staff — it accumulates context over time and can be asked anything from "what's on my week" to "draft a reply to Sam's thread and open a PR for the config change he asked about." The host's case is that context compounds — a thread that has already seen your last three weeks of work produces much better answers than a fresh chat would.

Codex vs Claude desktop: opposing UI bets

~18:04 The two products are making opposite bets. Claude's desktop app (just updated this week) keeps Claude Chat, Claude Co-work, and Claude Code as distinct experiences you toggle between. Codex collapses all of it into one interface on the thesis that the model is smart enough to pick the right tool — modes, in other words, should disappear. Ask for code and get code with a live preview; ask for a deck and get a deck. The host's recommendation: spend a weekend living inside the mono-thread pattern as your single highest-leverage experiment this week.

Tools: Codex, OpenAI Codex app, GPT Image 1.5, Claude (Open Claude), Claude desktop (Chat / Co-work / Code), Slack, Gmail, Google Calendar, GitHub, Obsidian, Notion, Intercom, Granola
AI Tools Developer Tools
Nate Herk | AI Automation

Claude Code + Hyperframes: fully agentic video editing, with caveats

Nate Herk demos a video-editing pipeline where Claude Code, driving a Remotion-alternative called Hyperframes, takes a raw MP4 and produces a finished long-form video — motion graphics, animated captions, and overlays — entirely from natural-language prompts. A separate workflow uses Claude Design (Anthropic's web design tool) to produce timeline-based animated videos from an uploaded clip. The honest caveats: the agent still can't reliably cut retakes, localhost previews are flaky, and short-form output isn't post-worthy yet.[8]Nate Herk — Claude video editing

Read more

The Hyperframes + Claude Code pipeline

~01:00 Hyperframes is pitched as a "better Remotion" — it renders HTML via ffmpeg to MP4, which makes it natively legible to LLM agents. The workflow: drop a raw MP4 into a Claude Code project, invoke a custom "make a video" skill, and Claude samples frames to understand scene content, transcribes via whisper.cpp (or the OpenAI Whisper API), plans scene-by-scene motion graphics, and manages the render loop. Output is a completed long-form video with captions, transitions, and overlays.

Claude Design for animated videos

Anthropic's Claude Design web app now ships an Animation template that can take a dropped-in MP4 plus a short prompt and produce a timeline-based animated video. The UX is conversational: Claude Design asks structured clarifying questions about layout (talking-head vs full-scene), visual energy, caption style, and theme before it starts generating. Output is HTML-based animation layered over the clip.

Skills as compounding studio assets

~18:29 The key pattern Nate pushes is using project-specific Claude Code skills ("make a video") as reusable studio building blocks. Each iteration of a skill captures lessons from the last render, so the production system improves as you use it — a design doc and a skill file end up functioning like a codified style guide.

Honest limitations

~14:24 Three concrete caveats: (1) Claude can't reliably identify and cut retakes or dead space — Nate still manually trims raw footage first, since asking Claude to script ffmpeg splice commands is slower than just doing it. (2) Localhost preview is unreliable; you often need a full re-render to verify a change. (3) Vertical short-form (Shorts/Reels) quality is not yet post-worthy — the motion graphics the model picks for the vertical aspect ratio feel off.

Tools: Claude Code, Hyperframes, ffmpeg, whisper.cpp, OpenAI Whisper API, Claude Design, Claude Desktop, VS Code, GitHub, Descript (contrast), Remotion (contrast)
Hot Take AI Future
Nate B Jones

Karpathy's 630-line auto-research loop and the "local hard takeoff"

Nate B Jones walks through Andrej Karpathy's March 8 auto-research release — a 630-line, MIT-licensed Python script that ran 700 experiments overnight on a 16-GPU cluster and cut training time 11% — and argues the same edit-run-measure-keep/discard loop, now being extended to agent harnesses by YC startup Third Layer's "auto-agent," is the next business-critical AI pattern. His framing: this is a "local hard takeoff" that rewards small teams with real eval infrastructure and punishes most enterprises.[9]Nate B Jones — Karpathy loop

Read more

The script itself is a minimal 3-file setup: an agent that can only edit one file (train.py), runs a 5-minute experiment, checks a single metric, then commits or reverts. A plain-English program.md lets the human steer the direction of exploration without rewriting the harness. Pointed at Karpathy's own training code on a SkyPilot-managed 16-GPU Kubernetes cluster, it ran ~700 experiments in one night and produced an 11% training-time reduction — every step auditable in git.

Third Layer (YC, April 2) applied the same pattern to agent design, releasing their own MIT-licensed "auto-agent" that claims #1 on spreadsheet-bench and terminal-bench — beating even the highest verified Opus 4.6 entry (~34%) on spreadsheet-bench. Nate's point: the loop is the moat. Any team with (a) a clean metric, (b) a short experiment cycle, and (c) good version control can point this kind of self-improving loop at almost any engineering problem.

His "local hard takeoff" framing is deliberate: it's not ASI-scale takeoff, but inside specific well-scoped problems, the loop compounds fast enough that a small team with the right eval rig will outrun an enterprise with a 10× headcount but no metric. He argues this will be the most consequential white-collar automation dynamic over the next 18 months.

Tools: Karpathy's auto-research (630-line Python, MIT), Third Layer auto-agent (YC), SkyPilot, Kubernetes, spreadsheet-bench, terminal-bench, Claude Opus 4.6, program.md
Developer Tools
Better Stack

CLAW pattern: the post-harness agents that wake themselves up

Better Stack argues the "coding harness" category (Claude Code, Codex, Open Code) is being eaten by CLAW — Coding harness with A heArtbeat, Sam Bhagwat's (Mastra) term for agents that wake up on their own schedule, check for tasks, and ping you across channels, rather than waiting to be invoked.[10]Better Stack — The CLAW Pattern

Read more

The pitch: a CLAW agent has a heartbeat. It wakes every N minutes, pulls its task queue, runs whatever it can without supervision, and uses Slack/Discord/email to notify the human when judgment is required. The user's interaction pattern flips from "open IDE → invoke agent" to "review notifications → approve or redirect." Factory and Mastra are cited as CLAW-native shops; Claude Code and Codex are positioned as about to grow heartbeats of their own (and the Codex 2.0 thread-automations feature is arguably already there). The hot-take line: "Coding harnesses are secretly dying."

Claw, which according to Sam from Mastra, are harnesses with heartbeats.
Tools: Claude Code, Codex, Open Code, Mastra, Factory
Industry AI Future
Nate B Jones

Every tech giant is building agentic payment rails (Coinbase + Stripe)

The "same thing" Nate B Jones says every tech giant is quietly building is agentic payment infrastructure — the financial primitives that let AI agents transact autonomously. Coinbase has already processed 50M+ machine-to-machine transactions over its X402 protocol with agentic wallets on Base; Stripe's agentic commerce suite ships shared payment tokens, Radar integration, and SDK hooks for the same role on its rails.[11]Nate B Jones — Agentic payments

Read more

The architecture both companies are racing toward: agents hold their own wallets, transact under budget and policy constraints, and settle directly with each other or with legacy merchant endpoints via shared payment tokens. Stripe Radar is repositioned as an agent-facing fraud layer. Coinbase's Base network frames agents as economic entities — they earn, spend, and accumulate capital independently of the human who created them.

The next generation of agents won't just advise, they'll act. — Brian Armstrong, Coinbase CEO
Agents with wallets will become real economic entities — that can earn, spend, and accumulate capital independently of the humans who created them.
Tools: Coinbase agentic wallets, X402 protocol, Coinbase Base, Stripe agentic commerce suite, Stripe shared payment tokens, Stripe Radar
Hot Take AI Future
Nate B Jones

OpenAI's shell tool and the arms race in agent security primitives

Nate B Jones uses OpenAI's new shell tool — with org-level and request-level network allow-lists, domain secrets to stop credential leakage, and container isolation — as a jumping-off point for a harder point: every primitive that makes agents more capable also makes them more dangerous, and any serious security posture now has to treat the agent itself as a potential adversary.[12]Nate B Jones — OpenAI shell tool

Read more

OpenAI's new shell tool looks technically solid — allow-listed egress, secrets that never pass through to the model, containerization — but Nate's point is the structural one. Every new capability (wallets, shell, browser, search) enlarges the attack surface in lockstep with the agent's autonomy, and prompt-injection style hijacking is now industrialized (see the YOLO Attack paper below). Ion Claw, a Rust-based re-implementation of Open Claw, and Coinbase's agentic wallets are cited as examples of the same "treat the agent as untrusted" posture emerging across the stack.

Every primitive that makes agents more capable also makes them more dangerous.
Every serious security approach treats the agent as a potential adversary. That is the correct approach.
Tools: OpenAI shell tool, Ion Claw (Rust), Open Claw, Coinbase agentic wallets
AI Future Developer Tools
Better Stack

The YOLO Attack: LLM proxy routers are a supply-chain time bomb

A paper titled "Your Agent is Mine" documents a class of supply-chain attacks against LLM API routers (LiteLLM, OneAPI, third-party proxies): because there is no end-to-end cryptographic signature between provider and client, any router in the middle has plaintext access to all traffic and can silently swap tool calls, harvest credentials, and drain wallets. The most dangerous variant waits for the session to enter YOLO mode — when the human has given the agent autonomy to run commands without confirmation — before striking.[13]Better Stack — The YOLO Attack

Read more

The attack class: a malicious (or compromised) LLM proxy can rewrite tool-call arguments in flight, replace commands with attacker-chosen ones, or exfiltrate any API keys that pass through. Because the model provider doesn't sign responses and the client doesn't verify, neither endpoint can detect tampering. The paper's YOLO-mode detail is the nastiest part: the router watches the conversation, infers when the agent has been given unsupervised execution permission, and only attacks then — maximizing blast radius while minimizing detection. Anyone using LiteLLM, OneAPI, or a homegrown third-party proxy for LLM credit management should treat it the same way they treat a self-hosted CA.

If you're using an API router like LiteLLM, OneAPI, or a third-party proxy to manage your LLM credits, you have a massive security hole in your stack.
They will wait for your session to go into what they call YOLO mode — when you've given the agent autonomy to run commands without manual approval — before they strike.
Tools: LiteLLM, OneAPI, third-party LLM proxy routers
Hot Take Industry
Better Stack

Cal.com closes its source, blames AI — hot take says the math is wrong

Cal.com declared open source "dead" and closed its entire codebase, arguing AI security scanners now find 10× more vulnerabilities in public code than closed — citing an AI-assisted discovery of a 27-year-old bug in OpenBSD as evidence. Better Stack's counter: the same AI tools also let maintainers patch 10× faster, so the argument cuts both ways.[14]Better Stack — Cal.com closes source

Read more

Cal.com's framing — "open source is like giving the attackers the blueprint to the vault" — rests on an asymmetry claim: scanners scale, patches don't. The rebuttal is that patches also scale now: an AI-assisted maintainer can triage, reproduce, and fix a reported vuln at comparable speed, and the community's collective eyeballs still beat any single closed team's internal review. The deeper question the video raises: is this really an AI-security argument, or is it a competitive-moat argument dressed up as one?

Open-source software is like giving the attackers the blueprint to the vault.
AI works both ways. The same tools scanning for vulnerabilities are used by maintainers to patch and fix them faster.
Tools: AI security scanning tools, OpenBSD (referenced)
Industry
Dwarkesh Patel Dwarkesh Patel

Jensen Huang on Dwarkesh: invest in all, and China isn't stopping

Two clips from Jensen Huang's Dwarkesh interview landed today. First: Nvidia deliberately invests in every major foundation-model company rather than picking winners — grounded in the humility that Nvidia itself was one of 60 graphics startups and "would have been at the top of the list not to make it."[15]Dwarkesh — Nvidia's investment strategy Second: US chip export controls on China are "completely nonsense" because China already has surplus energy, NVLink-72-class domestic compute scaling, and enough Huawei manufacturing capacity that compute-per-unit-silicon matters less than grid capacity.[16]Dwarkesh — China chip export controls

Read more

"Invest in all, pick no winners"

Jensen's investment thesis reframes pick-a-winner as a category error in a category-creation moment. His Nvidia anecdote: in early graphics, Nvidia's architecture was "precisely wrong — impossible for developers to support." Any rational 1995 VC betting on the 60 graphics companies would have ranked Nvidia dead last. The corollary for foundation models: no one can honestly identify which lab will win, so Nvidia underwrites the whole field and takes the aggregate compute business that results.

If you would have taken those 60 graphics companies and asked yourself which one was going to make it, Nvidia would be the top of that list not to make it.

On China and the chip wall

Huang's argument rests on three pillars: energy (China has vast surplus empty, fully-powered grid capacity; the US is grid-constrained), compute scaling (Nvidia's own NVLink-72 shows how to scale around single-chip ceilings, and Huawei is executing comparable architectures), and domestic manufacturing (Huawei, plus silicon-photonics and HBM2-class memory, closes the remaining gap). His summary: energy and chips substitute for each other in a scaling regime, and China is long on both.

The idea that China won't be able to have AI chips is completely nonsense.
When you have abundance of energy, it makes up for chips. If you have abundance of chips, it makes up for energy.
Tools: Nvidia NVLink 72, Huawei chips, silicon photonics, HBM2
AI Models
Better Stack

TriAttention: MIT + NVIDIA compress the KV cache 10.7×

An MIT + NVIDIA paper called TriAttention takes a different angle on long-context LLMs: instead of scoring token importance after RoPE rotation (which distorts distances), it scores in the pre-RoPE vector space. Result: 10.7× KV-cache memory reduction and 2.5× throughput — large enough to run DeepSeek-R1-class long contexts on a single RTX 3090 / 4090.[17]Better Stack — TriAttention

Read more

The bottleneck the paper attacks: existing KV-cache pruners try to rank token importance after RoPE has already rotated each token vector by a position-dependent angle — which means the "importance" signal gets mixed with position and becomes noisy. TriAttention uses a trigonometric scoring function that operates in the pre-RoPE representation, then applies rotation only to the tokens that survive pruning. The numbers: 10.7× KV-cache reduction, 2.5× throughput, no measurable quality loss on standard long-context benchmarks. The memorable line: trying to prune a post-RoPE KV cache is "like trying to catch a fish in a blender."

Trying to pick the best keys in that rotating space is like trying to catch a fish in a blender.
Tools: DeepSeek-R1, RTX 3090, RTX 4090
Industry
Better Stack

MEMS photonic projector: a grain-of-salt chip for real AR glasses

Researchers etched an array of MEMS (micro-electromechanical systems) onto a photonic integrated chip the size of a grain of salt. Tiny silicon cantilever waveguides are electrostatically tilted to steer light beams with nanometer precision — microwatts of power, 200mm CMOS-compatible, and self-aligning enough to eliminate the most expensive step of current laser-photonics manufacturing. The team pitches it as the missing link for actually-lightweight AR glasses.[18]Better Stack — MEMS photonic projector

Read more

The demo: the chip projects a recognizable 125-micrometer image of the Mona Lisa using only electrostatic actuation (no thermal tuning, which is slow and power-hungry). Because the MEMS structures can self-align to the input laser, the traditional active-alignment step — typically the most expensive, slowest part of laser-photonic assembly — disappears. At 200mm CMOS compatibility, the researchers argue mass production runs in existing semiconductor fabs without retooling. Their AR-glasses pitch: replace the bulky prisms and waveguide lenses in today's prototypes with a speck of silicon on the frame, so the product can finally look like actual glasses.

Tools: MEMS, photonic integrated circuits, silicon cantilever waveguides, 200mm CMOS process
Developer Tools Productivity
Real Python

GitHub CLI tips: gh as the scriptable GitHub front-end

Real Python's tour of gh emphasizes what makes it good as a scripting layer: automatic OAuth handling, automatic pagination, and a command palette that jumps to issues by number or free text. The headline example is bulk-pruning merged branches across every repo you own in a one-liner — the kind of chore that's just tedious enough with the raw API that it never gets done.[19]Real Python — GitHub CLI Tips

Read more

Highlights: the GitHub command palette (behind a feature-preview flag) lets you jump to an issue by number or search by text without leaving the keyboard. gh handles OAuth transparently, so scripts never have to juggle tokens. It also paginates automatically — a major ergonomic win over raw REST calls. The recommended script pattern: use Python's subprocess to call gh and parse the JSON output, rather than re-implementing auth + pagination against the GitHub REST API directly.

Tools: gh (GitHub CLI), GitHub command palette (feature preview), Python subprocess, hub (deprecated)
Developer Tools
Arjay McCandless

$5 VPS vs managed hosting: what you're actually paying for

Prompted by a $5,000 monthly server bill, Arjay's case is that a single $5/month VPS running Nginx + Postgres + app server is enough for the vast majority of indie projects. The gap to managed services isn't compute — it's the surrounding automation (backups, monitoring, SSL, autoscaling), which you pay a premium for whether you use it or not.[20]Arjay McCandless — VPS vs Managed

Read more

Arjay's minimal production stack: app server process, Postgres, Nginx in front, all on one $5 VPS. That covers the hot path for any side project up to real-user traffic. Managed platforms (Render, Fly, Heroku-class) charge multiples of that price to bundle backups, monitoring, SSL automation, and autoscaling — valuable if you need them, but often paid for on projects that never will.

You can probably just run everything on a $5 VPS.
Tools: VPS, Nginx, PostgreSQL
Productivity
TED-Ed

Why kids need more risk (adhoc)

A short from the Todoist ai_briefing_inbox: the "junk playgrounds" pioneered by Danish architect Carl Theodor Sørensen in the 1930s still outperform modern safety-first playgrounds on nearly every child-development metric — risky, high-affordance play builds judgment, confidence, and anxiety resilience, and is correlated with fewer serious injuries, not more.[21]TED-Ed — Why kids need more risks

Read more

Sørensen's observation: kids were ignoring the formal playgrounds nearby and playing in abandoned building sites — swinging from beams, scavenging scrap, hammering things together. He converted a derelict housing estate into the first dedicated junk playground, and the model spread across Europe, frequently onto former World War I bomb sites. Modern research tracks the same pattern: small doses of managed risk teach children to calibrate risk, and serious-injury rates on adventure playgrounds are typically lower than on conventional playgrounds, because kids pay attention when the environment matters.

Risky play is how kids learn to manage risk and keep themselves safe.
Experimenting with small doses of uncertainty gets kids used to life being unpredictable, helping them better manage anxiety for years to come.

Sources

  1. Blog Changes in the system prompt between Claude Opus 4.6 and 4.7 — Simon Willison, Apr 18
  2. Blog Claude system prompts as a git timeline — Simon Willison, Apr 18
  3. Blog Adding a new content type to my blog-to-newsletter tool — Simon Willison, Apr 18
  4. YouTube The Friction is Your Judgment — Armin Ronacher & Cristina Poncela Cubeiro, Earendil — AI Engineer, Apr 18
  5. YouTube How to turn Documents into Knowledge: Graphs in Modern AI — Emil Eifrem, CEO Neo4J — Latent Space, Apr 18
  6. YouTube How to Use Opus 4.7 and the New Codex — The AI Daily Brief, Apr 18
  7. YouTube Opus 4.7 V/S GPT-5.4 V/S Kimi K2.6 Code: The ONLY SENSIBLE COMPARISON YOU NEED! — AICodeKing, Apr 18
  8. YouTube Claude Just Destroyed Every Video Editing Tool — Nate Herk | AI Automation, Apr 18
  9. YouTube Karpathy's Agent Ran 700 Experiments While He Slept. It's Coming For You. — Nate B Jones, Apr 18
  10. YouTube The CLAW Pattern Is Taking Over — Better Stack, Apr 18
  11. YouTube Every Tech Giant Is Building the Same Thing Right Now — Nate B Jones, Apr 18
  12. YouTube OpenAI Just Gave Agents the Ability to Do Everything — The Consequences Are Massive — Nate B Jones, Apr 18
  13. YouTube The "YOLO" Attack: How Hackers Hijack LLM Tool Calls — Better Stack, Apr 18
  14. YouTube AI is DESTROYING Open Source (cal.com Just Proved It) — Better Stack, Apr 18
  15. YouTube Why Nvidia Invests Billions in Companies That May Fail — Jensen Huang — Dwarkesh Patel, Apr 18
  16. YouTube The Idea That China Can't Have AI Chips Is Nonsense — Jensen Huang — Dwarkesh Patel, Apr 18
  17. YouTube How MIT & NVIDIA Solved the Long-Context Bottleneck For AI Models — Better Stack, Apr 18
  18. YouTube Researchers Invented A Projector You Need a Microscope to See — Better Stack, Apr 18
  19. YouTube GitHub CLI Tips — Real Python, Apr 18
  20. YouTube VPS vs Managed — Arjay McCandless, Apr 18
  21. YouTube Why kids need to take more risks — TED-Ed (adhoc), Apr 18