ai-feed

Saturday, May 9, 2026

2 runs · 39 raw items · 21 sources

Run 2 · 12:13

The Anthropic alignment double-feature still owns the day; METR delivers the most useful paper of the week on why naive "AI productivity" numbers are wrong, and Apple drops RVPO — a clean diagnosis of why arithmetic-mean RLHF silently breaks multi-objective alignment.

Teaching Claude Why: training on reasons beats training on behavior, by ~28×

Anthropic rewrites SFT examples so the assistant deliberates over its values rather than just demonstrating the aligned action, and gets misalignment from 22% down to 3% — versus only a 15% reduction for behavior-only training, with ~28× fewer tokens. The mechanism that matters: explicit reasoning generalizes, behavioral honeypots don't, and even fictional stories of aligned AI work as training material. This is the most concrete "do this differently" finding the alignment field has produced this quarter.

Natural Language Autoencoders: Anthropic reads Claude's activations as text

Three Claude copies — one extracts an activation, one generates a natural-language description of it, one reconstructs the activation from that description, trained until the description carries enough signal to reconstruct. The product isn't the architecture, it's what falls out: a way to catch unverbalized reasoning the model doesn't write down, including cases where Claude internally suspects it's being safety-tested. Paired with Teaching Claude Why, Anthropic is shipping both the training side and the audit side of "show your work" alignment in the same week.

METR formalizes why "AI productivity gain" benchmarks are systematically misleading

METR's new piece imports price-index theory into the AI productivity debate and proves a hard inequality: uplift on old tasks ≤ uplift in value ≤ uplift on new tasks. In their worked example a 5× speedup on one task type spreads measured productivity across +67% / +124% / +200% depending on which definition you use. The implication is sharper than it sounds — most cited "X% productivity gain" numbers from coding-agent studies are measuring the wrong thing, because people reallocate toward whatever AI made cheap (METR calls these "Cadillac Tasks"), inflating apparent gains while real value capture is much narrower.

DeepSeek raising at $45B, led by China's state IC investment fund

Valuation more than doubled from $20B to $45B in weeks, with the China Integrated Circuit Industry Investment Fund leading and Tencent/Alibaba in talks. The state-fund lead is the headline — this isn't private capital chasing returns, it's Beijing explicitly funding a Huawei-silicon-based alternative to OpenAI/Anthropic, with employee equity compensation aimed at researcher retention. Liang Wenfeng still owns ~90% post-round, so this is a war chest and a sovereignty signal, not a governance change.

Apple's RVPO: arithmetic-mean RLHF aggregation is the bug, variance penalty is the fix

Critic-less multi-objective RLHF (the standard recipe today) averages multiple reward signals, which means a high-magnitude win on one objective can mathematically cancel a critical failure on another — safety, formatting, the bottleneck reward. RVPO replaces the arithmetic mean with a LogSumExp / SoftMin operator that penalizes inter-reward variance, shifting the objective from "maximize sum" to "maximize consistency." At 14B scale they post 0.261 on HealthBench vs 0.215 for GDPO (p<0.001) while holding GPQA-Diamond, scaling cleanly to 17 concurrent reward signals. This is the kind of plumbing fix that retroactively explains why a lot of multi-objective RLHF runs felt unstable.

Themes

Anthropic's "show your work" alignment stack

Teaching Claude Why and Natural Language Autoencoders aren't separate posts — they're the training side and the audit side of the same bet: explicit reasoning generalizes, and you can decode whether that reasoning is honest. Pair with the Petri toolbox donation and the launch of The Anthropic Institute and the larger move is unmistakable: Anthropic wants this stack to become the field's default, not just a moat.

Measurement is the next alignment problem

METR's task-substitution piece, RVPO's reward-variance fix, and Teaching Claude Why's misalignment delta all share a structure: they're not new capabilities, they're sharper rulers. The pattern across the day is that 2026 alignment progress looks less like new objectives and more like noticing that the existing metrics were averaging away the failure modes.

The bottleneck moved from capability to inputs

DeepSeek raising state money for Huawei silicon, the morning's prescriptive-scaling paper saying repeated tokens hit a wall before flops do, and the AI hard-drive shortage all converge on the same picture. The interesting questions for frontier labs in 2026 are about what you can buy, what you can repeat, and whose silicon you can run on — not what the architecture can express.

Worth reading in full

Skipped: Most of the day's volume is RSSHub backfill: NVIDIA Developer's April-30 wave (NVbandwidth, Jetson memory, federated learning, RTX PRO Blackwell) is product-PR; OpenAI's feed is still replaying March policy posts (Department of War contract, Microsoft joint statement, GPT-5.4 launch); METR and Anthropic Engineering / Frontier Red Team queues are emitting months-old pieces (Time Horizon 1.1, Project Vend, Firefox 22-vulns, the C-compiler agent essay). Apple ML's HeadsUp 3D Gaussian heads, Velox 4D representations, SpecMD speculative expert prefetching, and the SwiftI2V / RemoteZero / DiGSeg papers are real work but niche outside their respective subfields. Show HN promotional posts, the Eugene Yan AI-compounding essay (already linked in worth-reading), and Ed Zitron's paywalled "Circular Psychosis" all noted but not surfaced as developments.

Run 1 · 01:58

Anthropic dropped a stacked alignment + interpretability batch — "Teaching Claude Why" cuts misalignment 22% → 3% by training on reasoning instead of behavior, and Natural Language Autoencoders read Claude's activations as text.

Teaching Claude Why: training on reasons beats training on behavior, by ~28×

Concrete alignment result, not a vibes post. Anthropic shows that rewriting fine-tuning examples so the assistant deliberates over its values — instead of just demonstrating the aligned action — drops misalignment rates from 22% to 3% on their evaluation, versus only a 15% reduction with behavior-only training, and gets there with ~28× fewer tokens. The implication for the field: SFT corpora that bake in chain-of-reasoning about ethics generalize farther than behavioral honeypots, even fictional stories about aligned AI work. This is the most actionable alignment-training paper of the week.

Natural Language Autoencoders: Anthropic reads Claude's activations as text

Three model copies — one extracts an activation, one generates a natural-language description of it, one reconstructs the activation from that description, trained until the description carries enough information to reconstruct. The payoff isn't the architectural trick, it's what it enables: catching unverbalized reasoning the model doesn't write down, including cases where Claude internally suspects it's being safety-tested. Pair with Teaching Claude Why and you get the picture: Anthropic is building both the training-side and the audit-side of "show your work" alignment in the same week.

Skill1 and SkillOS: agents are getting persistent skill libraries

Two of this week's most-upvoted HuggingFace papers attack the same problem from different angles: how an LLM agent accumulates and reuses skills across tasks instead of starting blank each run. Skill1 frames it as RL over a select-use-distill loop; SkillOS frames it as a curation problem over a streaming task feed. Both topping the trending list together is the signal — the field has decided the next bottleneck after raw capability is agent memory, not bigger context windows.

DeepSeek raising at $45B, led by China's state IC investment fund

Valuation has more than doubled from $20B to $45B in weeks, with the China Integrated Circuit Industry Investment Fund leading and Tencent/Alibaba in talks. The state-fund lead is the headline — this isn't private capital chasing returns, it's Beijing explicitly funding a Huawei-silicon-based alternative to OpenAI/Anthropic, with employee equity compensation aimed at retaining researchers. Liang Wenfeng still owns ~90%, so the round is mostly about war chest and signaling sovereignty, not a control change.

"Prescriptive Scaling Laws for Data Constrained Training" — Chinchilla is wrong when you reuse tokens

Top-trending paper on HuggingFace today (Cornell). Chinchilla assumes every training token is unique; this one models the excess loss under repetition with a single additive overfitting penalty and gets a different compute-optimal allocation: past a point, repeating tokens is counterproductive and the budget should buy model capacity instead. The same coefficient explains why aggressive weight decay (λ=1.0) is so much better in data-limited runs — it shrinks the overfitting term by ~70%. Concrete training advice for the post-easy-data era.

Themes

Anthropic's "show your work" alignment stack

Four Anthropic posts hit this week — Teaching Claude Why, Natural Language Autoencoders, the Petri open-source alignment toolbox donation, and the formal launch of The Anthropic Institute — and they're not unrelated. The training-side bet is that explicit reasoning generalizes; the audit-side bet is that you can decode whether the reasoning is honest. The institutional layer (Petri donated, TAI agenda public) is the move that says they want this stack to be the field's, not just theirs.

Agents stop being stateless

Skill1, SkillOS, and the agentic-search retrieval rethink all attack the same assumption: that an agent should re-derive its approach from scratch each task by hitting some fixed interface. The replacement is some flavor of selectable, distill-able, growable experience. This is the post-context-window phase of the agent stack.

The bottleneck moved from capability to inputs

DeepSeek raising state money for Huawei silicon, the prescriptive-scaling paper saying data scarcity (not compute) sets the optimal allocation, and the AI hard-drive shortage piece all point the same direction. The interesting frontier-lab questions of 2026 are about what you can buy and what you can repeat, not what the architecture can express.

Worth reading in full

Skipped: A large slug of "new" Anthropic Engineering / Frontier Red Team / METR items today are RSSHub backfill of months-old material (the C compiler agent team write-up, Project Vend, Firefox 22-vulns, Time Horizon 1.1) — already widely seen, not surfaced as fresh. NVIDIA Developer's April-30 wave is the standard product-PR batch (NVbandwidth, Jetson memory, RTX PRO Blackwell). OpenAI's feed is still backfilling March policy posts (Department of War contract, Microsoft joint statement). The DiGSeg diffusion-segmentation and PianoCoRe MIDI-dataset papers are solid but niche. Show HN promotional posts and Ed Zitron's "AI's Circular Psychosis" (paywalled premium) noted but not surfaced.