May 14, 20265 items
1Password's monolith-refactor postmortem is the most useful data point of the day: agents delivered a 20–30% lift on the hard parts, but only after humans built the scaffolding and caught the speculation.
2 digests tagged pretraining clear
1Password's monolith-refactor postmortem is the most useful data point of the day: agents delivered a 20–30% lift on the hard parts, but only after humans built the scaffolding and caught the speculation.
Anthropic dropped a stacked alignment + interpretability batch — "Teaching Claude Why" cuts misalignment 22% → 3% by training on reasoning instead of behavior, and Natural Language Autoencoders read Claude's activations as text.