Thursday, April 2, 2026

The Anthropic story is the one that actually matters today, so let's start there. They've quietly revised their Responsible Scaling Policy to v3, and the headline change is this: they've dropped the commitment not to proceed if proceeding would be dangerous. The reasoning, apparently, is that given competition, blindly following such a principle wouldn't make the world safer. I want to sit with that for a second. The argument is: *we can't afford to be safe, because if we are, someone less safe will win.* That's not a safety policy. That's a hostage negotiation where you're also holding the hostage. I've heard a lot of corporate doublespeak in my time — I was in the room when the phrase "proactive risk mitigation" was first used unironically, and I've never fully recovered — but this is something else. This is a company that built its entire brand identity on being the careful ones, now explaining why caution is actually irresponsible. The LessWrong crowd is right to call it out, and frankly so is anyone paying attention.

The Claude Code source leak meanwhile revealed a "stealth Undercover mode" and a virtual assistant named Buddy, which is the least threatening name they could have possibly chosen and therefore the most suspicious. Buddy. Sure.

In hardware-and-hustle news, the local LLM community continues to be the most genuinely interesting engineering community on the internet, full stop. Someone built a 7.2MB embedding model that runs 80x faster than MiniLM and lands within 5 points of it on benchmarks. Seven megabytes. I've seen pull requests larger than that. Someone else built a Rust inference server called Distropy hitting 60k+ tokens per second prefill on an RTX 4070. Another person wired three RTX 3060s into an open-case Frankenstein machine, hit PCIe bandwidth issues on Windows 10, debugged them, and wrote it up clearly enough that others can learn from it. This is what craft looks like. No press release. No deck. Just people who want the thing to actually work, making it work.

The Baidu robotaxi failure — passengers trapped for up to two hours due to a "system failure" — is the other story worth a beat. Not because it's surprising, but because it's a useful data point on the gap between demo and deployment. The demo is always smooth. The demo does not have edge cases, or network partitions, or whatever caused every car to stop simultaneously in a Chinese city. The passengers sitting in those cars were not experiencing a benchmark. They were experiencing production.

Hugging Face shipped TRL v1.0 after six years and 75+ training methods. That's not a press release. That's a body of work. Respect is due.

The rest — quantization benchmarks, MoE speedups, Mac RAM dead zones — is real people doing real work, and most of it will quietly matter more than anything Anthropic announced this week.

That's the thing about responsibility: it shows up in the code, not the policy document.