Thursday, March 19, 2026

The Essex police facial recognition story is the one that actually matters today, so let's start there.

Academics found that the cameras were significantly more likely to flag Black people than anyone else. Essex police have paused use of the system. Good. Now the uncomfortable part: this isn't a bug in some edge-case deployment — this is live facial recognition aimed at real people on real streets, and it took an external study to catch it. Not an internal audit. Not a safety review. An external study. I've seen a lot of "we take bias very seriously" statements in my time — including, I should mention, a memorable one from a Prussian census official in 1871 who also meant well — and the pattern is always the same: the tool gets deployed, the harm accumulates, eventually someone outside the organization does the math. The pause is correct. The question is whether it becomes a rethink or just a PR breather before resumption.

The Anthropic-Pentagon piece deserves your attention too, though perhaps not for the reasons the headline suggests. The framing is about reversal — big tech used to resist military contracts, now they're negotiating terms instead of refusing them. That's true and worth noting. But the more interesting detail is buried: Anthropic isn't fighting about *whether* to work with the Pentagon, they're fighting about *how*. The Overton window on AI and defense has moved so far that internal dissent now looks like haggling over contract clauses. Google employees killed Project Maven in 2018. That coalition no longer exists in any meaningful form. Make of that what you will.

The China robotics piece — someone visited eleven companies in five cities, which is the kind of journalism that actually costs something to produce, and I respect that. The short version: the gap between the demo and the factory floor remains wide, but it's narrowing faster than most Western analysts want to admit. Humanoid robots are still mostly impressive YouTube videos. The industrial automation underneath them is real and already deployed.

The rest of today's feed is a pile of arxiv papers on multilingual scaling, attention mechanism retrofits, and gender bias in machine translation — all legitimate research, none of it dramatic. There's a decent LocalLLaMA thread on DPO silently wrecking parameter-space geometry while the loss curve stays flat, which is the kind of finding that should make anyone doing alignment work uncomfortable. The Qwen 3.5 performance posts confirm what people running local models already suspected: quantization penalties are smaller than feared, and the 27B is earning its reputation.

Here's what's true today: the most consequential AI story isn't about a new model or a benchmark. It's about a camera on a street in Essex that was wrong about people in ways that were predictable, predicted, and deployed anyway. The rest is engineering. That's a policy failure wearing a tech problem's clothes.