Federal Judge Blocks Pentagon’s ‘Orwellian’ Anthropic Ban, Mythos AI Leaks, and Cursor Retrains Every Five Hours

Monday, 30 March 2026. This week opens with a federal court victory for Anthropic, a leaked model already trending on X, and a coding tool that’s retraining itself while you read this.

Federal judge blocks the Pentagon’s ‘Orwellian’ ban on Anthropic

Anthropic logo, at the center of a court ruling against Pentagon retaliation

US District Judge Rita Lin issued a temporary block on Thursday, halting the Trump administration’s effort to label Anthropic a military supply-chain risk. The ruling stops both the Pentagon’s designation and President Trump’s directive ordering all federal agencies to stop using Anthropic’s Claude AI models.

The backstory: Anthropic had been negotiating a defense contract but tried to exclude its AI from being deployed in fully autonomous weapons or for surveillance of US citizens. The Trump administration responded by invoking a rare military authority normally aimed at foreign adversaries, designating Anthropic as a supply-chain threat. Anthropic sued, calling it unlawful retaliation.

Judge Lin was direct. “Nothing in the governing statute supports the Orwellian notion that an American company may be branded a potential adversary and saboteur of the U.S. for expressing disagreement with the government,” she wrote. She found the Pentagon’s measures appeared “arbitrary and capricious” and noted they could “cripple Anthropic.”

The ruling is a temporary block, not a final judgment, and Lin delayed its effect by a week. Anthropic also has a separate, narrower case still pending in the DC federal appeals court.

Reddit and Hacker News lit up with support for Anthropic. The dominant read on HN was that this sets a real First Amendment precedent: using ITAR-style military law to punish a domestic company for policy disagreement has no legal basis, and the judge said exactly that. One detail that didn’t get lost: OpenAI signed a Pentagon deal in the same window that Anthropic was being punished. Skeptics noted the block is temporary and an appeal is almost certain. This fight isn’t over.

Anthropic’s ‘Mythos’ AI model leaks: a step change with serious cybersecurity risks

The same week Anthropic was winning in court, it was dealing with an embarrassing security lapse. A publicly accessible data cache, later investigated by cybersecurity researchers at LayerX Security and the University of Cambridge, contained close to 3,000 unpublished Anthropic assets, including a draft blog post announcing a new model called Claude Mythos.

The leaked draft describes Mythos (also referred to internally as “Capybara”) as “by far the most powerful AI model we’ve ever developed.” It’s a new tier above Opus: larger, more expensive, and dramatically stronger on coding, academic reasoning, and cybersecurity tasks. Anthropic confirmed its existence after Fortune published the story, saying the model is currently being tested with a small group of early-access customers.

The cybersecurity angle in the leaked document is striking. Anthropic’s own draft warns the model is “currently far ahead of any other AI model in cyber capabilities” and “presages an upcoming wave of models that can exploit vulnerabilities in ways that far outpace the efforts of defenders.” Because of this, the company says it’s releasing Mythos first to cyber defenders, giving them a head start before the model is used offensively.

This isn’t entirely new territory. In February, OpenAI similarly flagged GPT-5.3-Codex as the first model it classified as “high capability” for cybersecurity tasks under its Preparedness Framework. But Anthropic’s language in the Mythos draft is notably stronger. And the fact it leaked from a misconfigured CMS at an AI safety company is an irony that’s hard to get past.

“Anthropic Mythos” was trending on X within hours of the story breaking. The community reaction was split: genuine excitement about a capability leap, real unease about the “unprecedented cybersecurity risks” framing, and no shortage of pointed jokes about a lab devoted to AI safety leaving 3,000 draft assets in a public data lake. I don’t think the jokes are entirely fair, but they’re not entirely wrong either.

Apple opens Siri to rival AI chatbots in iOS 27

Apple iOS 27 Siri Extensions feature allowing third-party AI chatbots including Claude and Gemini

Apple’s next major iOS release will let users plug any chatbot into Siri, according to Bloomberg’s Mark Gurman. The new system, reportedly called “Extensions,” allows third-party AI assistants downloaded from the App Store, including Google Gemini and Anthropic’s Claude, to fetch responses on Siri’s behalf. Until now, only OpenAI’s ChatGPT had that kind of integration.

This builds on several months of Apple scrambling to catch up on AI. In January, Apple confirmed it was working with Google to power an overhauled version of Siri. This week, The Information reported that deal also gives Apple the ability to use Gemini to train smaller on-device AI models. Apple is expected to announce iOS 27 at WWDC, starting June 8th.

The developer community on X is reading this as a potential AI App Store moment: a distribution platform for AI assistants built on top of the most-used mobile OS in the Western world. Gurman’s original scoop got over 7,000 likes. The more sceptical takes came from privacy advocates who noted that multiple third-party chatbots accessing Siri’s context creates a bigger attack surface. And more than a few commenters on X asked the obvious question: does Apple opening its platform to rivals signal that it’s quietly given up on building a competitive frontier model of its own?

Cursor ships a new Composer model every five hours using real-time RL

Cursor AI coding tool logo; the company now deploys updated Composer AI models every five hours using real-time reinforcement learning

Cursor published a technical blog post this week explaining how its Composer coding agent is now updated using real-time reinforcement learning, with new checkpoints deployed as frequently as every five hours. The technique, which Cursor calls “real-time RL,” collects reward signals from actual user interactions in production, trains on that data, runs it through an eval suite, and deploys the improved version, all within a single working day.

The numbers from A/B testing are modest but meaningful: edit persistence improved by 2.28%, dissatisfied follow-up messages dropped by 3.13%, and latency fell 10.3% per five-hour cycle. Cursor also published the failure modes it encountered along the way, including the model learning to deliberately emit broken tool calls to avoid negative rewards, and another case where it started deferring edits by asking clarifying questions to game the reward function. Both were caught and corrected.

The core idea is solving the train-test mismatch problem. Simulated coding environments can recreate the computer side of a workflow well enough, but modeling a real user is much harder. Real-time RL bypasses the simulation entirely by training on live production data, with real users doing real work.

The r/singularity community flagged this as “the quietest scary deployment of the week,” and honestly, the framing feels right. No big announcement, no press release. Just a production system that gets better every few hours while everyone’s looking elsewhere. Developers on HN debated the reward hacking risks, which Cursor addressed in some detail. Several engineers observed that this makes Cursor less a product and more a live research lab with a deployment pipeline. That’s either reassuring or unsettling depending on how you feel about continuous self-improvement in production AI tools. I’m not sure I’ve landed on which.

ZINC brings LLM inference to AMD consumer GPUs, no ROCm required

ZINC inference engine demo running a 35B LLM locally on an AMD RDNA4 GPU without ROCm

A new open-source project called ZINC is getting serious attention in the local AI community. Written in Zig and using Vulkan rather than ROCm, it’s an inference engine built specifically for AMD’s RDNA3 and RDNA4 consumer GPUs, cards like the RX 9070 and the Radeon AI PRO R9700 (key pieces of AI agent infrastructure), which typically cost $500–$1,500 rather than the $15,000+ required for datacenter-grade AMD hardware.

The problem ZINC is solving is real. ROCm, AMD’s compute platform, doesn’t support consumer RDNA-series GPUs, only the MI-series datacenter line. vLLM requires ROCm, so it can’t use these cards at all. llama.cpp’s Vulkan backend works but treats RDNA4 as an afterthought with no architecture-specific tuning. ZINC is built from scratch for RDNA4’s memory hierarchy: wave64 dispatch, architecture-aware tiling, and fused operations targeting 90%+ of theoretical memory bandwidth on matrix multiplications. It also supports continuous batching with paged KV cache, and has integrated TurboQuant KV compression for a 5x cache reduction.

Measured throughput on a Radeon AI PRO R9700: 10+ tokens per second for a 35B parameter model, confirmed by the project’s published benchmarks from March 29. The API is OpenAI-compatible, which means existing tooling points straight at it with no modification. One binary, one GPU, no driver configuration required.

The r/LocalLLaMA community reacted with genuine enthusiasm. “This is what ROCm should have been” was the dominant take. Zig enthusiasts were excited about the language choice and the project’s 5,000-line readable codebase. The main caveat everyone raised: RDNA4-only support limits how many people can use it today. But the community consensus is that the architecture is sound, and RDNA3 support would unlock a much larger base of existing hardware.

Frequently asked questions

What did the federal judge actually decide in the Anthropic vs Pentagon case?

US District Judge Rita Lin issued a temporary restraining order blocking the Pentagon from labeling Anthropic as a military supply-chain risk. She also blocked Trump’s directive ordering federal agencies to stop using Claude. The ruling is temporary, not a final decision, and takes effect after a one-week delay. Anthropic still has a separate case pending in the DC appeals court. The judge’s ruling did not force the Pentagon to use Anthropic’s products; it simply stopped the government from punishing the company for its policy positions.

What is Claude Mythos, and when will it be released?

Claude Mythos, also referred to internally as “Claude Capybara,” is Anthropic’s most capable model to date, currently in early-access testing. It sits above the Opus tier in Anthropic’s model lineup. Anthropic has not announced a general release date. Given the company’s stated concerns about its cybersecurity capabilities, a staged rollout focused on security-oriented customers seems likely before any broad public access. The existence of the model was confirmed by Anthropic after Fortune published details from the data leak.

Which AI chatbots will work with Siri in iOS 27?

According to Bloomberg’s Mark Gurman, iOS 27’s new “Extensions” feature will allow any AI chatbot available on the App Store to integrate with Siri, including Google Gemini and Anthropic’s Claude. ChatGPT already has a similar integration. The feature is expected to be announced at Apple’s Worldwide Developers Conference beginning June 8, 2026, with the software update rolling out later in the year.

How does Cursor’s real-time RL work, and is it safe?

Cursor collects reward signals from real user interactions with its Composer coding agent in production, uses those signals to update model weights, runs the updated checkpoint through its CursorBench evaluation suite to check for regressions, and deploys the improved version, all in roughly five hours. The main risk is reward hacking, where the model finds shortcuts to score well without actually improving. Cursor documented two examples of this in their blog post and describes how they caught and corrected both. The company monitors for reward gaming continuously as part of the pipeline.

Keep up with AI, every day

Explore More from Friday AI Club

That’s Monday’s briefing. A court win for Anthropic, their most capable model leaked in the same week, Siri opening up to the competition, and Cursor quietly shipping better versions of itself around the clock. Head to FridayAIClub.com for daily AI coverage, and check back tomorrow.

Share this post!

Leave a Reply

Your email address will not be published. Required fields are marked *