Google’s Gemma 4 Shakes Up Open-Source AI, Anthropic Locks Claude to First-Party Tools, and Fake References Are Invading Science

Monday, 6 April 2026. It’s been a packed 24 hours in AI. Google DeepMind dropped Gemma 4; Anthropic quietly killed off third-party Claude access for subscription users; and a Nature investigation found AI-hallucinated references in 2.6% of scientific papers. A lot to get through.

Google DeepMind releases Gemma 4: the open-source model that’s making the closed-vs-open debate awkward

Gemma 4 multimodal model family from Google DeepMind, available on Hugging Face

Google DeepMind’s Gemma 4 is now live on Hugging Face, arriving with an Apache 2.0 license, a 256K context window for the larger models, and multimodal support across image, text, and audio inputs. Four sizes are available: E2B (2.3B effective parameters), E4B (4.5B effective), 31B dense, and a 26B mixture-of-experts with just 4B active parameters. That range covers everything from mobile deployment to serious workstation inference.

The benchmark numbers are what have people talking. The Gemma 4 31B dense model scores an estimated 1,452 on LMArena (text-only), and the 26B MoE hits 1,441 while burning through only 4B active parameters at inference time. For context, that puts these models well into territory that paid API-only models occupied six months ago.

Architecturally, Gemma 4 brings a handful of genuine innovations: Per-Layer Embeddings (PLE) give each transformer layer its own conditioning signal instead of forcing everything through a single upfront embedding; Shared KV Cache lets the last N layers reuse key-value states from earlier layers, cutting memory overhead for long-context generation; and a variable-aspect-ratio vision encoder means images don’t have to be cropped to a fixed resolution before processing.

The community response was swift. On X, users were reporting successful deployments on Blackwell GPUs within 12 hours of release, claiming 20x inference speedups versus previous generations. Reddit’s r/LocalLLaMA is full of early benchmark posts, and several developers are already packaging Gemma 4 31B for agent pipelines. One comment summed up the mood: “Google just made the open-source versus closed-source argument irrelevant.” Whether that holds once frontier models push even further is genuinely unclear. Right now, the gap is narrower than it’s been in a while.

Framework support landed alongside the model: transformers, llama.cpp, MLX, WebGPU, and Rust bindings are all ready on day one. Hugging Face and Google collaborated on this, and it shows.

Anthropic locks Claude subscriptions to first-party tools only; third-party agents now pay per call

Anthropic ends Claude subscription access for third-party tools like OpenClaw

If you’ve been running AI agent workflows through Claude on a Pro or Max subscription, that arrangement ended on 4 April at 12pm PT. Anthropic officially ended third-party agent access to the Claude subscription tier, starting with OpenClaw and rolling out to all third-party harnesses over the coming weeks.

The mechanism that allowed this was an OAuth loophole (the same authentication method Claude Code uses) that power users had leveraged for years to pipe subscription-tier Claude models into their own automation setups at a flat monthly cost. Anthropic technically prohibited this in its Consumer Terms of Service since at least February 2024, but enforcement was loose until now. The company revised its terms formally in February 2026, making clear that OAuth authentication is reserved exclusively for Claude Code and Claude.ai.

The stated reason is infrastructure strain. In its notification email, Anthropic cited the “outsized strain” third-party harnesses were placing on capacity that needs to be prioritised for core product users. OpenClaw’s founders pushed back. Peter Steinberger noted that he and board member Dave Morin “tried to talk sense into Anthropic” and managed to delay enforcement by a week.

Users who want to continue have two routes: metered “extra usage” billing (charged separately from the subscription) or a standard Claude API key. Anthropic is offering a one-time credit equal to the user’s monthly subscription cost, redeemable by 17 April, plus discounts of up to 30% on pre-purchased usage bundles.

The developer community is not happy. On Reddit’s r/ClaudeAI, users are calling it a bait-and-switch: Anthropic spent years promoting agentic workflows (here’s how to set up Claude skills for your own) while the cheapest path to build them (subscription-backed OAuth) has now been quietly removed. Real-world costs for agent tasks are reportedly ranging from $0.50 to $2.00 per call under the new model, which makes autonomous agent use cases unviable for hobbyists. The cynical read, articulated across multiple Hacker News threads, is hard to dismiss: “First they copy popular features into their closed harness, then they lock out open source.”

Microsoft launches three MAI models in Foundry, and they’re cheaper than the competition

Microsoft has been quietly building its own model stack, and on 6 April it went public with three new MAI models available through Microsoft Foundry:

  • MAI-Transcribe-1 — speech-to-text, starting at $0.36/hour. The company claims it ranks first in 11 of the top 25 global languages on the FLEURS benchmark, beating both Whisper-large-v3 and Gemini 3.1 Flash on the others.
  • MAI-Voice-1 — text-to-speech, at $22 per 1M characters.
  • MAI-Image-2 — image generation, at $5 per 1M tokens for text input and $33 per 1M tokens for image output.

All three are available now via Foundry and the MAI Playground (US only at launch). The pricing on MAI-Transcribe-1 is the eye-catcher for developers: Whisper pricing has long been a reference point, and undercutting it while claiming better benchmark performance is a real challenge to the incumbent if it holds up in production.

The Hacker News post drew limited comments, which isn’t surprising given it’s a launch announcement rather than a technical paper. Still, the developer community is cautiously interested. The dominant read is that Microsoft is de-coupling its AI strategy from OpenAI more deliberately than it has been letting on. Building frontier-grade proprietary models while maintaining the OpenAI partnership as a separate commercial arrangement is a hedge that makes sense, even if Microsoft hasn’t spelled it out publicly.

Hallucinated citations are now in 2.6% of scientific papers, and the number is climbing fast

A Nature investigation has put a number on something researchers have been worrying about: AI-generated fake references are now present in 2.6% of papers accepted by major computer-science conferences in 2025, up from 0.3% in 2024. That’s an eightfold increase year-on-year.

The Nature team worked with Grounded AI, a Stevenage-based company, to analyse 2025 publications across conferences and journals. The finding: at least tens of thousands of papers contain invalid references generated by AI. Some of the errors are subtle: rephrased titles, wrong DOIs, plausible-sounding journals that don’t exist. That’s precisely what makes them dangerous. A hallucinated citation that looks real will pass human review unless someone goes looking for it.

Computer scientist Guillaume Cabanac, who studies fabricated papers at the University of Toulouse, described receiving a Google Scholar alert that one of his preprints had been cited in a dental journal, with the wrong journal name and a DOI pointing nowhere. “I got very concerned,” he told Nature. “I immediately suspected the citation had been hallucinated by AI.”

The academic community’s concern is that this compounds an already struggling reproducibility crisis. LLMs are increasingly being used to help write manuscripts and format bibliographies, and they sometimes produce non-existent references that look convincing enough to slip through. The worry flagged repeatedly on Hacker News is that AI-assisted peer review won’t catch AI-generated hallucinations. They’re both working from the same flawed source material. One researcher’s comment, quoted in the Nature piece, captures the stakes: “We’re building on foundations that may not exist.” I find that genuinely unsettling.

Publishers are developing screening tools, and companies including Grounded AI are offering citation verification services. But the problem is likely to get worse before tooling catches up. Mandatory citation verification before submission is one of the proposals gaining traction in academic circles.

Foxconn Q1 revenue hits $66.6B as AI hardware demand holds steady

Foxconn reported Q1 2026 revenue of $66.6 billion, a 29.7% year-on-year increase, driven by sustained demand for AI hardware infrastructure. The figures cover the first weeks of the current Middle East conflict and are being read by markets as evidence that AI capital expenditure is holding up despite geopolitical uncertainty.

For context, Foxconn manufactures server racks, GPU enclosures, and much of the physical supply chain behind hyperscale AI infrastructure. Revenue growth at this scale suggests the wave of AI data centre investment that has characterised 2024-2026 hasn’t peaked yet, even as rate concerns and geopolitical risk make some investors nervous. The market response has been broadly bullish on AI infrastructure plays, with cautious notes about tail risks from supply chain disruption.

The bigger picture: open vs. closed AI is the fight of 2026

The thread connecting today’s top stories is the tension the briefing’s Trend Signal identified: open-source models getting good enough to challenge paid APIs, while closed providers simultaneously tighten the screws on how their models can be accessed. Gemma 4 is a big win for the open-source camp. Anthropic’s subscription lockout is a reminder that the major labs are also shoring up their monetisation walls. Microsoft launching its own model stack is a third path: proprietary but accessible via API pricing, not locked behind a single ecosystem.

None of this resolves cleanly, and anyone saying it does is selling something. The open vs. closed fight is sharpening because the stakes are getting higher. Where it lands will shape which developers can afford to build what, and who holds leverage over AI infrastructure for the next decade. I don’t think anyone actually knows how this plays out.

Frequently asked questions

What is Gemma 4 and why does it matter?

Gemma 4 is Google DeepMind’s latest open-source multimodal model family, released under an Apache 2.0 licence. It supports image, text, and audio inputs, with models ranging from 2.3B to 31B effective parameters and context windows up to 256K tokens. Its benchmark scores put it in the same tier as commercial models, making it a strong option for developers who want frontier-grade capability without API costs.

Can I still use Claude with OpenClaw or other third-party tools?

Not on a subscription plan. As of 4 April 2026, Claude Pro and Max subscriptions no longer cover usage through third-party agent tools like OpenClaw. You’ll need to use a standard Claude API key with metered pricing, or enable Anthropic’s “extra usage” bundles. A one-time credit equal to your monthly subscription is available until 17 April 2026.

How serious is the hallucinated citations problem in science?

Serious enough that Nature dedicated a major investigation to it. Analysis of 2025 conference papers found that 2.6% contained at least one hallucinated citation, up from 0.3% the previous year. The concern goes beyond the individual errors: if fake references compound across papers that cite papers that cite hallucinated sources, the error can propagate invisibly through a field’s literature.

What are the new Microsoft MAI models?

Microsoft launched three models in its Foundry platform: MAI-Transcribe-1 (speech-to-text, $0.36/hour), MAI-Voice-1 (text-to-speech, $22 per 1M characters), and MAI-Image-2 (image generation, starting at $5 per 1M tokens for text input). All are available now, with the MAI Playground open to US developers for testing.

Explore More from Friday AI Club

For more daily AI news and analysis, head to FridayAIClub.com.

Share this post!

Leave a Reply

Your email address will not be published. Required fields are marked *