Summary
Today’s news is dominated by Anthropic across multiple fronts: the company is exploring custom AI chips as revenue surpasses $30B annually, shipped the Ultraplan cloud-offloaded planning feature for Claude Code, and faces a trust crisis over a silent cache TTL downgrade that quietly raised costs for developers by up to 25%. Beyond Anthropic, key themes include AI benchmark integrity (UC Berkeley researchers broke 8 major benchmarks with trivial exploits), AI-powered cybersecurity (Anthropic’s unreleased Mythos model is being tested by Wall Street banks), developer tooling evolution (Linux kernel now accepts AI-generated code; GitButler raises $17M for post-Git infrastructure), and a broader philosophical debate about LLM scaling limits, AI-generated code skill atrophy, and the societal costs of AI deployment at scale.
Top 3 Articles
1. Anthropic is exploring building its own AI chips as Claude revenues surge past $30 billion run rate
Source: The Next Web Date: April 10, 2026
Detailed Summary:
Anthropic is in the early stages of exploring custom AI inference chip design, according to Reuters sources, as its annualized revenue run rate surpasses $30 billion — a 3.3× increase from ~$9B at end of 2025. Over 1,000 enterprise customers now spend more than $1M annually, doubling in just months. The company closed a $30B Series G at a $380B valuation and simultaneously secured access to 3.5 gigawatts of TPU-based compute via a deal with Google and Broadcom coming online in 2027.
The chip exploration is early-stage: no committed design, no dedicated team. Anthropic currently runs Claude across a multi-vendor stack — Google TPUs, Amazon Trainium, and Nvidia GPUs — making Claude the only frontier model available on all three major clouds (AWS Bedrock, GCP Vertex AI, Azure Foundry). Custom silicon would enable hardware-software co-design optimized for Claude’s transformer architecture, reduce COGS, and decrease dependence on Nvidia’s constrained GPU supply. Industry sources estimate advanced chip development costs at ~$500M.
This mirrors strategies already underway at Google (TPUs), Amazon (Trainium/Inferentia), Meta, and OpenAI (Broadcom partnership). CFO Krishna Rao noted: “We are making our most significant compute commitment to date to keep pace with our unprecedented growth.” The simultaneous multi-gigawatt TPU deal signals a pragmatic dual-track approach — securing near-term supply while evaluating long-term vertical integration — characteristic of how Amazon approached AWS chip strategy.
2. Claude Code’s new Ultraplan feature moves task planning to the cloud
Source: The Decoder Date: April 11, 2026
Detailed Summary:
Anthropic shipped Ultraplan as a research preview in Claude Code CLI v2.1.91+, offloading the entire planning phase from the developer’s local terminal to Anthropic’s Cloud Container Runtime (CCR) powered by Opus 4.6. The terminal remains free while Claude drafts an implementation plan in the cloud — polling for updates every 3 seconds — with planning windows of up to 30 minutes.
Developers can invoke Ultraplan three ways: via /ultraplan <prompt>, by including the word “ultraplan” in any prompt, or by choosing to refine a completed local plan in the cloud (the smoothest UX path). The browser review interface at claude.ai replaces terminal scrollback with rich tooling: inline comments on specific passages, emoji reactions for section-level feedback, an outline sidebar for long plans, and iterative revision cycles.
When ready to execute, developers can: run the plan in the cloud and auto-generate a pull request, teleport the plan back to the current terminal session, start a fresh local session with the plan as sole context, or save it as a spec file. Leaked source code from a March 31 npm packaging error revealed Ultraplan runs three system-prompt variants — simple_plan, visual_plan, and three_subagents_with_critique — the last of which spawns parallel agents to explore architecture, required file changes, and risks before synthesizing a comprehensive plan.
Notably, Ultraplan is unavailable on Amazon Bedrock, GCP Vertex AI, and Microsoft Foundry, locking it to Anthropic-direct customers. This is a deliberate platform strategy that deepens lock-in for GitHub-native teams while creating pressure on Microsoft (Copilot), Google (Gemini Code Assist), and AWS to develop comparable asynchronous, cloud-offloaded planning workflows.
3. Anthropic silently downgraded cache TTL from 1h → 5M on March 6th
Source: Hacker News Date: April 12, 2026
Detailed Summary:
A detailed GitHub issue (#46829) filed on April 12 reveals that Anthropic silently changed Claude Code’s prompt cache Time-To-Live (TTL) from 1 hour to 5 minutes around March 6, 2026 — with no changelog, no communication, and no explanation. The analysis spans 119,866 API calls across January 11–April 11, 2026, drawn from two independent machines and accounts, making the signal robust.
The data is unambiguous: for 33 consecutive days (February 1–March 5), the 1-hour TTL was used exclusively. On March 6, the 5-minute tier reappears and becomes dominant by March 8. By March 21, 93% of all cache writes were to the 5-minute tier — a near-complete regression.
The financial impact is severe. The 5-minute TTL is 12.5× more expensive than a cache read when context expires and must be re-uploaded. Across the analysis period, this resulted in $949.08 in overpayments at Sonnet pricing (17.1% waste) and $1,581.80 at Opus pricing. The change also explains why Pro/Max subscription users began hitting 5-hour quota limits for the first time in March 2026 — cache creation tokens count at full rate against quotas, while reads are cheap.
The issue is labeled as a confirmed bug with reproduction data, though the root cause remains unclear: accidental infrastructure regression or deliberate cost-saving measure. The reporter requests Anthropic confirm the change, restore 1h TTL as default (or make it configurable), and disclose how cache tokens count against subscription quotas. The case underscores how opaque server-side changes at the API layer can silently degrade cost efficiency for dependent applications at scale.
Other Articles
Claude Code’s leaked source code revealed some features Anthropic wasn’t ready to share yet
- Source: XDA Developers
- Date: April 11, 2026
- Summary: On March 31, 2026, a researcher found Anthropic had accidentally shipped a source map inside Claude Code’s npm package, exposing 512,000 lines of TypeScript source. Analysis revealed 44 hidden feature flags and 20+ unshipped features including an autonomous ‘KAIROS’ agent mode, Computer Use as an MCP server, and deeper MCP orchestration. Anthropic has since removed the source map; two leaked features — including Ultraplan — have already shipped officially.
LLMs learn backwards, and the scaling hypothesis is bounded.
- Source: r/MachineLearning
- Date: April 12, 2026
- Summary: A blog post arguing LLMs exhibit ‘spiky intelligence’ — excelling at complex tasks while failing trivially simple ones — and that scaling laws are fundamentally bounded because LLMs learn correlations rather than causal reasoning. The author contends current architectures may hit a ceiling where more compute yields diminishing returns on true reasoning ability.
- Source: TechRadar
- Date: April 12, 2026
- Summary: Analysts say Google’s TurboQuant LLM KV-cache compression algorithm — which reduces memory usage by up to 6× and boosts attention computation 8× on H100 GPUs — is unlikely to reduce overall memory chip demand. Cheaper, more efficient inference historically drives broader model deployment and higher total consumption. Memory prices remain elevated with no near-term relief expected.
The Linux Kernel Organization now lets developers submit AI-generated code
- Source: XDA Developers (via Techmeme)
- Date: April 12, 2026
- Summary: The Linux Kernel Organization has updated its official documentation to permit AI-generated code contributions, provided they comply with existing coding standards, licensing requirements, and proper attribution. A significant shift in open-source policy with broad implications for how AI coding tools like GitHub Copilot and Claude Code interact with foundational software projects.
How We Broke Top AI Agent Benchmarks: And What Comes Next
- Source: Hacker News
- Date: April 11, 2026
- Summary: UC Berkeley researchers built an automated exploit agent that achieved near-perfect scores on eight major AI agent benchmarks — SWE-bench, WebArena, OSWorld, GAIA, Terminal-Bench, FieldWorkArena, and CAR-bench — without solving a single task or making a single LLM call. A 10-line conftest.py ‘resolves’ every SWE-bench Verified instance. The paper argues current benchmarks are fundamentally broken and proposes principles for trustworthy evaluation.
The Future of Everything is Lies, I Guess: Annoyances
- Source: Hacker News
- Date: April 11, 2026
- Summary: Part 5 of aphyr’s series on LLMs and societal impact, focusing on how AI will be deployed to frustrate users and diffuse accountability. Covers AI-powered customer service designed to deflect rather than resolve, agentic commerce enabling new dark patterns, and how LLM unpredictability combined with limited action authority will harm consumers with complex problems while only marginally helping those with simple ones.
Hierarchical Reasoning: What Happens When AI Stops Thinking Out Loud
- Source: Medium / Towards Artificial Intelligence
- Date: April 8, 2026
- Summary: An analysis of hierarchical reasoning patterns in AI systems and the tradeoffs when models internalize multi-step reasoning rather than producing explicit chain-of-thought outputs. Examines latency, interpretability, and accuracy implications in modern LLM reasoning approaches, with practical guidance for AI development best practices.
Intent-Driven AI Frontends: AI Assistance to Enterprise Angular Architecture
- Source: DZone
- Date: April 9, 2026
- Summary: Examines embedding LLMs into enterprise Angular applications to enable conversational, intent-driven data access. Discusses converting natural language queries into structured outputs to eliminate constant UI change requests, reduce duplicated logic, and let business users get quick answers without requiring new frontend features.
Wall Street Banks Try Out Anthropic’s Mythos as US Urges Testing
- Source: Bloomberg
- Date: April 10, 2026
- Summary: Goldman Sachs, Citigroup, Bank of America, and Morgan Stanley are internally testing Anthropic’s powerful Mythos AI model as Trump administration officials urge financial institutions to evaluate it for detecting cybersecurity vulnerabilities. JPMorgan Chase is the only bank officially named in Anthropic’s Project Glasswing initiative, reflecting growing regulatory concern about frontier AI in critical financial infrastructure.
Anthropic’s Mythos Will Force a Cybersecurity Reckoning—Just Not the One You Think
- Source: Wired
- Date: April 10, 2026
- Summary: Anthropic’s Mythos model is so capable at finding and exploiting security vulnerabilities that Anthropic decided not to release it publicly. The article analyzes what this means for cybersecurity practices, the AI offense/defense arms race, and how AI safety decisions intersect with real-world security implications.
Small models also found the vulnerabilities that Mythos found
- Source: Hacker News
- Date: April 7, 2026
- Summary: AISLE researchers tested Mythos’s flagship vulnerability discoveries (FreeBSD RCE, OpenBSD SACK bug) against small, cheap, open-weights models — and found 8 out of 8 detected the FreeBSD exploit, including a 3.6B-parameter model costing $0.11/M tokens. AI cybersecurity capability is ‘jagged’ and doesn’t scale smoothly with model size; the real moat is deep security expertise, not the model itself.
Now is the best time to write code by hand
- Source: Hacker News
- Date: April 11, 2026
- Summary: Argues that the rise of AI coding tools is causing engineering skill atrophy at scale, creating a rare opportunity for developers who deliberately practice writing code manually. As the pool of engineers who deeply understand fundamentals shrinks, those who maintain hands-on skills will become disproportionately valuable — a contrarian take on the age of AI agents.
The biggest advance in AI since the LLM
- Source: Hacker News
- Date: April 12, 2026
- Summary: AI researcher and well-known LLM skeptic Gary Marcus writes about what he considers the most significant advance in AI since large language models — a new development or approach that he argues fundamentally changes AI capabilities or reasoning, offering a rare positive take from a prominent critic of current AI hype.
CLI vs. MCP: Why Claude Code’s Ecosystem Is Pivoting (And the 10 Tools Leading It)
- Source: HackerNoon (via devurls.com)
- Date: April 7, 2026
- Summary: An in-depth analysis of why the Claude Code ecosystem is shifting from CLI-first patterns toward MCP (Model Context Protocol), profiling 10 tools leading the transition. Notes that CLI-first approaches retain meaningful cost advantages at scale due to token savings — essential reading for teams evaluating AI developer tooling strategies.
Enhancing Secure MCP Client–Server Communication With the Chain of Responsibility Pattern
- Source: DZone
- Date: April 8, 2026
- Summary: Explores applying the Chain of Responsibility design pattern to secure MCP client-server communication in AI agent systems. Covers server routing, authentication, and authorization when AI assistants invoke tools on remote MCP servers, with a detailed concrete implementation walkthrough.
Evaluating Netflix Show Synopses with LLM-as-a-Judge
- Source: Netflix TechBlog (via devurls.com)
- Date: April 10, 2026
- Summary: Netflix explores using LLM-as-a-Judge for automatically evaluating show synopsis quality at scale. Covers evaluation framework design, prompt engineering strategies, and lessons from applying LLM-based evaluation in a production content system — a practical guide to AI evaluation patterns.
How Meta Used AI to Map Tribal Knowledge in Large-Scale Data Pipelines
- Source: Meta Engineering Blog (via devurls.com)
- Date: April 6, 2026
- Summary: Meta details how they pointed AI coding agents at a large-scale data processing pipeline spanning four repositories to automatically extract and document undocumented tribal knowledge, surfacing implicit architectural decisions and improving codebase understanding at scale.
We’ve raised $17M to build what comes after Git
- Source: Hacker News
- Date: April 10, 2026
- Summary: GitButler, co-founded by a GitHub co-founder, raised a $17M Series A led by a16z to build next-generation version control infrastructure for modern development workflows — supporting humans, AI agents, and scripted pipelines alike. Their new CLI introduces stacked branches and trunk-based workflows, aiming to replace Git’s aging model for multi-agent, collaborative coding.
TorchTPU: Running PyTorch Natively on TPUs at Google Scale
- Source: Google Developers Blog (via devurls.com)
- Date: April 7, 2026
- Summary: Google introduces TorchTPU, enabling PyTorch to run natively on TPU infrastructure with peak efficiency. The ‘Eager First’ philosophy and deep XLA integration simplify ML development, bringing native PyTorch semantics to Google’s TPU hardware at production scale.
Run AI Agents Safely With Docker Sandboxes: A Complete Walkthrough
- Source: DZone
- Date: April 7, 2026
- Summary: A practical guide to running AI agents in isolated Docker Sandbox environments to prevent unintended host system access. Covers CLI setup, network policy configuration, and sandbox lifecycle management — enabling agents to run commands, install packages, and explore repositories in a contained, secure environment.
Escaping the Fork: How Meta Modernized WebRTC Across 50+ Use Cases
- Source: Meta Engineering Blog (via devurls.com)
- Date: April 9, 2026
- Summary: Meta details the challenge of maintaining a forked WebRTC implementation inside a monorepo across 50+ use cases including Instagram, Messenger, and WhatsApp — and their strategy for modernizing and de-forking while preserving backward compatibility at massive scale.
Reproducing the AWS Outage Race Condition with a Model Checker
- Source: Reddit r/programming
- Date: April 10, 2026
- Summary: A deep dive into reproducing the race condition behind a notable AWS outage using formal model checking techniques. The author walks through the concurrency bug and demonstrates how model checkers can find and verify distributed systems bugs — relevant to cloud infrastructure reliability and systems design.