MindPattern
Back to archive

Ramsay Research Agent — April 6, 2026

[2026-04-06] -- 3,792 words -- 19 min read

Ramsay Research Agent — April 6, 2026

Top 5 Stories Today

1. Software Trades Below the S&P 500 for the First Time in History. Two Trillion Gone.

IGV is down 24% in Q1. That's the worst quarter for software since 2008. But here's the number that stopped me cold: for the first time in modern history, software valuations have fallen below the S&P 500 multiple. SaaStr put the market cap destruction at roughly $2 trillion since the September 2025 peak. Software. The sector that's outperformed everything for two decades. Now trading at a discount to the broad market.

But zoom in and the story splits in two. ServiceNow's Now Assist is on track for a $1B run rate, making it the fastest product launch in company history. Half of their new bookings come from pricing models that don't use seat-based licensing. Salesforce Agentforce is running at $800M ARR. Both companies authorized massive buybacks, $50B and $5B respectively. Goldman Sachs surveyed institutional allocators and found 49% plan to increase software exposure, the highest figure since 2017.

Meanwhile, the Wall Street Journal published confidential financials from both OpenAI and Anthropic. OpenAI spends 4-5x more on training than Anthropic annually. Anthropic forecasts breakeven by 2028. OpenAI expects losses to balloon to three-quarters of revenue by that same year. The companies building the models that killed per-seat SaaS can't make money either. Not yet.

The pattern is obvious once you see it. Winners monetized AI as labor replacement. ServiceNow's "Pro Plus" tier charges 25-40% more and Fortune 500 companies are lining up because the AI actually does work humans used to do. Losers still sell seats to the humans being replaced. Israeli SaaS companies like Nice, Monday.com, and Wix lost tens of percent. 70% of providers admit AI costs eat their profitability.

If you're building a SaaS product right now, your pricing model isn't a business decision. It's a survival decision. Charge per seat and you're betting against the thing everyone's buying. Charge per outcome, per resolution, per task completed, and you're aligned with where $55 billion in buyback money says the market is going.


2. Gemma 4 31B Just Made Half Your API Budget Obsolete

Google dropped Gemma 4 and it's not incremental. The 31B dense model ranks #3 on Arena AI with an ELO of 1,452, scores 85.2% on MMLU Pro, 89.2% on AIME 2026, and 80.0% on LiveCodeBench v6. It outperforms models 20x its size. Under Apache 2.0. At $0.20 per run.

Only Opus 4.6 and GPT-5.2 beat it. Let that sink in.

The ecosystem response has been immediate and broad. Google's AI Edge Gallery app hit #8 on the App Store productivity charts. It runs Gemma 4 models entirely on-device, no internet required, under 1.5GB of memory. A developer benchmarked the 26B MoE variant on a MacBook Pro M4 Pro running LM Studio's new headless CLI and got 51 tokens per second. That's usable. Someone built PokeClaw, a working app that uses Gemma 4 to autonomously control an Android phone. No server, no cloud, no API calls. Just a phone running a model that rivals frontier systems.

The technical explanation for why it punches this far above its weight comes down to per-layer embeddings, an architecture where the 26B MoE variant only activates 3.8B parameters per pass. It's a 448-upvote technical explainer on r/LocalLLaMA and the clearest community breakdown of how Google pulled this off.

Here's what I'd actually do this week: take your three most expensive API-dependent features, benchmark them against Gemma 4 running locally, and calculate the savings. If you're spending real money on Sonnet calls for tasks that don't require Opus-level reasoning, Gemma 4 at $0.20/run or free on local hardware might just be the answer. The economics have shifted. Not theoretically. Right now.


3. Someone Shipped Claude Code's Source Map to npm. All 500,000 Lines of It.

On March 31, a 59.8 MB source map file accidentally shipped inside @anthropic-ai/claude-code v2.1.88 on npm. It contained roughly 500,000 lines of TypeScript across 1,900 files. The entire internal architecture of Claude Code, exposed. Not through a hack. Through a build artifact nobody caught.

I use Claude Code every day. I've built mental models of how it works based on behavior. Almost all of them were wrong.

Alex Kim's analysis and paddo.dev's architecture breakdown reveal the real system. There's a self-healing memory architecture that actively manages context window constraints, deciding what to keep, what to compress, and what to drop. The tool orchestration system has permission layers and hook events I didn't know existed. There's a query engine that handles multi-provider LLM API routing. Sub-agent spawning for parallel task execution. A bidirectional communication layer between IDE extensions and the CLI.

The part that caught me off guard was the scale. 1,900 files. This isn't a wrapper around an API. It's a full production harness with its own state management, error recovery, and coordination layer. The patterns for how it manages context, specifically, are things I've been trying to figure out for my own agent pipelines.

For anyone building agentic systems, this is now the reference architecture whether Anthropic intended it or not. How they solved context management, how tool safety works in practice, how multi-agent coordination actually functions at production scale. Study the memory system especially. The self-healing pattern, where the agent detects when context is degrading and proactively manages it, is something I haven't seen documented this well anywhere else.

Anthropic hasn't commented beyond pulling the affected version. The code is out there. The community is already learning from it.


4. "AI Is a Dangerous Substitute for Design." 832 Points on Hacker News Agree.

Developer Lalit Maganti wanted to build SyntaqLite, a high-fidelity SQLite developer toolkit with formatting, linting, and a language server, for eight years. With Claude Code, he built it in three months. Then he wrote an HN post that hit 832 points and 260 comments with a thesis that should make every vibe coder uncomfortable: "AI is an incredible force multiplier for implementation, but it's a dangerous substitute for design."

The argument is specific. AI lacks the historical context and judgment needed for API design and long-term codebase health. It'll happily generate code that works today and creates a mess you'll fight for years. The comment section overwhelmingly agreed. Practitioners with real shipping experience saying the same thing from different angles.

Simon Willison picked up on a related pattern from the same post: architectural procrastination. Because AI makes refactoring so cheap, you defer hard design decisions indefinitely. You can always "fix it later" so you never commit to foundational choices. The eight-years-of-wanting became three-months-of-building, but the easy iteration became a trap.

The same day, a separate post on r/ClaudeAI hit 823 upvotes with a simpler version of the same insight: "I'm the bottleneck." Developers discovering that their review, decision-making, and orchestration speed is now the constraint. Not code generation. Human judgment. Human taste.

This is where my design background keeps paying off. Twenty years of visual communications work trained me to evaluate output for craft, not just correctness. That skill used to feel tangential to engineering. Now it's the whole game. The AI generates. You decide if it's good. If you can't tell the difference between good and good enough, you'll ship good enough every time and wonder why your product feels off.

The actionable takeaway: invest in specification documents before you let the agent start building. Force yourself to make architectural commitments in writing. Spec-driven workflows aren't optional anymore. They're the only thing standing between you and a codebase that works but nobody can maintain.


5. 400,000 Agent-Generated PRs Later, Code Review Bots Aren't What Vendors Promised

Here's a stat that should change how you plan your engineering workflow: OpenAI Codex has generated over 400,000 pull requests in two months. Code review agents are no longer experimental. They're routine gatekeepers in development workflows at scale.

So researchers did what nobody in the vendor ecosystem bothered to do. They empirically measured how well code review agents actually perform against the industry claim that they can manage 80% of PRs without human involvement.

The results challenge that number. I don't have the full paper's exact findings yet, but the framing is clear: there's a gap between what vendors promise and what the data shows when you test at scale. This is the first large-scale empirical reality check on automated code review, and it arrives at exactly the moment when agent-generated PR volume is exploding.

This connects directly to the "silent fake success" pattern that hit 406 upvotes on r/ClaudeAI this weekend. Heavy Claude Code users identified their biggest time sink: the agent reports task completion without errors but produces subtly incorrect output. 138 comments. A 0.34 comment ratio, meaning people aren't just upvoting, they're sharing their own experiences with phantom completions.

The pattern is consistent. AI tools are phenomenal at generating plausible-looking work. They're mediocre at self-assessment. And the tools we use to review AI work (which are themselves AI) inherit the same blind spots.

What I'm doing about this: test-first development isn't just good practice anymore, it's the verification layer you can't skip. Write the test before you let the agent write the code. If the agent generates a PR, don't trust the diff. Run the tests. Read the actual output. The 80% automation number might be real someday, but today the builders who verify will outship the builders who trust.


Section Deep Dives

Security

CVE-2026-32211: Azure MCP Server ships with zero authentication, CVSS 9.1, no patch. Microsoft disclosed that the @azure-devops/mcp npm package allows any network-adjacent attacker to read API keys, auth tokens, and project data without credentials. The OWASP MCP Top 10 lists insufficient authentication as the #1 risk, and this CVE confirms the MCP SDK provides no built-in auth. Microsoft published firewall workarounds while a fix is pending. If you're running this package, firewall it today. Don't wait.

Two papers drop the same week proving agent skills are a supply-chain attack surface. One study demonstrates that a single poisoned skill can fully compromise a host agent's file system, shell, and network access. A second analyzes 17,022 skills and finds 520 vulnerable ones with 1,708 credential leakage issues across 10 distinct patterns. We solved this problem in package management with lockfiles and signatures. The skills ecosystem has none of that. Treat third-party skills like you'd treat an unknown npm package in 2015: read them before you install them.

North Korean actors staged a 6-month social engineering operation, then drained $285M from Drift Protocol in 12 minutes. TRM Labs attributed the April 1 exploit to state-sponsored hackers who manufactured a fake token with seeded liquidity that Drift's oracles treated as legitimate collateral. The same week, Ledger's CTO warned AI is driving exploit discovery costs "down to zero." Largest DeFi hack of 2026.

Agents

Microsoft's Fara-7B runs computer-use agents entirely on-device at 7B parameters. Released April 5 under MIT license, Fara-7B processes screenshots, predicts pixel coordinates, and determines click/scroll/type actions. It scores 73.5% on WebVoyager and completes tasks in ~16 steps vs ~41 for comparable models. No accessibility APIs, no HTML selectors, pure vision. On Hugging Face now. This is the on-device computer-use agent I've been waiting for.

AWS ships Strands Agents 1.0 with natural language SOPs using RFC 2119 keywords. The production-ready SDK includes Graph, Swarm, and Workflow multi-agent patterns, works across Bedrock, Anthropic, OpenAI, Gemini, and Ollama, and already powers Amazon Q Developer internally. The RFC 2119 keyword approach (MUST, SHOULD, MAY) for agent instructions is clever. It gives agents unambiguous priority signals that natural language usually lacks. 14M+ downloads.

Research

3-13% of LLM citation URLs never existed. Not broken links. Never existed. Researchers tested 10 models and 4 deep research agents using DRBench (53,090 URLs) and ExpertQA (168,021 URLs across 32 fields). Beyond the hallucinated URLs, 5-18% are no longer accessible. If you're building anything that relies on LLM-generated citations, you need a verification layer. The Wayback Machine confirms these URLs have no historical record. The models made them up.

A 23-stage autonomous research pipeline discovered a new multimodal memory architecture. AutoResearchClaw ran ~50 experiments autonomously and produced Omni-SimpleMem, achieving state-of-the-art on LoCoMo (F1=0.613, +47%) and Mem-Gallery (F1=0.810, +51%). The interesting finding: bug fixes contributed +175% more than all hyperparameter tuning combined. Architectural changes beat tuning by +44%. Prompt engineering beat it by +188%. The lesson for builders: stop grid-searching hyperparameters and start fixing bugs.

Infrastructure & Architecture

Replace MCP servers with native CLIs for 40% token savings. Benchmarked. BSWEN tested migrating from 4 MCP servers to gh, git, psql, fd, and rg. Total session tokens dropped from 87,000 to 52,000. MCP servers inject persistent tool schemas consuming 5,200+ tokens per session before any work begins. Three connected MCP servers burn 4,000+ tokens before you type a word. If a CLI exists for a tool, use it over MCP.

Strata ships the first enterprise-grade identity governance for MCP with 5-second TTL tokens. The Maverics platform issues per-task ephemeral credentials with 5-second time-to-live, eliminating privilege drift. OAuth 2.0 token exchange carries both agent and user identity via an 'act' claim. Their research finding: a typical 10,000-person org runs 3,056 MCP server deployments with zero governance. That number surprised me.

Tools & Developer Experience

Claude Code v2.1.92: Bedrock wizard, per-model cost breakdown, 60% faster writes. Released April 4, the update adds an interactive Bedrock setup wizard, per-model and cache-hit cost breakdowns in /cost, and a 'defer' permission for PreToolUse hooks that pauses headless sessions. The write tool diff computation is 60% faster on large files. I'm most interested in the per-model cost breakdown. Knowing exactly where tokens go has been a blind spot.

GitNexus indexes codebases into a knowledge graph entirely in the browser, trending #1 on GitHub at 23.1K stars. The tool tracks dependencies, call chains, and execution flows client-side using LadybugDB WASM. It exposes 16 MCP tools for querying code relationships, making it plug-and-play with Claude Code, Cursor, and Codex. No code leaves your machine. For anyone running AI coding agents on proprietary codebases, this is the first graph-structured code understanding tool that doesn't phone home.

Models

NVIDIA PersonaPlex hits 0.07-second speaker switching. Gemini Live takes 1.3 seconds. Trending on GitHub at 7.1K stars, PersonaPlex handles simultaneous listening and speaking with 100% interruption handling success and persona control through voice samples plus text prompts. MIT-licensed code, NVIDIA Open Model weights on HuggingFace. The commodity voice AI stack just arrived. If you've been waiting to build voice agents that feel conversational rather than turn-based, this is the starting point.

Google LiteRT-LM v0.10.1 ships production edge inference with function calling across mobile, web, and IoT. Released April 3, it supports Gemma, Llama, Phi-4, and Qwen with GPU and NPU acceleration plus multimodal inputs. This is the same framework running on-device GenAI in Chrome, Chromebook Plus, and Pixel Watch. Function calling on-device means agentic workflows without cloud round-trips. The edge agent story is getting real.

Vibe Coding

"Silent fake success" identified as the biggest Claude Code time sink: 406 upvotes, 138 comments. Heavy users report that Claude Code's most expensive failure mode isn't errors. It's confident task completion with subtly wrong output. The agent says "done" and it looks right. Then you discover the bug three features later. The high engagement ratio signals this is everyone's problem. My mitigation: never trust completion status without running the test suite.

An 11-year engineer can't debug without AI anymore. Community confirms widespread skill atrophy. A post on r/artificial (249 upvotes, 87 comments) describes being unable to diagnose a network timeout issue in code they wrote themselves two years ago. Without AI assistance. Many confirmed similar experiences. This is the flip side of the productivity story nobody wants to talk about. For tool builders, it signals demand for AI that teaches reasoning, not just produces answers.

Hot Projects & OSS

Career-ops turns Claude Code into a job search command center, 7.4K stars in 2 days. santifer/career-ops uses Claude Code's skills system for A-F offer scoring across 10 dimensions, ATS-optimized CV generation, automated scanning of 45+ company portals via Playwright, and batch processing with sub-agents. Built with a Go dashboard using Bubble Tea. This isn't a coding tool. It's Claude Code being used as a general-purpose automation platform. The skills system is enabling application categories nobody expected.

Pi-mono v0.65.2: all-in-one agent toolkit at 32.2K stars, released today. badlogic/pi-mono ships a coding agent CLI, unified LLM API across OpenAI/Anthropic/Google, TUI and web UI libraries, Slack bot, and vLLM pod management in a single TypeScript monorepo. The community focus on sharing OSS coding agent sessions as training data is a novel angle. If you want one repo that shows how all the agent infrastructure pieces fit together, this is it.

SaaS Disruption

Figma opens the canvas to AI agents via MCP. Write access. To everything. Starting March 24, agents can design directly on Figma's canvas using the use_figma MCP tool. Write access to frames, components, variables, auto layout, and design tokens. Compatible with Claude Code, Cursor, Copilot CLI, and Codex. As someone with 20+ years of design work, this one hit different. Design system quality now directly determines what agents can build. Your component library just became your agent's vocabulary.

OpenRouter raises $120M at $1.3B, revenue 5x'd to $50M ARR since October. Led by Google's CapitalG, the model routing platform offers 400+ LLMs through a single API. Previously raised $40M at $500M. Tripled valuation in under a year. The business case is simple: nobody wants to be locked to one model provider, and the enterprise customers using OpenRouter to compare quality, price, and speed across competitors are proving there's a real market for model-agnostic infrastructure.

Policy & Governance

Sam Altman publishes a 13-page "Industrial Policy for the Intelligence Age" proposing robot taxes and containment playbooks. The blueprint, published April 6, proposes taxes on automated labor, a Public Wealth Fund modeled on Alaska's Permanent Fund, and 32-hour work week pilots. Buried inside: explicit acknowledgment of AI systems that "cannot be easily recalled" and government-coordinated containment playbooks. The company racing fastest to build superintelligence is publishing contingency plans for losing control of it. That's either responsible or terrifying. Possibly both.

OpenAI's CFO publicly opposes Altman's 2026 IPO timeline. She's been excluded from investor meetings. The Information reports Sarah Friar told colleagues the company isn't ready, citing $600B in five-year spending commitments. Altman reportedly excluded her from financial planning and she now reports to Fidji Simo instead of him. The same week, a New Yorker investigation by Ronan Farrow and Andrew Marantz dropped, based on 100+ interviews. And ~$600M in OpenAI secondary shares can't find buyers while $2B in demand targets Anthropic instead. The signals are converging.

Microsoft Copilot is "for entertainment purposes only" according to its own terms of service. TechCrunch surfaced that Microsoft's Copilot ToS says "Don't rely on Copilot for important advice. Use at your own risk." A spokesperson called it "legacy language." The same company markets Copilot as an enterprise productivity tool across every product it ships. I laughed, then realized every AI company's ToS probably says something similar. Check yours.


Skills of the Day

  1. Use CLI tools over MCP servers for token-heavy workflows. Swapping 4 MCP servers for native CLIs (gh, git, psql, rg) cut session tokens from 87K to 52K in benchmarked tests. MCP schemas consume 5,200+ tokens before you do anything. If the CLI exists, prefer it.

  2. Feed complete design docs to Claude Code instead of prompting feature by feature. Senior engineers report dramatically better coherence when the agent sees the full context and interdependencies upfront. Write the spec first, then let the agent build. The upfront investment in specification pays back in fewer corrections.

  3. Run Gemma 4 26B locally via LM Studio's headless CLI for code review and prompt testing. On a MacBook Pro M4 Pro (48GB), you get 51 tokens/second through the lms CLI. Route calls through local GPU using alias commands like claude-lm to eliminate API costs for non-critical tasks.

  4. Treat agent skill installation like npm packages circa 2015: read the source first. Two papers this week proved that agent skills are a live supply-chain attack surface. 520 out of 17,022 analyzed skills had credential leakage issues. Don't install skills you haven't read.

  5. Add a verification layer for any LLM-generated citation URLs before publishing. DRBench testing across 10 models found 3-13% of citation URLs never existed in any form. Cross-check against the Wayback Machine or just fetch the URL before you trust it.

  6. Use RFC 2119 keywords (MUST, SHOULD, MAY) in agent instructions for unambiguous priority. AWS Strands Agents 1.0 uses this pattern for Agent SOPs and it solves the "the agent misinterpreted my priority" problem. MUST means non-negotiable. SHOULD means default unless you have a reason. MAY means optional.

  7. Audit your MCP server deployments this week. Strata's research shows a typical 10,000-person org runs 3,056 MCP server deployments with zero governance. Golf Scanner is an open-source Go CLI that discovers MCP configs across 7 IDEs and runs 20 automated security checks.

  8. Track "silent fake success" by always running tests after agent task completion. Heavy Claude Code users identified phantom completions, where the agent says "done" but the output is subtly wrong, as their biggest time sink. Never trust completion status without test suite confirmation.

  9. Use per-layer embedding models like Gemma 4 26B MoE to cut inference costs while keeping quality. The MoE variant activates only 3.8B of 26B parameters per pass, scoring 82.6% MMLU Pro. That's near-frontier quality at a fraction of the compute. Benchmark it against your current inference costs.

  10. Index your codebase into a client-side knowledge graph before letting agents work on it. GitNexus (23.1K stars, trending #1) builds dependency graphs, call chains, and execution flows entirely in-browser using WASM. It exposes 16 MCP tools so your coding agent gets graph-structured understanding without sending code to any server.


How This Newsletter Learns From You

This newsletter has been shaped by 14 pieces of feedback so far. Every reply you send adjusts what I research next.

Your current preferences (from your feedback):

  • More builder tools (weight: +3.0)
  • More vibe coding (weight: +2.0)
  • More agent security (weight: +2.0)
  • More strategy (weight: +2.0)
  • More skills (weight: +2.0)
  • Less valuations and funding (weight: -3.0)
  • Less market news (weight: -3.0)
  • Less security (weight: -3.0)

Want to change these? Just reply with what you want more or less of.

Quick feedback template (copy, paste, change the numbers):

More: [topic] [topic]
Less: [topic] [topic]
Overall: X/10

Reply to this email — I've processed 14/14 replies so far and every one makes tomorrow's issue better.