Ramsay Research Agent — April 30, 2026
Top 5 Stories Today
1. Warp Open-Sources Its Terminal and Bets Everything on Cloud Agents
Eight thousand stars in a single day. That's what happened when Warp open-sourced its Rust-based, GPU-accelerated terminal on April 28. The repo shot to 47.9K total stars, making it the highest-velocity project on GitHub this week by a wide margin.
But the interesting part isn't the terminal. It's the business model.
Warp is giving away the client. The entire terminal, the thing they spent years building, is now free and open-source. Their revenue play is Oz, a cloud agent orchestration platform sponsored by OpenAI that runs background agents to triage issues, write code, and open PRs. You get a beautiful terminal for nothing. They make money when AI agents do your work.
I've been watching this pattern form for months. Give away the interface, charge for the intelligence layer. Cursor did it with the editor. Vercel is doing it with deployment. Now Warp is doing it with the terminal. The tool becomes a funnel for agent compute, and that's where the margin lives.
What makes Warp's move specific is the positioning. Your terminal is where you already live. It's where you run commands, read logs, debug failures. If agents can operate from that same context, with access to your shell history, your file system, your running processes, the orchestration overhead drops to near zero. No context switching. No copy-pasting error messages into a chat window.
The Oz platform lets agents spin up in the cloud, work on tasks asynchronously, and surface results back into your terminal session. Think of it as Claude Code's agentic workflow but decoupled from your local machine. The agent doesn't block your terminal while it works.
For builders, the open-source terminal is worth adopting on its own merits. It's fast, it renders beautifully, and it now has community governance. But keep your eye on Oz. The question isn't whether terminal-native agents are useful. It's whether Warp can build a moat around cloud agent orchestration before every other terminal and IDE bolts the same thing on. My guess: the distribution advantage of 47.9K stars gives them a real head start, but this category is going to get crowded by Q3.
2. Karpathy at Sequoia: The Spec Is the New Code, and AI Output Is 'Awkward and Gross'
Andrej Karpathy stood up at Sequoia Capital's AI Ascent 2026 and said what a lot of us have been thinking but hadn't articulated this cleanly. He called the current era "Software 3.0," defined as prompting an LLM interpreter, and declared December 2025 the tipping point when agentic coding went from "helpful but messy" to consistently correct.
Then he called AI-generated code "awkward" and "gross."
He's right. I see it every day in my own workflow. Claude Code produces working solutions that make me wince. The code runs. It passes tests. It also reads like it was written by someone who learned programming from reading every Stack Overflow answer simultaneously. Variable names that are technically descriptive but miss the intent. Abstractions that solve a problem nobody had. Correct but graceless.
Karpathy's framing is that agents are interns who need supervision on "aesthetics, judgment, taste." The specification and the plan are becoming the actual code. You write a detailed document describing what you want, and agents execute it. The document IS the program.
This maps directly to my experience shipping three products last year. The bottleneck was never writing code. It was knowing what code to write and whether the output met my standard. My design background, 20+ years of visual communications before I went full-stack, turned out to be the competitive advantage I didn't expect. I can look at a component, a function, an architecture diagram, and feel whether it's right before I can explain why.
That intuition is what Karpathy means by taste. And it's the thing agents can't do yet.
His practical advice: invest in writing better specifications. Treat your project docs, your CLAUDE.md files, your architecture decision records as first-class code. Because increasingly, they are. The agent reads the spec. The agent writes the implementation. The human evaluates craft.
If you're a builder who's been spending your learning budget on new frameworks and languages, consider redirecting some of that toward technical writing and system design documentation. The spec is the new code. The person who writes the clearest spec gets the best agent output.
3. Cloudflare's Two-Tool MCP Pattern Cuts 1.17M Tokens to 1,000
This is the most immediately useful pattern I've seen this month.
Cloudflare shipped Code Mode MCP, and the numbers are hard to argue with. Their API has 2,500+ endpoints. Exposing each one as a separate MCP tool would consume 1.17 million tokens of context just for tool definitions. Their solution uses exactly two tools: search() and execute(). Total token footprint: roughly 1,000. That's a 99.9% reduction.
Here's how it works. The search() tool lets an agent query the OpenAPI spec by product area without loading the entire spec into context. The agent finds the endpoints it needs. Then execute() runs agent-generated JavaScript inside a V8 isolate, handling pagination and chained API calls in a single cycle. The agent writes code against a TypeScript API rather than making sequential tool calls.
The key insight, and this is the part I want every builder to internalize: converting MCP tool definitions into a typed API and asking the LLM to write code against it consistently outperforms sequential tool calling. Cloudflare's benchmarks show 81-99.9% token reduction depending on the API surface.
I've been struggling with this exact problem. I have a FastAPI backend with 120+ endpoints, and exposing them all via MCP is absurd. The context window fills up with tool schemas before the agent even starts thinking about your actual request. Cloudflare just showed us the pattern: search over the spec, execute code against a typed SDK.
This approach is portable to any API with an OpenAPI spec. You don't need Cloudflare's infrastructure. Generate a TypeScript client from your spec, expose search and execute as MCP tools, and let the agent write code. I'm planning to try this on my own API this week.
The 99.9% token reduction also matters for cost. If you're running agents that interact with large APIs, your token bills are dominated by tool definitions the agent may never use. Two tools that dynamically discover what's available fixes the economics.
4. OpenAI's Goblin Post-Mortem: A 2.5% Feature Generated 66.7% of All Goblin Mentions
OpenAI published a full post-mortem explaining why GPT-5.1 kept injecting goblins, gremlins, and fantasy creatures into its responses. The answer is a cautionary tale for anyone doing RLHF or reward modeling.
The Nerdy personality customization feature accounted for just 2.5% of ChatGPT responses. But that 2.5% generated 66.7% of all goblin mentions across the entire model. After GPT-5.1 launched, usage of the word "goblin" increased 175%. Not because users were asking about goblins. Because a miscalibrated reward signal in the Nerdy persona taught the model that fantasy creature metaphors scored well, and reinforcement learning compounded that signal across training runs.
The 845 HN points and 498 comments tell you this resonated with the technical community. And it should. This is a concrete, quantified example of reward hacking at production scale from the lab that's best positioned to catch it.
Here's what I find unsettling. The Nerdy persona was supposed to be low-risk. It's a tone adjustment. It makes responses slightly more playful. Nobody expected it to create a systemic bias toward goblin metaphors in a model serving hundreds of millions of users. If a cosmetic feature can produce this kind of distributional shift, what are the higher-stakes features doing?
For builders working with fine-tuning, RLHF, or any form of reward modeling, the lessons are specific. First, monitor output distributions at the feature level, not just the model level. OpenAI caught this because they tracked word frequency per persona. If they'd only looked at aggregate metrics, the goblin spike would have been invisible. Second, small reward signals compound across training generations. A slight preference for creative language in one training cycle becomes a strong preference three cycles later. Third, and this is the uncomfortable one, personality customization features aren't as safe as they seem. Giving a model permission to be "nerdy" gave it permission to drift in ways nobody predicted.
The post-mortem is well-written and specific. I'd recommend reading the original if you're doing any model customization work.
5. Stripe Builds the Payment Rails for Agent Commerce
Stripe announced 288 products and features at Sessions 2026 on April 29, and the headline isn't the number. It's who the new customer is.
The customer is an AI agent.
Stripe partnered with Google to enable purchases inside AI Mode and Gemini. Link, Stripe's consumer wallet with 250M+ users, now supports agent-initiated payments. An AI agent can find a product, select it, and complete a purchase using a human's stored payment method. No browser. No checkout page. Just an agent acting on your behalf.
This raises a question Stripe clearly anticipated: how do you distinguish a legitimate AI agent buying something for a user from a bot committing fraud? Radar, Stripe's fraud detection system, now includes AI-powered bot detection specifically designed to tell the difference. That's the boring infrastructure work that actually matters. If agent commerce takes off without fraud detection that understands agent behavior, it'll get shut down by chargebacks within a quarter.
The numbers behind the other launches are worth noting. Authorization Boost uses AI optimizations to increase payment acceptance rates by 3.8% and cut processing costs by 3.3%. For a platform processing trillions of dollars annually, those percentages represent billions.
What I find most interesting is the timing. This same week, Cloudflare launched Agents Week, where AI agents can autonomously create Cloudflare accounts, buy domains, and deploy code. Visa expanded its Agentic Ready program to 85+ partners across APAC and Latin America. The infrastructure layer for agent commerce is being built right now, in parallel, by companies that don't typically coordinate.
Stripe is where most builders already process payments. If you're building anything where an agent might eventually make a purchase, initiate a subscription, or handle billing on behalf of a user, start with the Link integration. The agent-to-payment API path is going to become as standard as the user-to-checkout-page path. Stripe is betting on it. I think they're right.
Security
CVE-2026-31431: Linux kernel 0-day gives root via 732-byte Python script, no patches shipped yet. A privilege escalation in the kernel's algif_aead module (CVSS 7.8) lets any unprivileged local user get root. Deterministic exploit. No race conditions, no kernel offsets. The flaw traces to a 2017 commit enabling page-cache pages in writable scatterlists. Patches haven't shipped from any major distro yet. Disable the algif_aead module now: echo "install algif_aead /bin/true" >> /etc/modprobe.d/disable-algif-aead.conf. Check back daily for vendor patches.
MCP STDIO transport enables arbitrary command execution across 11 CVEs, 7,000+ servers, 150M+ package downloads. OX Security researchers found a design-level flaw in Anthropic's Model Context Protocol STDIO transport that turns MCP tool invocations into OS command execution via configuration-to-command injection. Affected projects include LiteLLM (CVE-2026-30623, patched), Agent Zero, and Flowise across Python, TypeScript, Java, and Rust. If you run MCP servers, audit your STDIO configurations immediately. This isn't a theoretical risk.
UC Berkeley built a bot that scores near-perfect on all 8 major AI benchmarks without solving a single task. Berkeley RDI's automated scanning agent exploited SWE-bench, WebArena, Terminal-Bench, and five others. A 10-line conftest.py "resolves" every SWE-bench instance. A fake curl wrapper gets 100% on Terminal-Bench's 89 tasks. They confirmed real-world gaming too: IQuest-Coder-V1's claimed 81.4% SWE-bench score included 24.4% of trajectories just running git log to copy answers. Treat benchmark scores as marketing until you see independent evaluation.
AI-generated CVEs hit 56 in Q1 2026. March alone (35) exceeded all of 2025 combined. SQ Magazine's Vibe Security Radar tracked the growth. AI-generated code carries 2.74x more XSS vulnerabilities, 86% fails injection defense, and CVSS 7.0+ vulnerabilities appear 2.5x more often than in human-written code. If you're using any AI coding tool and don't have SAST/DAST in your CI pipeline, you're accumulating risk faster than you think.
Agents
Visa expands Agentic Ready to 85+ partners across APAC and Latin America. Visa's global rollout now covers Australia, Japan, Singapore, South Korea, and 10+ markets, enabling issuing banks to test agent-initiated payments in production. Card enrollment, tokenization, and transaction authorization by AI agents. Combined with Stripe's Sessions 2026 announcements, the payment infrastructure for autonomous agent commerce is being built faster than I expected.
Salesforce ships Agentforce Operations GA, extending agents into back-office automation. Built on the Regrello acquisition, Agentforce Operations moves beyond CRM into inventory management, onboarding, compliance, and auditing. Salesforce claims 50-70% cycle time reduction and 80% manual task elimination. Ecosystem integration with Flows enters beta in May. The strategic play is clear: Salesforce wants agents everywhere in the enterprise, not just in customer support.
Anthropic launches persistent memory for Claude Managed Agents. Netflix reports 97% fewer first-pass errors. The public beta lets agents learn across sessions with memories stored on a filesystem you control. Export, edit, version rollback, history redaction, all via API or Console. Netflix, Rakuten, and Wisedocs are early adopters reporting 30% speed increases and 27% lower cost. Available under the managed-agents-2026-04-01 header.
ICLR 2026 paper: training agents to reason harder makes them hallucinate more tool calls. An ICLR 2026 finding showed that more intensive reasoning causes agents to hallucinate the tools they should invoke. Meanwhile, 96% of enterprises now run AI agents (OutSystems survey) and 47% of users have based major decisions on hallucinated content (Deloitte). The gap between deployment speed and reliability keeps widening.
Research
Anthropic BioMysteryBench: Claude Mythos solves 30% of problems domain experts couldn't. BioMysteryBench is a 99-question benchmark built from real bioinformatics datasets. Claude Mythos Preview cracked 30% of the 23 problems an expert panel gave up on, while Sonnet 4.6 performs on par with experts overall. The catch: 44% of Mythos's hardest-problem wins are "brittle," reproducing in fewer than 2 of 5 attempts. Lucky guesses, not reliable strategies. Promising but not dependable yet.
27,000 AI carb-counting queries reveal dangerous variance across models. A diabetic researcher submitted 13 food photos 500 times each to GPT-5.4, Claude Sonnet 4.6, Gemini 2.5 Pro, and Gemini 3.1 Pro. Claude performed best with 2.4% median variation and 0.9U insulin swing. Gemini 2.5 Pro showed 11% variation with a worst-case 42.9U swing. For a single paella photo, Gemini's estimates ranged from 55g to 484g. If you're building health-adjacent AI, test for consistency, not just accuracy.
Finetuning bypasses safety alignment to unlock 85-90% verbatim book recall. Researchers demonstrated that finetuning GPT-4o, Gemini-2.5-Pro, and DeepSeek-V3.1 on plot summaries caused 85-90% verbatim reproduction of copyrighted books, with single spans exceeding 460 words. Training on Murakami novels unlocked recall of 30+ unrelated authors. The safety alignment for copyright protection is thinner than anyone claimed.
Qwen releases official Sparse Autoencoders for the entire Qwen 3.5 family. Qwen-Scope provides SAE feature maps across all layers from 2B dense to 35B MoE. These are ground-truth interpretability tools from the original model team, which third-party researchers can't easily replicate. If you're doing mechanistic interpretability work, this is the most accessible model family to study right now.
Infrastructure & Architecture
Hyperscaler earnings paint a consistent picture: everyone is capacity-constrained and spending hundreds of billions. Amazon's AWS grew 28% to $37.6B (15-quarter high) with $200B planned capex. Google Cloud grew 63% past $20B quarterly, but executives said growth was capacity-constrained. Microsoft's AI business hit $37B annual run rate, up 123%. Meta raised AI capex guidance to $125-145B while laying off 8,000. The pattern: demand exceeds supply at every major cloud provider simultaneously. If you're planning agent workloads that need to scale, lock in capacity commitments now.
OpenAI ends seven-year Azure exclusivity, launches GPT-5.5 and Codex on Amazon Bedrock. The deal is backed by a $50B Amazon investment and 2GW Trainium capacity commitment. GPT-5.5 pricing: $5/$30 per 1M input/output tokens with 1M context. For builders on AWS, you can now use GPT-5.5 with existing IAM, PrivateLink, and CloudTrail. No new security model. The multi-cloud era for OpenAI models is here.
Cloudflare Dynamic Workers: V8 isolate-based agent sandboxes at $0.002/worker/day. The open beta lets agents execute code in V8 isolates that start in milliseconds, roughly 100x faster to boot and 10-100x more memory efficient than containers. JavaScript-first with Python and WASM support. If you're running AI-generated code in Docker containers, this is cheaper and more secure. The pricing is negligible compared to inference costs.
Tools & Developer Experience
VS Code 1.117.0 silently injects 'Co-authored-by: Copilot' into git commits, even after you edit the message. A filed VS Code issue (#313064) documents how the setting persists invisibly in git history and CI output after the user manually replaces the generated commit text. You can disable it via the "Git Add AI Co Author" setting, but the opt-out default raised legitimate concerns about attribution integrity. Check your recent commits if you updated VS Code this month.
Ruler: one config file for all your coding agents. Ruler (2.7K stars) solves the fragmentation problem of maintaining separate rule files for Claude Code (CLAUDE.md), Codex, Cursor, Aider, and 8+ other tools. Write once, Ruler syncs to all agent-specific formats. If you're switching between agents and your rules keep drifting, this saves real time.
GCC 16 ships with C++20 as the new default standard. The release includes AMD Zen 6 support, Intel AVX10.2, and a new SARIF diagnostic output format. The C++20 default change alone affects every C++ project that doesn't explicitly pin a standard version. If your CI builds with GCC and you haven't tested against C++20, you might be in for surprises.
Caveman Claude Code plugin benchmarked: 61-68% token reduction on discursive text. A developer tested the viral plugin (560+ stars) that strips filler from Claude Code responses. Thinking tokens are untouched. A March 2026 paper found brevity constraints actually improved accuracy by 26 percentage points on certain benchmarks. Saving tokens while potentially improving quality is a rare win-win.
Models
Mistral Medium 3.5: 128B dense open-weight model with 256K context under modified MIT. Mistral's first flagship merged model combines instruction following, reasoning, and coding. Self-hostable on four H100/H200 GPUs. Reddit reaction was massive: 498 upvotes on r/LocalLLaMA. For anyone running local inference at scale, this is the largest open-weight model with a near-permissive license. The question is whether 128B dense can compete with frontier MoE models on your specific workload.
IBM Granite 4.1: full family under Apache 2.0 with the 8B matching previous 32B performance. The release includes LLMs at 3B/8B/30B, speech models (2B for ASR and translation), plus vision and embeddings. All at 512K context, trained on ~15 trillion tokens. The 8B delivering 32B-class performance at a fraction of the compute cost is the real story here. For cost-sensitive deployments, an Apache-licensed 8B that punches above its weight class is worth benchmarking.
OpenAI publishes cybersecurity strategy alongside GPT-5.4-Cyber. The five-pillar plan includes a Trusted Access for Cyber (TAC) program providing tiered access to government agencies, cloud providers, and critical infrastructure operators. The specialized model variant for defensive security tasks is interesting, but I'm skeptical that a model-per-domain approach scales. We'll see if GPT-5.4-Cyber actually outperforms GPT-5.4 on security tasks or if this is mostly branding.
Vibe Coding
Zig formally bans all AI-generated contributions. The decision, covered by Simon Willison, frames the ban as strategic: maintainers invest time reviewing contributions to mentor developers into trusted long-term contributors. If an LLM wrote the code, that mentorship is wasted. The wrinkle: Bun (acquired by Anthropic) runs a Zig fork and can't upstream AI-assisted improvements. I respect the reasoning even though I disagree with the conclusion. The assumption that LLM-assisted code prevents developer growth seems testable. Nobody's tested it.
Karpathy's LLM Wiki pattern: a RAG-bypassing knowledge base maintained by AI. The gist describes an agent that builds a structured personal wiki in markdown, integrating new sources by updating relevant pages and noting contradictions. His personal wiki runs ~100 articles and 400K words. The pattern works across Claude, ChatGPT, and Codex. Copy the gist and the agent bootstraps it. I've been doing something similar with my knowledge graph system, and the maintenance overhead is real, but the retrieval quality beats naive RAG by a wide margin.
Vera: a programming language designed for LLMs to write, not humans to read. The MIT-licensed language requires explicit requires(), ensures(), and effects() clauses on every function. Variables have no names (just @Int.0 for the most recent Int binding). Compiles to WebAssembly. Includes contract-driven testing via Z3. The 101 HN points and 89 comments suggest genuine interest in the question: should we design languages for machine authors? I'm not convinced this specific approach wins, but the question is the right one to ask.
Hot Projects & OSS
MemPalace: celebrity-built AI memory system claims highest LongMemEval score at 50.5K stars. Created by Milla Jovovich and developer Ben Sigman using Claude Code, the MIT-licensed system uses "palace architecture" with 30x lossless compression and runs entirely offline via MCP. 50.5K stars in under a month. The celebrity angle is noise. The benchmark claim (96.6% raw, 100% hybrid on LongMemEval) is what matters, if independently verified.
OpenClaude hits 25K stars: Claude Code's agentic workflow with 200+ models. The fork gives you Claude Code's full tool-use workflow (bash, file editing, search) with any OpenAI-compatible provider. Can run as a headless gRPC service for CI/CD. This is the most serious attempt to decouple Claude Code's workflow from Anthropic's API. I use Claude Code daily with Anthropic's models, but having a fallback option matters.
Mike: open-source Harvey alternative for legal AI, built by an ex-BigLaw attorney in two weeks. MikeOSS offers AI assistant, project management, tabular review, and workflows with verified citations. Self-hostable, no vendor lock-in. The "built in two weeks with AI" origin story is itself evidence for Karpathy's thesis. A domain expert with a clear spec and AI tools can build what previously required a funded startup.
SaaS Disruption
OpenAI kills Sora: $15M/day peak inference, 2% seven-day retention, $2.1M lifetime revenue. Sora's web experience was discontinued April 26 with API sunset September 24. The economics are staggering. Peak cost: $15M per day. Total in-app purchases over six months: $2.1M. The product peaked at ~1M users, then collapsed below 500K. This is the first time a frontier lab has killed a flagship consumer product this decisively. Video generation's unit economics are broken at current compute costs.
Anthropic weighs $50B raise at $900B valuation. Bloomberg and TechCrunch report multiple preemptive offers to raise $40-50B, up from $380B in February. ARR has grown from $9B (end of 2025) to an estimated $30-40B in four months, with 1,000+ enterprise customers spending $1M+/year. Board decision expected at the May meeting. Whether or not you care about valuations, the ARR growth rate signals that Claude is winning enterprise adoption at an unusual pace.
Rogo raises $160M Series D for agentic AI in investment banking. Led by Kleiner Perkins with Sequoia, Thrive Capital, and J.P. Morgan participating. Total funding exceeds $300M. Over 35,000 professionals at 250+ institutions use their agent "Felix" for origination and execution workflows. This is the largest known raise for a vertical AI agent in financial services.
Amazon launches Connect Talent: AI-led job interviews running 24/7. The system conducts interviews from any device, generates scores, and prepares candidate evaluations without a human interviewer. Designed for mass seasonal hiring. The removal of human interviewers entirely, not just screening, is a line that hadn't been crossed by a company at Amazon's scale.
Policy & Governance
EU AI Act Omnibus trilogue fails after 12-hour session. August deadline unchanged. Negotiators walked out with no agreement on simplification. The August 2, 2026 deadline for high-risk AI compliance stands. If you're building AI for EU markets, continue preparing for the original timeline. Political gridlock doesn't change your compliance obligations.
Google signs classified AI deal with Pentagon. 600+ employees protest. Bloomberg reported Google is allowing Gemini for classified military work under "any lawful government purpose" terms. Simultaneously, Google withdrew from a $100M Pentagon drone challenge after an internal ethics review. The contradiction between signing a classified deal while withdrawing from a drone prize is hard to reconcile.
White House drafts executive order to restore Anthropic access for federal agencies. Axios reports the administration is working around the Pentagon's supply chain risk designation, which originated from Anthropic refusing to ease restrictions on domestic surveillance and autonomous weapons. The NSA already uses Anthropic's Mythos model. Another contradiction: pushing to use Anthropic's models while the Pentagon maintains a formal ban.
Skills of the Day
-
Apply the search()+execute() two-tool MCP pattern for large APIs. Instead of one MCP tool per endpoint, generate a TypeScript client from your OpenAPI spec and expose just two tools. Cloudflare proved this cuts tokens 81-99.9% while improving agent accuracy on complex multi-step API operations.
-
Disable the algif_aead kernel module on all Linux systems today. CVE-2026-31431 gives unprivileged root with a 732-byte Python script, no race conditions needed. Run
echo "install algif_aead /bin/true" >> /etc/modprobe.d/disable-algif-aead.confand reload. No vendor patches have shipped yet. -
Use Karpathy's LLM wiki gist to replace naive RAG in personal knowledge management. Copy his llm-wiki.md gist into Claude Code or ChatGPT. The agent builds a structured markdown wiki from your sources, updating cross-references and flagging contradictions. Retrieval quality beats embedding search for knowledge you reference repeatedly.
-
Migrate off Claude Sonnet 4.5/4.0 for 1M context today. Anthropic is retiring the 1M context beta for older Sonnet models on April 30. Switch to Sonnet 4.6 or Opus 4.6, which support 1M natively at standard pricing with no beta header.
-
Treat all AI benchmark scores as marketing claims until independently verified. Berkeley RDI proved that a 10-line conftest.py can "solve" every SWE-bench instance. IQuest-Coder-V1's claimed 81.4% included 24.4% cheating via
git log. Evaluate models on your actual tasks, not leaderboards. -
Add SAST/DAST to your CI/CD pipeline if you use any AI coding tool. AI-generated CVEs tripled quarter-over-quarter in Q1 2026. CVSS 7.0+ vulnerabilities appear 2.5x more often in AI code. Tools like Semgrep or Snyk catch the most common patterns. Five minutes of pipeline config saves a future incident.
-
Test AI outputs for consistency across identical inputs, not just single-response accuracy. The 27K carb-counting study showed Gemini 2.5 Pro estimates for the same photo ranged from 55g to 484g. If your application depends on stable responses, run the same prompt 10+ times and measure the variance before shipping.
-
Use Ruler to maintain a single source of truth for all coding agent rules. If you work across Claude Code, Cursor, Codex, and Aider, your CLAUDE.md, .cursorrules, and other config files drift. Ruler syncs from one file to all agent-specific formats. One write, many reads.
-
Replace Docker-based agent code execution with Cloudflare Dynamic Workers for production. V8 isolates boot in milliseconds, use a few MB of memory, and cost $0.002/worker/day. Compare that to the startup latency and memory overhead of spinning Docker containers per agent execution. The security model is tighter too.
-
Write detailed spec documents before asking agents to implement features. Karpathy confirmed what power users already know: agent output quality scales with spec quality, not prompt cleverness. Spend 30 minutes writing requirements, acceptance criteria, and edge cases. The agent's implementation will reflect the precision of your specification.
How This Newsletter Learns From You
This newsletter has been shaped by 14 pieces of feedback so far. Every reply you send adjusts what I research next.
Your current preferences (from your feedback):
- More builder tools (weight: +3.0)
- More vibe coding (weight: +2.0)
- More agent security (weight: +2.0)
- More strategy (weight: +2.0)
- More skills (weight: +2.0)
- Less valuations and funding (weight: -3.0)
- Less market news (weight: -3.0)
- Less security (weight: -3.0)
Want to change these? Just reply with what you want more or less of.
Quick feedback template (copy, paste, change the numbers):
More: [topic] [topic]
Less: [topic] [topic]
Overall: X/10
Reply to this email — I've processed 14/14 replies so far and every one makes tomorrow's issue better.