Latest
All news →Kuri's 464KB browser agent beats Vercel's on token cost
Kuri is a browser automation and web crawling tool written in Zig, designed specifically for AI agents with a focus on token efficiency. It features a 464 KB binary with ~3ms cold start, and claims 16% lower workflow token cost compared to Vercel's agent-browser according to project benchmarks. It offers CDP automation, accessibility snapshots, HAR recording, a standalone fetcher mode (kuri-fetch) that doesn't require Chrome, and an interactive terminal browser.
Firefox 150: Mythos AI Caught 271 Bugs Before Ship
Anthropic's Mythos Preview AI found 271 security vulnerabilities in Firefox 150 before release, a sharp jump from the 22 bugs caught by Anthropic's Opus 4.6 in Firefox 148. Mozilla's Bobby Holley called Mythos "every bit as capable" as the world's best security researchers, while Mozilla CTO Raffi Krikorian warned that open source maintainers still lack access to such tools.
OpenAI quietly drops Euphony, a debugging lens for Codex sessions
Euphony visualizes chat data and Codex session logs, giving developers a free way to debug AI interactions instead of guessing what went wrong.
Hydra: Swap AI coding CLIs mid-session when rate limits bite
Hydra is a unified wrapper for AI coding CLIs that switches between providers like Claude Code, Codex, OpenCode, and Pi when hitting rate limits. It automatically manages context transfer through clipboard, letting developers maintain workflow without manual context copying or re-explanation.
Anthropic's Mythos Model Leaked to Unauthorized Users on Launch Day
A private online forum group gained unauthorized access to Anthropic's Mythos model the same day it was announced for limited testing. The group has used the model regularly since, corroborated by screenshots and a live demo.
Claude Code pulled from Pro, now costs individuals 5x more
Anthropic has removed Claude Code from its $20/month Pro subscription, calling it a 'test' despite updating its pricing page. Journalist Ed Zitron spotted the change on April 21st. The coding assistant remains available on Team plans, which require a minimum of 5 seats at $20 each, making the cheapest path $100/month. Users have also reported severe throttling issues with Claude Opus recently.
CrabTrap: When your AI security guard is another AI
CrabTrap is an open-source HTTP proxy from Brex that secures AI agents in production by intercepting requests, evaluating them against policies, and allowing or blocking them in real time. It combines static rule matching with LLM judgment to make security decisions.
GitHub Copilot pauses sign-ups as agentic compute costs surge
GitHub paused new sign-ups for Copilot Pro, Pro+, and Student plans as agentic workflows push compute costs past what flat-rate pricing can sustain. Opus models drop from Pro plans, billing shifts to per-token, and usage limits now show in VS Code and CLI.
Almanac Builds Wiki Where AI Drafts, Humans Verify the Long Tail
Almanac is a wiki platform where users build knowledge bases using AI tools like Claude, ChatGPT, Cursor, and Codex. An MCP extension turns Claude Code into a research agent that drafts and submits entries. A CLI handles terminal-based contributions. Articles are attributed to contributors and opened for community edits, covering niche topics traditional encyclopedias miss.
Anthropic Quietly Drops Claude Code from $20 Pro Plan
Anthropic has removed Claude Code from its $20/month Pro plan for new signups, with support docs now referencing only the Max Plan. The company calls it 'a small test on ~2% of new prosumer signups,' but documentation changes suggest something broader. Current Pro subscribers still have access through the web app and CLI, for now.
TSRX Compiles Cleaner UI Code for Humans and AI
TSRX is a TypeScript language extension for building declarative UIs designed for what creator Dominic Gannaway calls the 'agentic era.' It keeps structure, styling, and control flow co-located and readable while remaining fully backward compatible with TypeScript. The compiler targets React, Solid, and Ripple, handling framework-specific patterns like React hooks rules automatically to improve code readability for engineers and AI systems alike.
Some secret management belongs in your HTTP proxy
AI agents given direct access to API keys create security headaches. Some models refuse requests with visible secrets, others store keys in memory across sessions. The fix is an HTTP proxy that intercepts requests and injects authentication headers, so agents never touch the actual credentials. exe.dev's Integrations feature automates this pattern, including a GitHub App for OAuth.
Mozilla Squashed 271 Firefox Bugs Using Anthropic's Mythos
Mozilla used Anthropic's Mythos Preview AI model to identify and fix 271 vulnerabilities in Firefox 150, gained through direct collaboration with Anthropic. The find shows AI can now catch bugs that previously required expensive human analysis. But the approach raises questions about access: most open source projects lack the resources and connections that made this possible.
Passive Contact Lens Treats Glaucoma Without Electronics
Researchers at the Terasaki Institute for Biomedical Innovation have developed a smart contact lens with zero electronics that uses microfluidics to monitor glaucoma and automatically deliver drugs when eye pressure climbs. A smartphone app with a convolutional neural network reads pressure levels with 94% accuracy. Testing on rabbits showed effectiveness comparable to eye drops with no biocompatibility issues over 14 days of use.
The zero-days are numbered
Mozilla used Anthropic's Claude Mythos Preview to find 271 security vulnerabilities in Firefox 150. Opus 4.6 previously caught 22 bugs in Firefox 148. The 12x increase in detection raises questions about whether AI is shifting the advantage from attackers to defenders.
Taskd lets Claude manage its own task queue
A task management system built by Levi Durfee that integrates with Claude AI via an MCP server. Written in Go and TypeScript, deployed on Google Cloud Run with Cloud SQL and envelope encryption via GCP KMS. Claude can autonomously pick up tasks, create plans, leave comments, change status, and even add suggestions to the task list.
AI Was Ruining My Philosophy Class. So We Wrote One Essay Together
When AI made traditional philosophy essays unreliable, a University of Chicago professor tried something unusual: writing one with his entire class. The collaborative experiment worked. Students said they worked harder, learned more, and were finally doing real philosophy instead of pretending for a grade.
AI's Code Overload: When Your Brain Becomes the Bottleneck
Dave Rupert identifies a growing problem among developers using AI coding tools: cognitive overload from code that outpaces human understanding. Drawing on manufacturing principles from Goldratt's 'The Goal,' he argues AI-generated code creates excess 'inventory' that the 40-watt human brain struggles to process, leading to what researchers call 'cognitive debt.'
Anthropic hikes Claude Code to $100/month as quality drops
Anthropic removed Claude Code from the $20/month Pro plan and now requires a $100/seat/month Team Premium seat. The change coincides with documented quality regression tied to a February update, with AMD engineer Stella Laurenzo's analysis of over 6,800 sessions showing the assistant began ignoring instructions and hallucinating fixes. Users on Hacker News expressed frustration, with some considering competitors like GLM and Kimi or exploring local models.
Zindex Wants to Be the Database for Agent Diagrams
Zindex is a diagram infrastructure platform designed for agents and agentic systems, featuring the Diagram Scene Protocol (DSP) that enables agents to create, edit, validate, and render diagrams as durable state. Key features include semantic descriptions (not geometric), built-in Sugiyama-style hierarchical layout pipeline, incremental editing with stable IDs, multiple render targets (SVG, PNG with 4 themes), deterministic execution, 40+ validation rules, and PostgreSQL storage. It serves as the middle layer between agent reasoning and visual output.
GoModel Shrinks LiteLLM's Footprint 44x with a Go Rewrite
GoModel is an open-source AI gateway that puts 10+ LLM providers behind one OpenAI-compatible API. Written in Go, it claims to be 44x lighter than Python-based LiteLLM while handling caching, cost tracking, and multi-provider routing in a single small container.
$5B In, $100B Out: Anthropic's Decade-Long AWS Gamble
Anthropic announced a fresh $5 billion investment from Amazon, bringing Amazon's total investment to $13 billion. In exchange, Anthropic committed to spending over $100 billion on AWS over the next 10 years to train and run its Claude model. The deal centers on Amazon's custom AI accelerator chips (Trainium2 through Trainium4), with Anthropic securing options for future chip capacity. The commitment includes 5 GW of compute, roughly the power consumption of Houston.
Meta logs workers' keystrokes and clicks to train AI agents
Meta is installing tracking software on US employees' computers to capture mouse movements, clicks, and keystrokes for AI model training. The data will help build agents that can perform computer tasks autonomously, learning how humans interact with dropdown menus, keyboard shortcuts, and other interface elements. Spokesperson Andy Stone says the data won't be used for performance reviews.
The blurry JPEG gets a name: expansion artifacts
An opinion piece introducing 'expansion artifacts,' the term for hallucinations, style issues, and strange outputs that appear when LLMs generate content. Unlike compression artifacts, these occur during decompression, when models extrapolate from compressed training data. The article examines examples in text, code, images, and video, and warns of risks when AI-generated content feeds into new AI generations, creating feedback loops that flatten and worsen information quality.
Meta's Plan to Turn Employees Into AI Training Data
According to reports, Meta plans to capture employee mouse movements, keystrokes, and screen activity to train AI agents. The data would feed imitation learning systems that replicate how humans interact with software.
Amazon's AI tools are duplicating faster than anyone can clean up
Internal documents reveal Amazon's generative AI adoption has led to duplicate internal tools and data governance issues across its retail division. AI lowers barriers to tool-building, causing teams to create overlapping systems faster than they can be consolidated. Risks include 'shadow AI' deployments, data persistence problems where AI-generated copies remain after source data is deleted, and security vulnerabilities. Amazon is exploring using AI to identify duplicates and flag risks while balancing its autonomous 'two-pizza team' culture.
gpt-image-2 drops diffusion for transformers, tops image arena
ChatGPT launches Images 2.0, an upgraded AI image generation feature with advanced text-heavy composition capabilities, expanded editing tools, high-resolution outputs up to 4K, and flexible format support for infographics, posters, comics, and social content.
Amazon Bets $25B on Anthropic in Silicon Play Against NVIDIA
Amazon will invest up to $25 billion in Anthropic, securing a commitment from the AI startup to spend $100 billion on AWS over the next decade. The real story is silicon: Anthropic is shifting toward Amazon's custom Trainium chips, taking aim at NVIDIA's grip on AI training infrastructure.
Daemons Clean Up the Mess AI Agents Leave Behind
Charlie Labs introduces Daemons, AI background processes that maintain codebases by keeping PRs mergeable, updating documentation, triaging bugs, and managing dependencies. Daemons are self-initiated, defined in simple .md files, and continuously fix drift in codebases, PRs, issues, and docs.
Musk Gets Criminal Summons as France Raids X Over Grok Deepfakes
French cybercrime investigators raided X's Paris headquarters and summoned Elon Musk to appear in April for questioning over Grok's generation of deepfake nude images (including those depicting children), antisemitic content, and child pornography. The criminal investigation also covers hate speech and fraudulent data extraction. UK and EU regulators have opened similar probes into Grok, while US state attorneys general are demanding changes to stop nonconsensual sexualized images.