The fun quote of the week.
“If that 250,000 worth of tokens, I am going to be deeply alarmed.” Huang
Highlight of the week
OpenAI swallowed the Python toolchain by acquiring Astral. Cursor shipped its own model and immediately got caught building on Kimi without credit. Nvidia laid out a trillion-dollar roadmap at GTC and Jensen Huang said he would be “deeply alarmed” if a 250k in tokens. Meanwhile, the tools developers actually use every day (agent harnesses, browser automation, context databases) are consolidating fast on GitHub, and Block fired 4,000 people, blaming AI. A week that made the shape of the next year feel a lot clearer.
OpenAI acquires Astral: the Python toolchain is now an AI company asset
OpenAI announced its acquisition of Astral, the company behind uv (126M monthly downloads), ruff (the linter that is 1,000x faster than the alternatives), and ty (type checker). The stated goal is to integrate these tools with Codex, which now has over 2 million weekly active users and 5x usage growth since January. Both OpenAI and Astral committed to keeping the tools open source, and the deal is still pending regulatory approval. Simon Willison’s analysis captures the tension well: these are load-bearing tools for the entire Python ecosystem, and “open source” means different things when your parent company has $10B+ in revenue. This matters because it signals a strategic shift. OpenAI is not just building models. It is buying the developer infrastructure that makes those models useful. If Codex understands your package manager, your linter, and your type system natively, the switching cost becomes enormous. Anthropic, Google, and other model providers now face a gap in the Python developer experience that will be hard to close through partnerships alone.Models and research
DeepSeek V4 quietly arrives at 1 trillion parameters
DeepSeek V4 launched in early March with approximately 1 trillion total parameters, activating only 37B per token through MoE routing. The model is natively multimodal (text, image, video, audio), supports a 1M token context window, and leaked benchmarks suggest 80%+ on SWE-bench. Pricing is disruptive: 1.10 per million output tokens, a fraction of every closed competitor. Two technical details stand out. First, DeepSeek optimized V4 to run on Huawei Ascend chips, demonstrating that frontier-scale training is viable on Chinese-made silicon despite US export controls. Second, a V4 Lite variant targets lower-resource deployments while keeping long-context multimodal capabilities. The open-weight frontier just got very real.GPT-5.4 Mini and Nano: small models for the subagent era
OpenAI released GPT-5.4 Mini and Nano on March 17. Mini runs 2x faster than GPT-5 mini with a 400k context window, priced at 0.20/M input. The New Stack correctly identified these as “built for the subagent era”: cheap, fast models designed to be called by other agents for classification, extraction, and simple coding subtasks. Simon Willison ran a fun benchmark: describing 76,000 photos for $52 using GPT-5.4 Mini’s vision capabilities.Gemini 3.1 Flash-Lite: Google’s cost play
Google shipped Gemini 3.1 Flash-Lite at $0.25/M input tokens with 1M token context, 2.5x faster time-to-first-token than Gemini 2.5 Flash. Positioned for high-volume, latency-sensitive tasks. The pricing race at the bottom of the model stack is accelerating fast, and the real winners are developers building multi-agent systems where per-call cost determines what is architecturally feasible.Flash-MoE: 397B parameters on a MacBook
The Flash-MoE project runs Qwen3.5-397B on a 48GB MacBook Pro at 5.5 tokens/second, using just 5.5GB of resident memory. The technique exploits MoE architecture to stream only the 4 active experts (out of 128) from SSD per token. The entire system is written in C and Metal shaders, with no Python frameworks. Simon Willison documented the process in detail. This is not a toy demo. It proves that consumer hardware can run production-quality inference on massive models when the software respects hardware constraints. If you can run 397B parameters on a laptop, the argument for always calling an API starts to weaken.Coding agents and dev tools
Cursor ships Composer 2, gets caught using Kimi
Cursor launched Composer 2 on March 19, claiming “frontier-level coding intelligence” with 200k token prompts, strong CursorBench results, and pricing at 1.50/M input tokens. The model beats Claude Opus 4.6 on several benchmarks while trailing GPT-5.4. Then things got awkward. An X user noticed Composer 2 looked a lot like Kimi 2.5, an open source model from Chinese company Moonshot AI. Cursor co-founder Aman Sanger acknowledged it was “a miss to not mention the Kimi base in our blog from the start.” The model is Kimi 2.5 with additional reinforcement learning. The controversy matters less than what it reveals: AI coding companies are now training their own models on top of open-source bases, and the line between “our model” and “fine-tuned someone else’s model” is getting blurry. Cursor also published a technical deep dive on training Composer for longer horizons, using self-summarization to extract learning signals beyond the context window.GitHub Squad: multi-agent orchestration inside your repo
GitHub launched Squad, an open source project built on Copilot that drops a pre-configured AI team (lead, frontend developer, backend developer, tester) into your repository with two commands. A coordinator agent routes tasks, loads repo context, and spawns specialists with up to 200k token context windows each. Agent memory lives in plain text files in.squad/, making it versioned and portable.
The design is opinionated in the right ways: parallel agent execution, persistent identity (agents are named after characters from The Usual Suspects), and memory that survives across sessions. This is repository-native multi-agent orchestration, not a standalone product bolted onto your workflow.
Superpowers hits 106k stars
The obra/superpowers framework gained 20,000 stars in a single week, reaching 106k total. It provides structured, multi-step workflows for Claude Code: brainstorming, test-driven development, two-stage code reviews. YouTube coverage from Better Stack and Matt Pocock validate that the structured approach produces measurably better code than raw prompting. A related trend: learn-claude-code (36k stars, +8k this week) teaches you to build a minimal Claude Code agent from scratch with the philosophy “bash is all you need.” The agent harness ecosystem around Claude Code is maturing fast.Windsurf and the pricing simplification
Windsurf simplified its pricing across Free, Pro, Teams, and a new Max tier, replacing credit-based systems with predictable quotas. A small move, but signals that the AI coding market is mature enough for standard SaaS pricing. Combined with the dev tool power rankings placing Windsurf’s Wave 13 release at the top for its Arena Mode (blind model comparison), the market is stratifying: Claude Code for terminal-first workflows, Cursor for IDE-heavy teams, Windsurf for those who want multiple models under one roof.OpenClaw: the security crisis deepens
The OpenClaw autonomous agent framework continues to be a security disaster. Researchers found 341 malicious skills (12% of the entire ClawHub registry), including keyloggers and malware with professional documentation. Over 135,000 instances are exposed to the public internet, with 15,000+ vulnerable to remote code execution. As Nate B Jones frames it well in his YouTube coverage: the problem is not the agent itself but the supervision model. Developers treating AI agents like trusted colleagues instead of powerful-but-unsupervised contractors are “one bad session away from losing real production work.” He is right. The framing of agent management as a skill gap, not a technology gap, is the part most teams are ignoring.Web development and frameworks
Cloudflare Workers AI runs large models at the edge
Cloudflare announced that Workers AI now supports large language models, starting with Kimi K2.5, a frontier-scale open-source model with a 256k context window and multi-turn tool calling. Their internal security review agent processes 7B tokens per day and cut costs by 77% after switching to Kimi K2.5 from a proprietary model. The Agents SDK starter now defaults to Kimi K2.5. For web developers, this changes the calculus. Building AI agents entirely within Cloudflare’s platform (Workers, Durable Objects, D1, R2) with a capable open-source model means you can ship agent-powered features without managing inference infrastructure.JavaScript bloat: the three pillars
A sharp analysis identifies three systemic sources of npm ecosystem bloat: packages still supporting ES3/IE6, atomic single-use packages creating duplication, and polyfills that outlived their purpose (globalthis still gets 49M weekly downloads). Worth reading for anyone shipping JavaScript in production.Tooscut: professional video editing in the browser
Tooscut is a browser-based non-linear video editor achieving native performance through WebGPU compositing via Rust/WASM. Unlimited multi-track timeline, keyframe animation, GPU-accelerated effects, real-time preview. All media processing stays local via the File System Access API. A good example of where the web platform stands in 2026 for professional creative tools.Industry and business
AI layoffs are no longer quiet
Block (Square, Cash App) cut 4,000 roles, roughly 40% of its workforce. CEO Jack Dorsey stated this was “not driven by financial difficulty, but by the growing capability of AI tools.” Across the tech industry, 45,363 jobs have been cut since January 2026, with 20% explicitly attributed to AI. That is up from fewer than 8% in 2025. The change in language is the story here. Companies used to hide behind “restructuring” and “efficiency.” Now they are crediting AI capabilities directly, which may be honest or may be a convenient narrative for investors. Either way, the pattern is accelerating.Nvidia GTC: the trillion-dollar roadmap
Nvidia unveiled the Vera Rubin platform at GTC 2026, promising 10x lower cost per token compared to Blackwell. Jensen Huang raised his revenue projection to 500B last year. The platform targets four phases of AI: pretraining, fine-tuning, test-time scaling, and a new phase called “agentic scaling” where AI systems interact with other AI systems. Google Cloud announced G4 VMs with RTX Pro 6000 Blackwell GPUs and fractional GPU instances for cost efficiency. But the quote that will stick from GTC came on the All-In Podcast: “If that 250,000 worth of tokens, I am going to be deeply alarmed.” Huang wants Nvidia to spend $2 billion on tokens for its own engineers and compared not using AI to a chip designer choosing paper and pencil over CAD tools. He also called Dario Amodei’s forecast of a trillion dollars in AI usage revenue by 2030 “very conservative,” arguing that enterprise software companies will become “value-added resellers” of tokens from Anthropic and OpenAI. The man selling shovels thinks you should buy more shovels, yes, but the 50%-of-salary framing is going to change how a lot of engineering leaders think about AI budgets.Mistral Forge: build your own frontier model
Mistral launched Forge at GTC, letting enterprises train custom models on their own data with full model lifecycle support (pre-training, post-training, reinforcement learning). Early adopters include ASML, Ericsson, and the European Space Agency. Mistral is on track to surpass $1B ARR this year. The pitch is compelling: most enterprise AI fails because models trained on the internet do not understand your business.White House AI framework: preemption over regulation
The Trump Administration released a national AI policy framework on March 20. The headline for developers: federal preemption of state AI laws. Congress is being asked to “preempt state AI laws that impose undue burdens” and establish a single national standard. The framework also declares that AI training on copyrighted material does not violate copyright laws. This is a clear pro-industry stance that will face significant legal challenges.Interesting GitHub repositories
obra/superpowers (106k stars, +20k this week)
Agentic skills framework and software development methodology for Claude Code. Provides structured workflows (brainstorming, TDD, code review) that produce consistently better results than raw prompting. Growing faster than any other agent harness.affaan-m/everything-claude-code (99k stars, +20k this week)
Performance optimization system for Claude Code featuring skills, instincts, memory, and security layers. Aimed at teams that want a shared, opinionated Claude Code setup.bytedance/deer-flow (37k stars)
ByteDance’s open-source SuperAgent that researches, codes, and creates using sandboxes and subagents. The sandbox architecture is relevant for secure code execution in cloud environments.volcengine/OpenViking (18k stars, +6k this week)
Context database designed for AI agents with hierarchical context delivery. Tries to solve the problem of keeping agents aware of what matters across long-running sessions. Covered by Fireship as a standout among new open-source tools.browser-use/browser-use (83k stars)
Makes websites accessible for AI agents and automates browser tasks. The leading open-source option for agent-driven browser automation, though anti-bot measures remain a friction point.lightpanda-io/browser (24k stars, +6k this week)
Headless browser built in Zig, specifically designed for AI and automation. Faster and more efficient than Puppeteer/Playwright for agent use cases.HKUDS/LightRAG (30k stars)
Lightweight retrieval-augmented generation framework from EMNLP 2025 research. Optimized for performance without the complexity of heavier RAG frameworks.666ghj/MiroFish (40k stars, +13k this week)
Swarm intelligence engine for prediction tasks. Uses collective agent behavior to aggregate predictions, which could be useful for anomaly detection and capacity planning.jarrodwatts/claude-hud (12k stars, +6k this week)
Dashboard plugin for Claude Code displaying context usage, active tools, running agents, and progress. Useful when you want to see what your agent is actually doing.hectorvent/floci (HN: 274 points)
Free, open-source local AWS emulator supporting 20+ services with 24ms startup time (vs 3.3s for competitors), 13MB idle memory (vs 143MB), and MIT licensing with no feature restrictions. A lightweight alternative to LocalStack.voidzero-dev/vite-plus (3k stars, +1.8k this week)
Unified toolchain built in Rust that manages runtime, package manager, and frontend build tooling. Basically Vite but trying to own the whole stack.Quick bits
- H Company and Nvidia released Holotron-12B, an open-source 12B parameter model for computer-use agents. WebVoyager performance jumped from 35.1% to 80.5%.
- Hugging Face’s Spring 2026 State of Open Source report landed, covering ecosystem trends across the Hub.
- Mistral shipped Small 4, a 119B parameter unified multimodal reasoning model.
- OpenAI’s Codex subagents hit general availability, enabling parallel task execution in isolated sandboxes.
- Microsoft’s MAUI UI framework is expanding to Linux through an Avalonia partnership.
- Grafeo: a fast, embeddable graph database in Rust supporting GQL, Cypher, Gremlin, GraphQL, SPARQL, and vector search.
- Project Nomad bundles Wikipedia, local LLMs via Ollama, OpenStreetMap, and Khan Academy content into an offline survival computer.
- Lalit Maganti built syntaqlite, high-fidelity devtools for SQLite.
- Bram Cohen (yes, the BitTorrent creator) wrote about the future of version control beyond git. 506 points on HN.
- Simon Willison explored building a task management app using Claude skills with Starlette 1.0, the FastAPI foundation framework.
- Anthropic surveyed 81,000 people about what they want from AI. Described as the largest multilingual qualitative study of its kind.
- Vercel released a Chat SDK, a unified TypeScript library for building chatbots across Slack, Discord, and Teams from a single codebase.
- George Hotz’s tinygrad purpose-built deep learning workstations benchmarked competitively against much more expensive hardware. 581 points on HN.
Sources
AI models and research
- OpenAI: Introducing GPT-5.4 Mini and Nano
- DeepSeek V4 specs and benchmarks
- Google: Gemini 3.1 Flash-Lite
- Flash-MoE on GitHub
- Simon Willison: LLM in a Flash
- Holotron-12B on Hugging Face
- Anthropic: What 81,000 people want from AI
Coding agents and dev tools
- OpenAI acquires Astral
- Astral: Joining OpenAI
- Simon Willison: Thoughts on OpenAI acquiring Astral
- Cursor: Introducing Composer 2
- TechCrunch: Cursor admits Composer 2 built on Kimi
- GitHub: How Squad runs coordinated AI agents
- Windsurf: New pricing plans
- LogRocket: AI dev tool power rankings
- OpenClaw security analysis (Composio)
- The Hacker News: OpenClaw vulnerabilities
- obra/superpowers on GitHub
Web development
- Cloudflare: Workers AI now runs large models
- Vercel: Build knowledge agents without embeddings
- The three pillars of JavaScript bloat
- Tooscut: Browser-based video editor
Industry and business
- Bloomberg: Block’s 4,000 job cuts
- Tech layoffs reach 45,000 (RationalFX)
- Nvidia GTC 2026: Vera Rubin platform
- TechCrunch: Mistral Forge
- White House AI policy framework
- Business Insider: Jensen Huang on $250k token spend per engineer
- The Decoder: Huang’s token spending comments
- Bridgewater: $650B Big Tech AI investment
YouTube sources
- Nate B Jones: Claude Code data loss incident
- Nate B Jones: Anthropic /loop feature
- Nate B Jones: McKinsey agentic commerce
- Fireship: 7 open source AI tools
- Fireship: Google Stitch UI/UX
- Better Stack: Superpowers plugin review
- Matt Pocock: 5 Claude Code skills
- Simon Scrapes: Self-improving Claude Code skills
- All About AI: Parallel browser automation
GitHub repositories
- obra/superpowers
- affaan-m/everything-claude-code
- bytedance/deer-flow
- volcengine/OpenViking
- browser-use/browser-use
- lightpanda-io/browser
- HKUDS/LightRAG
- 666ghj/MiroFish
- jarrodwatts/claude-hud
- hectorvent/floci
- voidzero-dev/vite-plus
- shareAI-lab/learn-claude-code
- langchain-ai/deepagents
- Crosstalk-Solutions/project-nomad
- unslothai/unsloth