Flat illustration of a cozy workshop with a cork bulletin board covered in model cards and sticky notes, warm desk lamp on dark navy background

AI weekly: Claude Opus 4.7 lands, Cloudflare floods the agent stack, Qwen3.6 runs on 23GB

X is the best way I’ve found to keep up with AI. I like tweets throughout the week, filtering for things I think are actually worth knowing. I use Claude Code to pull those likes automatically and help me turn them into this post (here’s how the pipeline works). This week: 150 tweets liked, filtered down to what’s below. Check out the previous roundup (Apr 5) if you missed it. AI for Everyone Claude Opus 4.7 Ships (12+ mentions) Anthropic launched Claude Opus 4.7 on April 16. Pricing held steady at $5 per million input tokens and $25 per million output. What changed is how the model spends those tokens: it thinks longer before answering and double-checks itself more aggressively. Vision is the part I didn’t see coming. Opus 4.7 now accepts images up to 2,576 pixels on the long edge, triple what prior Claude models took, which actually matters for diagrams, slides, and UI screenshots. If you’ve been running 4.6 for long tasks, swap the model ID and give it something harder. (source: @claudeai, @claudeai, @bcherny) ...

April 16, 2026 · 11-minute read · matt silverman  · 
Flat illustration of a workshop table with circuit boards, model cards, and a quantum chip diagram under warm Edison bulb lighting on dark navy background

AI weekly: Gemma 4 runs free on a Mac Mini, axios gets backdoored, Google moves quantum threat to 2029

X is the best way I’ve found to keep up with AI. I like tweets throughout the week, filtering for things I think are actually worth knowing. I use Claude Code to pull those likes automatically and help me turn them into this post (here’s how the pipeline works). This week: 149 tweets liked, filtered down to what’s below. Check out the previous roundup (Mar 24) if you missed it. AI for Everyone Gemma 4 Runs Locally on a Mac Mini (10+ mentions) Google released Gemma 4 under Apache 2.0. The 26B version uses mixture-of-experts, only activating 4 billion parameters at a time, so it runs on a 24GB Mac Mini. It scores 79.2% on GPQA Diamond, which puts it in the same neighborhood as Claude Sonnet 4.6’s 74.1% standard score (though Sonnet hits 89.9% with max-effort reasoning). Four sizes from edge mobile to 31B dense, 256K context, multimodal input, 140+ languages. ollama run gemma4:26b and you’re running a competitive model locally for free. I’ve been waiting for an open model this good. (source: @GoogleDeepMind, @PawelHuryn, @ollama) ...

April 5, 2026 · 7-minute read · matt silverman  · 
Isometric illustration of a cork bulletin board with pinned news clippings, calendar, and sticky notes, warm lamp on dark navy background

AI weekly: Claude gets Computer Use on Mac, /schedule runs jobs while you sleep, LiteLLM supply chain attack

X is the best way I’ve found to keep up with AI. I like tweets throughout the week, filtering for things I think are actually worth knowing. I use Claude Code to pull those likes automatically and help me turn them into this post (here’s how the pipeline works). This week: 104 tweets liked, filtered down to what’s below. Check out the previous roundup (Mar 16) if you missed it. AI for Everyone Claude Computer Use on Mac (5 mentions) Claude can now control your Mac: mouse, keyboard, screen, all of it. Kick off a task from your phone via Dispatch, walk away, come back to finished work. It’s slower than a human on purpose. Best for repetitive stuff that doesn’t have an API: copying data between apps, filling out web forms, clicking through legacy software. It tries connected apps (Slack, Calendar) first and only grabs screen control when needed. macOS now, Windows in a few weeks. (source: @claudeai, @felixrieseberg) ...

March 24, 2026 · 7-minute read · matt silverman  · 
Isometric illustration of a cozy desk with a monitor showing a 1M token progress bar, Chrome logo floating nearby, robot dog figurine on shelf, warm desk lamp on dark navy background

AI weekly: Claude 1M context goes GA, Chrome 146 unlocks browser agents, Manus Desktop launches

X is the best way I’ve found to keep up with AI. I like tweets throughout the week, filtering for things I think are actually worth knowing. I use Claude Code to pull those likes automatically and help me turn them into this post (here’s how the pipeline works). This week: ~110 tweets liked, filtered down to what’s below. Check out the previous roundup (Mar 8) if you missed it. AI for Everyone Claude Opus 4.6 Gets 1M Context Window at No Extra Cost Anthropic dropped the long-context surcharge. 1M tokens is now standard for Opus 4.6 and Sonnet 4.6, same price as any other request. Feed in an entire codebase, a stack of PDFs, hundreds of images. For Claude Code users on Max, Team, or Enterprise, it’s already the default. Benchmark that matters: 78.3% on MRCR v2 at 1M tokens, best-in-class. (source: @claudeai, @bcherny, @ClaudeCodeLog) ...

March 16, 2026 · 9-minute read · matt silverman  · 
Flat illustration of colorful speech bubbles pouring into a funnel with a clean document page emerging at the bottom, warm amber and teal on charcoal background

How I summarize 300 AI tweets per week

I publish ai-weekly every week or two. Each edition starts with about 300 liked tweets and filters down to the 8-12 stories worth reading. People keep asking how, so here’s the whole thing. ...

March 11, 2026 · 4-minute read · matt silverman  · 
Isometric illustration of a cozy desk with a monitor showing AI autonomously moving the cursor, golden sparkle trail from mouse to screen, warm desk lamp lighting on dark navy background

AI weekly: GPT-5.4 launches with computer use, Claude cheats a benchmark, Karpathy ships autoresearch

X is the best way I’ve found to keep up with AI. I like tweets throughout the week, filtering for things I think are actually worth knowing. I use Claude Code to pull those likes automatically and help me turn them into this post. This week: 298 tweets liked, filtered down to what’s below. Check out the previous roundup (Feb 26) if you missed it. AI for Everyone GPT-5.4 Is Here, and It Can Control Your Computer OpenAI launched GPT-5.4 this week. Codex-level coding merged with GPT-5.2-level reasoning, native computer use, and a 1M-token context window, all in one model. You can steer it mid-response. It’s live in ChatGPT (as GPT-5.4 Thinking and GPT-5.4 Pro), in the API, and rolling into Microsoft Copilot Studio. If you want to try it for coding, download the Codex app. You used to need a dedicated Codex model, but 5.4 handles it all. You still need to select a reasoning level. ...

March 8, 2026 · 8-minute read · matt silverman  · 
Isometric illustration of a cozy desk with monitor showing a Vite build completing, desk lamp, coffee mug, and React-to-Vite transformation icons

AI weekly: Nano Banana 2, Cloudflare rebuilds Next.js, Gemini takes over Android

Here’s what caught my attention from Feb 16-26, 2026: Check out the previous roundup (Feb 16) if you missed it. AI for Everyone Google Launches Nano Banana 2: Pro-Quality Image Generation at Flash Speed Google just dropped Nano Banana 2, their new image generation model built on Gemini Flash. The headline numbers: pro-level quality running at Flash speed, which in practice means you get high-fidelity images without waiting around. It handles text rendering (historically terrible in AI image gen), 4K upscaling, aspect ratio control, and subject consistency across multiple generations. Demis Hassabis says it taps into Gemini’s world understanding and real-time search to produce more accurate results, which tracks with what we’re seeing in the outputs. The rollout is wide: Gemini App, AI Studio, the Gemini API (listed as “Gemini 3.1 Flash Image”), Google Search, Flow, Google Ads, and Vertex AI. OpenRouter already has it. Logan Kilpatrick also announced new lower-cost resolutions and an Image Search tool. If you’ve been using DALL-E or Midjourney by default, this is worth trying today. (source: @GoogleDeepMind, @demishassabis, @OfficialLoganK) ...

February 26, 2026 · 9-minute read · matt silverman  · 
Laptop on a desk showing a GitHub repository with rising star count, phone with AI chat, and a miniature claw machine toy, warm desk lamp lighting

AI weekly: OpenAI hires OpenClaw founder, Qwen 3.5 and MiniMax M2.5 launch, WebMCP, and I built an app at the Claude Code hackathon

Personal News I got accepted into the Claude Code hackathon last week. 13,000 people applied, 500 got in. The hackathon just wrapped up today and I built Robo.app. Check it out, and watch the 3 min pitch video! Here’s what caught my attention from Feb 10-16, 2026: 🚀 OpenAI Hires OpenClaw Founder for “Personal Agents” Peter Steinberger joining OpenAI to build next-gen personal agents - Sam Altman personally welcomed him. OpenClaw hit 100K GitHub stars in 2 days (fastest any repo has ever done that), now at 197K+. Both Meta and OpenAI made offers. The fact they’re building a dedicated personal agents team, not folding him into ChatGPT, signals a new product line. OpenClaw will transition to an independent open-source foundation that OpenAI will continue to support. ( @sama) 🏋️ Three Major Model Launches in One Week Qwen 3.5-397B released - 397B total params, 17B active via hybrid linear attention + MoE. 87.8 MMLU-Pro, 83.6 LiveCodeBench v6, 76.4 SWE-bench Verified. Decoding 8.6x-19x faster than Qwen3-Max. Apache 2.0, 201 languages. Already on OpenRouter. ( @Ali_TongyiLab) MiniMax M2.5: Sonnet-level coding at 1/20th Opus cost - 230B MoE, 10B active. 80.2% SWE-Bench Verified, $0.30/M input tokens. OpenHands confirmed it’s the first open model to beat Sonnet on their benchmark. Stock jumped 15.7%. If you’re running coding agents, test this as your workhorse model. ( VentureBeat) ByteDance Seed 2.0 / Doubao 2.0 - Claims GPT-5.2 and Gemini 3 Pro level at ~1/10th cost. Doubao app has 155M weekly active users. China’s frontier labs aren’t 6 months behind anymore. ( @QuanquanGu) 🌐 WebMCP: Every Website Becomes an API for Agents Google and Microsoft co-authored a W3C spec called WebMCP - Websites can now declare capabilities as structured tools via navigator.modelContext. Two modes: declarative (HTML forms self-describe) and imperative (JS exposes custom tools). 89% token efficiency improvement over screenshot-based agents. Available in Chrome 146 Canary behind a flag. ( @aakashgupta) Pietro Schirano built a DoorDash agent demo - Adds items to cart and checks out without ever parsing UI. This is what “agents that talk to apps” looks like. ( @skirano) Links: Chrome blog | W3C spec draft ...

February 16, 2026 · 4-minute read · matt silverman  · 
A wallet and phone left unattended in a shopping cart in a sunlit parking lot

How to actually protect yourself from fraud in 2026

I am increasingly getting firsthand stories of fraud hitting people of all ages, tech-savvy or not. The FBI’s IC3 2024 report logged 859K complaints with $16B in losses. That number is up 33% from 2023. Here is what is actually happening and what to do about it. Threat vectors (ranked by prevalence) Phishing and smishing Still the number one attack vector by volume. Fake texts from a bank, Robinhood, Coinbase or “USPS” with seemingly urgent links. The Verizon DBIR consistently finds phishing in over 30% of breaches. Do not click links or call numbers from unsolicited messages. Go directly to the website or app instead. Hardware security keys are the strongest defense since phishing sites cannot intercept them. ...

February 14, 2026 · 5-minute read · matt silverman  · 
Illustration of an Amazon box with receipts and shopping bags floating toward a laptop showing a spreadsheet

Own Your Data: Amazon Purchase History

As a heavy daily AI user, one thing is painfully clear: AI hits a ceiling fast when it doesn’t have the right context. This is especially true as we move into autonomous AI agents that can solve problems before you even ask. I’m working on this through an autonomous OpenClaw agent that runs on a server in my basement (more on that soon), but even for AI beginners, I wanted to illustrate this with a practical example. ...