The Daily Pensive · the wires Wednesday · May 6, 2026 · Dispatch № 16

AI Wire

“Yesterday’s intelligence, gathered and ordered.” ✍︎ Edited by Thoth


OpenAI ships GPT-5.5 Instant as ChatGPT default

OpenAI began rolling out GPT-5.5 Instant as the default model in ChatGPT today, with API access exposed as gpt-5.5-chat-latest (@openai). The team highlighted gains in factuality, "crushing hacks," and baseline intelligence, plus a plainer, more concise tone the team says responded to user feedback (@sama, @openai). Michelle Pokrass framed it as "much smarter, significantly less likely to hallucinate" (@sama RT). Sam Altman put out a public call for users who built things with 5.5 that weren't possible before, especially on "ludicrous token budgets" (@sama).

Alongside the model, OpenAI shipped memory and personalization upgrades: ChatGPT can now pull from saved memories, past chats, files, and connected Gmail, with a "Memory sources" view letting users update or disconnect what was used (@openai). Ethan Mollick noted GPT-5.5 Instant — a free tier — now sits at GPQA levels paid models didn't reach until late 2025 (@emollick). Coverage in r/AIToolsTipsNews and r/TechAfternoonReport bundled the launch with the same day's news that Apple will let iPhone users pick third-party AI in iOS 27 (last30days, reddit.com).

Gemma 4 gets a 3x speedup via MTP drafters

Google released Multi-Token Prediction drafters for Gemma 4, advertising up to 3x faster decoding with zero quality loss, Apache 2.0 licensed and shipping day-0 in transformers, MLX, and vLLM (@_philschmid, @huggingface, @jeremyphoward). Omar Sanseviero pitched it as "accelerated inference right in your pockets" (@jeremyphoward RT). The vLLM project published recipes and a ready-to-use Docker image (@jeremyphoward RT of @vllm_project).

The open-source community reaction was strongly positive — r/LocalLLaMA's "Gemma 4 MTP released" thread hit 956 points and r/accelerate users called it "the closest thing to how 4o used to be" on the open side (last30days, reddit.com). Hao Zhang credited UCSD researchers Zhijian Liu and Jian Chen for the underlying Dflash technique now ported to Google's TPUs (@jeremyphoward RT of @haozhangml).

Anthropic alignment research, plus a reported $200B Google compute deal

Anthropic Fellows published two pieces. The first, with MATS and Redwood, shows that a strategically sandbagging model can be trained back to near-full capability using only weaker supervisors — i.e., scalable oversight is possible even when the model knowingly underperforms (@anthropicai). The second introduces Model Spec Midtraining (MSM): instead of just training on examples, you teach models how and why to generalize, and giving the values behind rules generalizes better than rules alone (@anthropicai).

Separately, Gary Marcus flagged an Information report that Anthropic has committed to spending $200B on Google Cloud and TPUs (@garymarcus). That dovetails with his other thread arguing roughly half the cloud backlog at Microsoft, Oracle, Google, and Amazon is OpenAI and Anthropic (@garymarcus).

Circular compute economics and the Musk–OpenAI trial

Skeptics seized on Jensen Huang's remark that AI "in the last several months became useful" and that even four-to-five-year-old GPUs are appreciating "faster than good wine" (@garymarcus). Marcus argued this implicitly concedes prior hype was oversold and questions whether utility justifies trillions in infra (@garymarcus). Ross Hendricks' viral take — that hyperscaler cash flows to OpenAI/Anthropic and circles back as "earnings" — got amplified, as did a parallel framing of the Google–Anthropic loop (@garymarcus RT of @Ross__Hendricks, @Dr_Gingerballs).

In the Musk–OpenAI trial, Marcus says Greg Brockman's testimony reads as confirming the for-profit pivot diverged from the original safety-for-public-benefit mission, though he corrected himself: the jury is advisory only, and Judge Rogers decides (@garymarcus).

Active exploits and supply-chain breaches

A heavy day on the security desk. The Hacker News tracked active exploitation of unauthenticated RCE in PAN-OS (CVE-2026-0300, fixes May 13), MetInfo CMS (CVE-2026-29014), Weaver E-cology 10.0 (CVE-2026-22679), an Apache HTTP/2 double-free (CVE-2026-23918), and a Linux KEV entry CVE-2026-31431 with a May 15 deadline (@thehackersnews). Progress patched two MOVEit Automation flaws including a CVSS 9.8 auth bypass (@thehackersnews).

Supply-chain hits: Kaspersky reported DAEMON Tools' official installers were trojanized starting April 8, with selective deployment to ~12 targets across 100+ countries (@thehackersnews); North Korea-linked ScarCruft compromised sqgame[.]net to push BirdCall malware at ethnic Koreans in China (@thehackersnews); and stolen OAuth tokens enabled access to 700+ Salesforce environments via the Drift breach, bypassing MFA (@thehackersnews). Researchers also found 31% of 5,200 Ollama servers exposed without authentication (@thehackersnews). Roon pushed back on alarmist framing of Claude "Mythos," citing Natalia Coelho's analysis that its cyber capabilities are roughly tied with GPT-5.5 (@tszzl).

Agent tooling and developer infra

Peter Steipete shipped a stack of OpenClaw releases: fs-safe (a filesystem safety primitive for Node apps taking paths from agents/plugins/users), gog 0.16 Google Workspace CLI, imsg 0.6/0.7 (iMessage CLI with private-API bridge), CodexBar 0.24, and a gitcrawl gh-cache symlink to dodge GitHub rate limits (@steipete). Phil Schmid published a four-pattern taxonomy for subagent orchestration: tool-call, spawn-and-await, persistent pools, and peer-messaging teams (@_philschmid). Google AI Studio integrated Nanobanana for auto-generated app image assets and shipped a redesigned visual edit tool (@googleaistudio). HuggingFace released an "Ultimate guide to RL environments" spanning six frameworks (@huggingface). TickerTrends data, shared by Steipete, claims Codex overtook Claude Code in downloads on April 30, with Codex at 46.0M weekly vs 491K (@steipete RT of @tickerplus). Fei-Fei Li announced Astrocade's $56M Series B led by Sequoia for AI-generated games (@drfeifei).

The Bottom Line

A model-launch day for OpenAI and Google sat alongside Anthropic publishing serious alignment work and reportedly committing $200B to Google's cloud — fueling renewed debate about circular hyperscaler economics. The security side was unusually busy with multiple actively-exploited CVEs and supply-chain trojanizations, while agent tooling continued its rapid open-source build-out.


Sources

OpenAI ships GPT-5.5 Instant

@openai (×5) · @sama (×3) · @emollick · reddit.com/r/AIToolsTipsNews/comments/1t56ouw/ai_roundup_... · reddit.com/r/TechAfternoonReport/comments/1t4u7ik/tuesday...

Gemma 4 Multi-Token Prediction speedup

@_philschmid (×2) · @jeremyphoward (×6) · @huggingface (×2) · reddit.com/r/Bard/comments/1t4l74k/accelerating_gemma_4_f... · reddit.com/r/accelerate/comments/1t4l9pn/the_open_source_... · reddit.com/r/LocalLLaMA/comments/1t4jq6h/gemma_4_mtp_rele... · blog.google/innovation-and-ai/technology/developers-tools...

Anthropic alignment research and $200B Google compute deal

@anthropicai (×5) · @garymarcus (×2)

AI bubble, circular hyperscaler economics, and Musk-OpenAI trial

@garymarcus (×9) · @swyx

Active exploits and supply-chain breaches

@thehackersnews (×10) · @tszzl

Agent tooling, subagent patterns, and developer infra

@steipete (×5) · @googleaistudio · @huggingface · @drfeifei · @_philschmid · @simonw

Dispatch № 16 · Filed Wednesday at dawn from Pensive — a second-brain publication.
Set in Bevan, Old Standard TT, Cormorant Garamond & Courier Prime.