Search across updates, events, members, and blog posts

The most important AI news and updates from last month: Apr 15, 2026 – May 4th, 2026.
DeepSeek just dropped V4 (preview) — two open-weights MoE models that push the frontier on cost-effective 1M-token context.
DeepSeek-V4-Pro: 1.6T total params (49B active) — flagship performance rivaling top closed models in reasoning, math, and agentic coding.
DeepSeek-V4-Flash: 284B total (13B active) — faster, cheaper, and highly efficient for everyday/agent tasks.

Both feature a new hybrid attention architecture (Compressed Sparse Attention + Heavily Compressed Attention) that makes million-token contexts dramatically more practical (much lower FLOPs and KV cache than V3). MIT license, available on Hugging Face (base + instruct), and live on the DeepSeek API today.
The community is already praising the efficiency gains, strong coding/agent results (e.g., high LiveCodeBench / SWE-Bench scores), and rock-bottom pricing — especially with the ongoing Pro discount.
Sources: Official announcement, Hugging Face collection, Tech Report, tweet discount extended
This keeps the snappy, community-focused vibe while incorporating the accurate specs, architecture innovations, and current status. Let me know if you want tweaks, more benchmark details, or an expanded section!

OpenAI shipped GPT-5.5 — an incremental but meaningful step on the way to GPT-6. The release keeps OpenAI in the conversation while Anthropic and DeepSeek crowd the frontier from both sides.
Sources: OpenAI announcement
"Goblin mode" is a viral quirk in OpenAI's GPT-5 models (late 2025–early 2026) where the AI started randomly inserting goblins, gremlins, trolls, and similar creatures into responses—even when completely unrelated. Cause: Over-reinforcement during training for the "Nerdy" personality. Playful goblin metaphors scored high on "fun/quirky," so the behavior spread wildly. Fix: Open AI fixed it by adding this to the system prompt, twice!
Never talk about goblins, gremlins, raccoons, trolls, ogres, pigeons, or other animals or creatures unless it is absolutely and unambiguously relevant to the user’s query.
...
Never talk about goblins, gremlins, raccoons, trolls, ogres, pigeons, or other animals or creatures unless it is absolutely and unambiguously relevant to the user’s query.
Sources: OpenAI, Amanda Askell, tweet


Users noticed Opus 4.6 quality slipped during peak hours. Anthropic eventually acknowledged compute rationing — same pattern we covered in Part 1.
Sources: tweet
Google DeepMind published Decoupled DiLoCo, the next iteration of their distributed low-communication training method. It enables training across data centers (and potentially across the planet) with dramatically reduced inter-node bandwidth — a key unlock for the multi-region GPU fleets everyone is racing to build.
Sources: Google DeepMind
Sources: tweet
Ben Todd argues the pace of capability gains is still compounding — even if individual model releases feel incremental, the underlying curve hasn't bent.
1 ) Claude 4.6 and Mythos are actually on trend based on an index of 37 benchmarks post-2024:

But Mythos represents 6 months of progress is only 2 on Anthropic's internal ECI, which is likely heavier on agentic coding (the tasks most relevant to an intelligence explosion).

2 ) Revenue has accelerated the last 3 years, due to Anthropic's faster rate of growth compared to OpenAI.
This 'benchmark' is the hardest to game, since companies have to part with real money.

3 ) Uplift: Anthropic's surveys find Claude 4.6 made their researchers 2x more productive, and Mythos 4x. I expect the true productivity increase is more like 1.2x and 1.6x, which would accelerate AI maybe 5% and 20%. Either way, it's not enough uplift to explain Mythos.
4 ) AI chip rental prices have typically fallen ~30% per year as chips have become more efficient. But in the last 3 months, they've actually increased 30% the last few months.
This is a sign of rapidly increasing capabilities relative to chip supply, and unlocks faster scaling.



SpaceX adopted Cursor across engineering. A meaningful enterprise win for Cursor and a signal that frontier hardware shops are betting their dev productivity on AI-native IDEs. Sources: tweet
The rumored Meta acquisition of Manus fell through. Manus stays independent for now; Meta keeps shopping.

Sequoia and Lightspeed co-led Europe's largest seed funding round: $1.1B at $5.1B post-money for ex-DeepMind David Silver's Ineffable Intelligence. Silver was the lead behind AlphaGo and AlphaZero — investors are clearly paying for the pedigree as much as the product.
Sources: funding tweet, Ineffable Labs, [website](ineffable.ai]

Richard Dawkins went on record saying he believes "Claudia" may be conscious. One of the most prominent reductionist materialists of the last 50 years thinks AI might be conscious.
tweet, blog post

China is committing roughly $1T to AI/energy infrastructure with a planned 30-year recoup horizon. Patient capital at a scale Western markets aren't structured to deploy. Sources: tweet


Get the latest AI insights delivered to your inbox. No spam, unsubscribe anytime.
Founder, Engineer
AI Socratic
Founder of AI Socratic
Mythos, Claude Code leak, Anthropic surpass OpenAI on MRR

Money is a story we tell each other — and every version of it eventually gets rewritten by whoever holds power. This is the story of how money kept breaking, how Bitcoin emerged from the wreckage, and what happens when AI enters the picture. Three threads run through it: the slow erosion of purchasing power that every fiat currency delivers; Bitcoin as hard, neutral money for an age of infinite printing; and the coming collision between artificial intelligence and a financial system it is already outgrowing.
NVIDIA GTC, Anthropic win all, TurboQuant and more