AI Intel: Claude Tops App Store as #QuitGPT Hits 2.5M, DeepSeek V4 MIA, OLMo Hybrid Drops

Anthropic's Claude is now the #1 free app on the iPhone App Store in five countries — the US, Canada, Germany, Ireland, and Luxembourg. A month ago it wasn't even in the top 40. The catalyst: OpenAI's Pentagon deal, a consumer boycott that won't quit, and Anthropic's refusal to play ball with military surveillance. This is the biggest shift in consumer AI loyalty we've seen since ChatGPT launched.

The Great Migration: Claude Hits #1 as ChatGPT Users Jump Ship

The numbers tell the story. According to Appfigures data, Claude's daily US downloads surpassed ChatGPT's for the first time ever last Saturday. Anthropic confirmed that "every single day last week was an all-time record for Claude sign-ups." That's not a blip — that's a trend with legs.

The trigger was OpenAI closing a deal with the US Department of War (formerly Defense) while Anthropic publicly refused to participate unless the government could guarantee the tech wouldn't be used for surveillance or killing. The DoW said no. Anthropic walked. OpenAI stepped in. And then the internet did what the internet does.

The #QuitGPT movement, which started on Reddit and spread to Instagram (the @quitGPT account gained 10,000 followers in days), has now reached an estimated 2.5 million participants. The Guardian ran an op-ed from historian Rutger Bregman calling it one of the most significant consumer boycotts in recent history. Euronews, Axios, Business Insider, and The Hill all covered Claude's rise to #1.

For developers, this matters beyond the headlines. More Claude users means more demand for Claude API access. Anthropic's infrastructure was already strained — they had a notable outage on March 3 that sparked a whole Reddit thread about AI dependency. If you're building on Claude, having a fallback routing strategy isn't paranoia, it's engineering. API gateways like KissAPI that can route between Claude and other models become less of a nice-to-have and more of a reliability requirement.

DeepSeek V4: The Most Anticipated No-Show of 2026

DeepSeek V4 was supposed to drop this week. The Financial Times reported on February 28 that the release was "imminent," timed to coincide with China's Two Sessions parliamentary meetings starting March 4. Reddit's r/DeepSeek compiled every leak: Tuesday release pattern, FT confirmation, political timing. Community consensus pointed to March 3-4.

It's now March 7. No V4.

The silence is unusual for DeepSeek, which has historically been punctual with its release cadence. The company hasn't commented on the delay. Speculation on r/LocalLLaMA ranges from last-minute benchmark tuning to strategic timing against GPT-5.4's launch buzz. Some users think DeepSeek is waiting for the Two Sessions to conclude so the model gets its own news cycle rather than competing with political coverage.

What we do know from leaks: V4 is expected to be a significant jump in reasoning capability, potentially matching or exceeding Claude Opus 4.6 on coding benchmarks. If the pricing follows DeepSeek's pattern — aggressively cheap, open-weight — it could reshape the API pricing landscape overnight. For developers running multi-model stacks, V4 is worth keeping a slot open for. The moment it drops, expect every API gateway to race to add support.

GPT-5.4 Lands: 1M Context, $2.50 Input, Mixed Reviews

OpenAI shipped GPT-5.4 this week with a headline feature: 1 million token context window. That's 5x Claude's 200K limit and a genuine technical achievement. The model also includes computer use capabilities, putting it in direct competition with Claude's computer use feature.

Pricing came in at $2.50 per million input tokens and $15-20 per million output tokens — positioning it between Sonnet and Opus on cost. Cursor added support within hours. The AI gaming arena on Reddit had users pitting it against Opus 4.6 in real-time coding challenges.

Community reception has been split. The 1M context window is genuinely useful for large codebase analysis, long document processing, and RAG pipelines that previously required chunking. But early benchmarks from r/LocalLLaMA suggest the quality degrades noticeably past 500K tokens — a pattern we've seen before with extended context models. Several users reported that Opus 4.6 still produces better code on complex multi-file tasks, even with its smaller context window.

The timing is awkward for OpenAI. Launching a technically impressive model while your users are boycotting you is... a choice. The r/OpenAI subreddit has become a strange mix of "this model is actually great" posts and "I just cancelled my subscription" posts sitting side by side.

For API consumers, GPT-5.4's pricing creates interesting arbitrage opportunities. At $2.50/1M input, it's cheaper than Sonnet 4.6 ($3/1M) for input-heavy workloads. If you're processing large documents and don't need peak coding quality, the math favors GPT-5.4. Through a unified API gateway, you can route different task types to different models without changing your code.

OLMo Hybrid 7B: Open Source Gets a 2x Efficiency Boost

While the big labs fight over app store rankings and Pentagon contracts, AI2 quietly dropped something that might matter more in the long run. OLMo Hybrid 7B is a new open-source model that combines transformer attention with linear recurrent layers, achieving 2x data efficiency compared to standard transformer architectures.

Trained on 512 NVIDIA Blackwell GPUs in partnership with Lambda, OLMo Hybrid isn't trying to beat Opus on benchmarks. It's trying to prove that you can train competitive models with half the data — which means half the compute, half the cost, and half the environmental impact. The model comes with an instruct checkpoint, and a reasoning variant is coming soon.

This is the most open AI artifact released in months. Full training code, data, metrics, and checkpoints are public. For researchers and companies building custom models, OLMo Hybrid is a reference implementation that could change how the next generation of models gets built.

The developer angle: if you're running local models for privacy-sensitive workloads or edge deployment, hybrid architectures like this are the future. The 7B parameter size means it runs on consumer hardware, and the efficiency gains compound when you're paying for your own compute.

⚡ Quick Hits

  • Claude outage fallout: The March 3 Claude downtime sparked a heated r/ChatGPT thread about single-provider dependency. Top comment: "If your production app goes down because one API is down, that's not Anthropic's fault — that's your architecture." Fair point.
  • AI singularity discourse: r/singularity is having its weekly moment, this time fueled by GPT-5.4's computer use capabilities. The "we're 18 months away" crowd is loud. The "we've been 18 months away for 3 years" crowd is louder.
  • Jailbreak season: New GPT-5.4 jailbreaks are already circulating on r/ChatGPTJailbreak. OpenAI patched two within 48 hours. The cat-and-mouse game continues, but the speed of patches is notably faster than with GPT-4.

Access Claude, GPT-5.4, and More Through One API

Route between models based on task, cost, or availability. OpenAI-compatible format, pay-as-you-go pricing, no subscription lock-in.

Try KissAPI Free →

AI Intel is a daily briefing from KissAPI covering what's actually happening in AI — sourced from Reddit communities, developer forums, and industry reporting. No hype, no fluff, just signal.