Ethics Fight Reshapes the AI App Rankings

Public backlash to military AI partnerships pushes users toward companies taking a harder line on surveillance and autonomous weapons.

3/2/20262 min read

Ethics Fight Reshapes the AI App Rankings

Public backlash to military AI partnerships pushes users toward companies taking a harder line on surveillance and autonomous weapons.

Moms for Ethical AI

Mar 02, 2026

Anthropic’s Claude has climbed to the top of Apple’s U.S. App Store at a moment when public opinion, industry politics, and federal pressure have collided to reshape the AI landscape. The shift is a referendum on how Americans want AI companies to handle government demands for surveillance and military use.

Claude’s sudden rise

Claude’s ascent to the #1 spot and overtaking of ChatGPT, follows a rapid surge in downloads beginning in last weekend. According to multiple reports, Claude was outside the top 100 at the end of January and then vaulted into the top 20 through February before hitting #1 on February 28 and holding the position into early March. (TechCrunch)

Anthropic confirmed that free users are up more than 60% since January, daily signups have tripled since November, and paid subscriptions have more than doubled - breaking internal records throughout the week. (Mashable)

What triggered the public shift

Several developments converged:

  • OpenAI’s Pentagon partnership
    OpenAI’s announcement that it would support the Department of Defense in deploying AI inside classified networks sparked immediate backlash and fueled the “Cancel ChatGPT” trend. (CancelChatGPT.com)

  • Anthropic’s refusal to loosen safety guardrails
    Anthropic rejected Pentagon requests to remove restrictions that prevent Claude from being used for domestic mass surveillance or autonomous weapons. CEO Dario Amodei publicly warned about the dangers of both. (Hacks Technology News)

  • Federal retaliation
    Defense Secretary Pete Hegseth designated Anthropic a national‑security “supply‑chain risk,” a move that would block the company from future government contracts. Anthropic called the designation legally unsound and vowed to challenge it.

  • Cross‑company employee revolt
    More than 700 employees from Google and OpenAI signed a joint letter urging their companies not to comply with government demands for mass surveillance or autonomous lethal systems.

  • Political escalation
    President Trump publicly attacked Anthropic, framing the company’s stance as an attempt to “strong‑arm” the Department of War. His administration then barred federal agencies from using Claude, a move that ironically amplified public interest. (Engadget)

Why consumers are responding

The App Store rankings reflect a broader public sentiment: people are increasingly sensitive to how AI companies handle government pressure. Claude’s rise appears to be driven not by a new feature release but by a perception that Anthropic is taking a principled stand on civil liberties and weapons oversight.

This aligns with reporting that the Pentagon sought to use AI systems for domestic surveillance and autonomous weapons - requests that Anthropic refused.

What this means for the future of AI

The episode signals a shift in how the public evaluates AI platforms:

  • Ethics are becoming a competitive differentiator, not just model quality. (Moms for Ethical AI)

  • Government–industry conflict is now a visible market force, shaping consumer behavior.

  • App Store rankings are emerging as a proxy for public trust, not just popularity.

  • Companies may face pressure to declare their stance on military and surveillance uses as part of their brand identity.

This moment suggests that the next phase of the AI race may hinge on governance, transparency, and how companies navigate political demands, which is good news for the population and the planet.