Close Menu
Crypto Breaking News
    Crypto Breaking News
    • News
      • Press Release
      • Featured
      • Events
      • Exchanges
      • Bitcoin
      • Ethereum
      • Solana
      • Cardano
      • Ripple
      • Press Releases by PR Newswire
      • News by CoinPedia
      • News by Coincu
      • News by Blockchain Wire
      • Binance News
    • Crypto
      • Companies
      • Events
      • Partners
      • Buy Crypto
      • Timers
    • Advertise
      • Submit a Press Release
      • Logos
      • About
      • Services
    • Offers
      • Marketing Services
      • Wallets & Tools
    • Account
    • Video
    • Contact
    Submit PR
    Crypto Breaking News
    Crypto News

    US military used Anthropic for Iran strike despite Trump’s ban: WSJ

    28 seconds ago
    FacebookTwitterLinkedInCopy Link
    News Feed
    Google NewsRSS
    Us Military Used Anthropic For Iran Strike Despite Trump's Ban: Wsj
    Us Military Used Anthropic For Iran Strike Despite Trump's Ban: Wsj

    The US military reportedly relied on Anthropic’s Claude AI during a major air strike in Iran, a development that surfaced just hours after President Donald Trump ordered federal agencies to halt use of the model. Commands in the region, including CENTCOM, reportedly used Claude to support intelligence analysis, target vetting, and battlefield simulations. The episode highlights how deeply AI tooling has been woven into defense operations even as policymakers push to cut ties with certain vendors. The episode underscores a tension between executive directives and on-the-ground automation that could influence procurement and risk management across defense programs.

    Key takeaways

      <li Claude AI was reportedly deployed for intelligence analysis, target vetting, and battlefield simulations in connection with a major air strike, hours after a White House directive to pause use of the system.

      <li Anthropic had previously secured a multiyear Pentagon contract worth up to $200 million, with collaborations involving Palantir and Amazon Web Services to enable classified workflows for Claude.

      <li The Trump administration instructed agencies to stop working with Anthropic and directed the Defense Department to treat the company as a potential security risk after contract talks broke down over unrestricted military use.

      <li The Pentagon began identifying replacement providers and moved to deploy other AI models on classified networks, including a collaboration with OpenAI for such deployments.

      <li Anthropic CEO Dario Amodei publicly pushed back against the ban, arguing that certain military applications cross ethical boundaries and should remain under human oversight rather than automated, mass surveillance or autonomous weaponization.

    Sentiment: Neutral

    Market context: The episode sits at the intersection of defense procurement, AI ethics, and national-security risk management as agencies reassess vendor dependencies and the classification of AI tools for sensitive operations.

    Why it matters

    The incident offers a rare glimpse into how commercial AI models are integrated into high-stakes military workflows. Claude, originally designed for broad cognitive tasks, reportedly supported intelligence analysis and the modeling of battlefield scenarios, suggesting a level of operational trust that extends beyond lab environments into real-world missions. This raises important questions about the reliability, auditing, and controllability of AI in combat planning, especially when government policy signals shift rapidly around vendor usage.

    At the policy level, the friction between a contracting relationship and a presidential directive highlights a broader debate about how AI vendors should be treated in secure environments. Anthropic’s refusal to grant unrestricted military use aligns with its stated ethical boundaries, signaling that private-sector providers may increasingly push back against configurations they deem ethically problematic. The Pentagon’s response—turning to alternative suppliers for classified workloads—illustrates how defense departments may diversify AI ecosystems to reduce risk exposure, while maintaining capability in sensitive operations.

    The tension also touches on the competitive dynamics of the AI-as-a-service market. With OpenAI reportedly stepping in to provide models for classified networks, the sector is likely to witness continued experimentation and renegotiation of terms around security classifications, data governance, and supply-chain risk. The situation underscores the need for rigorous governance frameworks that can adapt to rapid technological change without compromising operational security or ethical standards.

    What to watch next

    • Regulatory and policy updates from the Defense Department and the White House regarding AI vendor usage and security classifications.
    • Any new procurement or partnerships that extend AI capabilities for classified missions, including potential agreements with alternative providers to replace or supplement Anthropic’s offerings.
    • Public statements from Anthropic and OpenAI about the nature of deployments on secured networks and any new restrictions or guardrails.
    • Further details on the outcome of the earlier unrestricted-use negotiations and how that will shape future defense contracting with AI vendors.

    Sources & verification

    • Reports about Claude’s use in a Middle East operation and the administration’s halt order, including evidence discussed with sources familiar with the matter.
    • Background on Anthropic’s Pentagon contract, including the multiyear arrangement worth up to $200 million and partnerships with Palantir and AWS for classified workflows.
    • Statements from Anthropic’s leadership and public comments on military use and ethical boundaries, including interviews and official responses to regulatory actions.
    • OpenAI’s deployment on classified networks and related discussions, including public discourse around a deal with the U.S. military and associated coverage.
    • Public discussions and social-media references connected to the OpenAI arrangement with the military, such as posts documenting industry reactions.

    Anthropic’s Claude in the crosshairs: AI, ethics and policy collide in defense operations

    Officials described Claude as playing a role in intelligence analysis and operational planning during a major air strike in Iran, a claim that illustrates how close AI tools have moved to battlefield decision-making. While the Trump administration moved to sever ties with Anthropic, the operational use of Claude reportedly persisted in certain commands, underscoring a disconnect between policy statements and day-to-day defense workflows. The practical reality is that AI-driven analyses, simulations, and risk assessments can slip into mission planning even as agencies reassess vendor risk and compliance requirements across departments.

    The Pentagon’s prior engagement with Anthropic was substantial: a multiyear contract valued at up to $200 million and a network of partnerships, including Palantir and Amazon Web Services, that enabled Claude’s use in classified information handling and intelligence processing. The arrangement highlighted a broader strategy: diversify AI capabilities across a trusted ecosystem to ensure resilience in sensitive settings. Yet when policy directions shifted, the administration moved to reframe the vendor relationship, signaling a risk-based recalibration rather than a wholesale retreat from AI-enabled defense operations.

    Behind the scenes, tensions between public policy and private sector ethics came to the fore. Defense Secretary Pete Hegseth reportedly pressed Anthropic to permit unrestricted military use of its models, a request that Anthropic’s leadership rejected as crossing ethical lines the company would not cross. The firm’s stance centers on the belief that certain uses—mass domestic surveillance and fully autonomous weapons—raise profound ethical and legal concerns, and that meaningful human oversight should survive the transition from concept to execution. This position aligns with ongoing debates about how to balance rapid AI adoption with safeguards against abuse and unintended consequences.

    For its part, the Pentagon did not stand still. Facing a potential supplier gap, it began lining up replacements and reportedly reached an agreement with OpenAI to deploy models on classified networks. The shift underscores a broader strategic move to ensure continuity of capability, even as vendors re-evaluate their terms for sensitive deployments. The contrast between Anthropic’s ethical boundaries and the department’s operational needs reveals a broader policy tension: how to harness transformative technology responsibly while preserving national security imperatives.

    Industry observers also noted the ecosystem effects of such transitions. The AI market is evolving toward more modular, security-cleared configurations that can be swapped or upgraded as policy and risk assessments shift. The OpenAI arrangement, in particular, signals continued appetite for integrating leading models into defense networks, albeit under stringent governance and oversight. While this trajectory promises enhanced capability for military analysts and planners, it also elevates scrutiny around data handling, model interpretability, and the risk of over-reliance on automated systems for critical decisions.

    Anthropic’s CEO, Dario Amodei, has argued that while AI can augment human judgment, it cannot replace it in core defense decisions. In public remarks, he reaffirmed the company’s commitment to ethical boundaries and to maintaining human control in pivotal moments. The tension between maintaining access to cutting-edge tools and upholding ethical standards is likely to shape future negotiations with federal agencies, particularly as lawmakers and regulators scrutinize AI’s role in civilian and national-security contexts.

    As the landscape evolves, the broader crypto and tech communities will be watching how these policy and procurement dynamics influence the development and deployment of advanced AI systems in high-stakes environments. The episode serves as a case study in balancing rapid technological advancement with governance, oversight, and the enduring question of where human responsibility ends and automated decision-making begins.

    Risk & affiliate notice: Crypto assets are volatile and capital is at risk. This article may contain affiliate links. Read full disclosure

    Crypto Breaking News
    • Website
    • Facebook
    • X (Twitter)
    • Pinterest
    • Instagram
    • Tumblr
    • LinkedIn

    The Crypto Breaking News editorial team curates the latest news, updates, and insights from the global cryptocurrency and blockchain industry.

    Related Posts

    Three Solana Data Points Indicate Resilience—is Sol Undervalued?

    Three Solana data points indicate resilience—is SOL undervalued?

    2 hours ago
    6 Polymarket Traders Net $1m On Us-Iran Strike, Insider Fears: Report

    6 Polymarket traders net $1M on US-Iran strike, insider fears: Report

    4 hours ago
    Bitcoin Rebounds To $68k After Death Of Iranian Supreme Leader

    Bitcoin Rebounds to $68K After Death of Iranian Supreme Leader

    6 hours ago
    Vitalik Buterin Says Ethereum Smart Accounts Are Coming Within A Year

    Vitalik Buterin Says Ethereum Smart Accounts Are Coming Within a Year

    8 hours ago
    Bitcoin Bottom Fractal Signals 130% Rally: Is Model Valid In 2026?

    Bitcoin Bottom Fractal Signals 130% Rally: Is Model Valid in 2026?

    10 hours ago
    Anthropic Ceo Responds To Pentagon Ban On Military Use

    Anthropic CEO Responds to Pentagon Ban on Military Use

    12 hours ago

    Search Crypto News

    Featured Crypto News

    Tether USDT Price Outlook 2026-2030

    Tether USDT Price Outlook 2026-2030

    27 February 2026
    Tangem Wallet Extends Promotion With Discounts And Btc Rewards

    Tangem Wallet Extends New Year Promotion: 20% Off Plus $10 in Bitcoin until 2 March!

    23 January 2026

    Latest News

    • US military used Anthropic for Iran strike despite Trump’s ban: WSJ
    • Three Solana data points indicate resilience—is SOL undervalued?
    • 6 Polymarket traders net $1M on US-Iran strike, insider fears: Report
    • Bitcoin Rebounds to $68K After Death of Iranian Supreme Leader
    • Vitalik Buterin Says Ethereum Smart Accounts Are Coming Within a Year
    • Bitcoin Bottom Fractal Signals 130% Rally: Is Model Valid in 2026?
    • Anthropic CEO Responds to Pentagon Ban on Military Use
    • ETH Drops 60% from 2025 High, Yet TradFi Bets on ETH: Here’s Why
    • Crypto Treasury Firms Likely to Consolidate in 2026, Says Exec
    • 11 US Senators Urge Federal Probe Into Binance Sanctions Compliance

    Join 17,000+ Crypto Followers

    • Facebook2.3K
    • Twitter4.3K
    • Instagram5.6K
    • LinkedIn4K
    • Telegram52
    • Threads800
    eToro Crypto 300x300
    Bitcoin Conference 2026 - Las Vegas

    About Crypto Breaking News

    About Crypto Breaking News

    Crypto Breaking News is a fast-growing digital media platform focused on the latest developments in cryptocurrency, blockchain, and Web3 technologies. Our goal is to provide fast, reliable, and insightful content that helps our readers stay ahead in the ever-evolving digital asset space.

    Web3 Digital L.L.C-FZ
    License Number: 2527596
    📞 +971 50 449 2025
    ✉️ info@cryptobreaking.com
    📍Meydan Grandstand, 6th floor, Meydan Road, Nad Al Sheba, Dubai, United Arab Emirates

    FacebookX (Twitter)InstagramPinterestYouTubeTumblrBlueskyLinkedInRedditTikTokTelegramThreadsRSS

    Links

    • Crypto News
    • Submit a Press Release
    • Advertise
    • Contact Us
    • Privacy Policy
    • Disclaimer
    • Terms and Conditions

    advertising

    AVATRADE
    © 2026 CryptoBreaking.com | All rights reserved | Powered by Web3 Digital & Osom One

    Type above and press Enter to search. Press Esc to cancel.

    Change Location
    Find awesome listings near you!