Back

The US DoD funds four frontier AI firms for advancing AI in defense

Deveshi Dabbawala

July 24, 2025
Table of contents

The U.S. Department of Defense (DoD) has committed $800 million to frontier AI, partnering with Anthropic, Google, OpenAI, and xAI. Each firm has been awarded an IDIQ (Indefinite Delivery, Indefinite Quantity) contract worth up to $200 million, signaling a strategic shift in how the U.S. plans to use AI in defense to maintain global military leadership.

This initiative, led by the Chief Digital and Artificial Intelligence Office (CDAO), aims to embed AI in defense across warfighting, intelligence, and enterprise functions. While the move promises a leap in military effectiveness and situational awareness, it also opens critical discussions on AI safety, governance, and ethical implications in high-stakes defense environments.

The strategic significance of frontier AI in defense

Frontier AI refers to highly capable models that exceed current benchmarks in generalization, reasoning, and autonomy. These models, trained on massive datasets, can power agentic systems, tools that reason, plan, and act with minimal human intervention.

The department intends to deeply embed AI in defense across its operations, from real-time battlefield intelligence to supply chain optimization and predictive maintenance. Expect frontier AI to support:

  • Warfighting operations by enhancing real-time awareness and tactical response.
  • Intelligence analysis through summarization, multi-modal data parsing, and pattern detection.
  • Enterprise efficiency by automating complex bureaucratic workflows.

“We are at the dawn of an era where AI can redefine strategic advantage,” said Dr. Doug Matty, Deputy Director for AI at the CDAO.

Who gets the funds for AI?

The $800 million investment is split across four major AI players:

  • Anthropic: Brings Claude’s alignment-first capabilities for safe decision-making.
  • OpenAI: Delivers GPT-4 and beyond, with scalable architecture and plugin ecosystems.
  • Google: Offers Gemini models with multi-modal reasoning for defense intelligence.
  • xAI: Elon Musk’s AI venture focused on long-context understanding and independent reasoning.

Notably absent from this funding allocation are other major frontier AI companies that have developed competitive models.

  • Meta AI: Noted for their Llama series of open-source models
  • Mistral AI: The French company with cost-effective frontier models
  • Cohere: Specializing in enterprise-focused models with privacy features

These models will be used across programs like:

  • Project Maven (computer vision for drone surveillance)
  • Advana (enterprise data analytics)
  • Edge Data Mesh (connecting AI at the tactical edge)
  • Army’s ELLM Workspace (secure language model sandbox for experimentation)

Together, they will power real-time analytics, battlefield planning, cybersecurity operations, and autonomous vehicle coordination.

The commercial-first approach and its implications

The DoD’s collaboration with commercial AI firms marks a pivotal shift: instead of building models internally, it’s leveraging private-sector speed and talent to scale AI in defense.

“The fact that frontier models are being integrated into mission-critical government systems is a clear signal to the enterprise space: these technologies are mature, resilient, and ready for real-world impact. It’s pushing the boundary of what’s possible, and raising the bar for what’s expected,” say Prashanna Rao, Head of Engineering, GoML

This approach is based on key benefits:

  • Speed to deployment: Commercial AI labs iterate faster than defense contractors.
  • Access to top talent: The AI brain trust primarily resides in Silicon Valley, not government agencies.
  • Cost efficiency: Shared R&D across markets reduces overhead.

A notable enabler is the partnership with the General Services Administration (GSA), ensuring AI procurement and tools can be shared across federal agencies, including DHS, VA, and the State Department.

However, this reliance on commercial players is not without trade-offs:

  • Speed vs. Oversight: Accelerated deployment could compromise safety checks.
  • Innovation vs. Control: The DoD may cede technical control to vendors.
  • Vendor lock-in: Heavy reliance on a few firms could limit future flexibility.

Ethical implications of frontier AI in defense

As AI becomes more powerful and autonomous, the stakes of its deployment rise dramatically. In a military context, even a small error could have catastrophic consequences. Therefore, the conversation must go beyond performance and deeply into ethics.

1. Bias and fairness

One of the most urgent concerns is the potential for algorithmic bias in frontier AI models. These models are trained on massive datasets scraped from the internet or curated by private firms, data that may contain historical, cultural, or geopolitical biases.

If these biases are not detected and mitigated, the consequences of AI in defense can be severe:

  • Target misidentification based on racial or ethnic profiling.
  • Faulty intelligence prioritization that ignores marginalized regions or communities.
  • Unjustified escalation due to misinterpretation of culturally nuanced data or behaviors.

2. Accountability and liability

As AI systems make increasingly consequential decisions, sometimes autonomously, the question of accountability becomes complex and critical in AI in defense.  

If an AI system incorrectly identifies a target, launches a response based on faulty reasoning, or contributes to a mission failure, who is to blame?

  • Is it the developer who wrote the model’s architecture?
  • The contractor who integrated it into military infrastructure?
  • The operator who followed its recommendation?
  • The command structure that approved its deployment?

3. Misuse, escalation, and global precedent

Another major ethical dilemma lies in the potential for misuse, either by adversaries, rogue actors, or even allied forces misinterpreting an AI’s output. AI in defense, especially those deployed in defense, can be:  

  • Hacked to provide false intelligence or override critical functions.
  • Manipulated through prompt injection or adversarial examples to change behavior.
  • Repurposed for use in authoritarian surveillance, misinformation campaigns, or autonomous lethal weaponry.

The use of AI in national defense also sets a precedent internationally. If democratic nations deploy autonomous systems without sufficient guardrails, it could normalize an AI arms race, where authoritarian regimes feel justified in deploying even riskier versions without transparency or ethics.

Building responsible AI governance and guardrails

The integration of AI in defense cannot proceed without robust governance frameworks. The CDAO, along with other DoD bodies, is actively working to build a policy infrastructure that ensures AI is deployed safely, responsibly, and transparently.

“What we’re seeing in defense is the early formation of scalable, responsible AI frameworks, guardrails, oversight mechanisms, alignment practices, that can and should influence how enterprises build and deploy AI. These patterns, tested under the highest stakes, offer a powerful foundation for safer, more trusted AI adoption in the commercial world.”
— Prashanna, Head of Engineering, GoML

Elements of effective AI governance

  • Policy and doctrine: Clear rules about where, when, and how AI can be used in combat and logistics.
  • Validation and testing: Rigorous red-teaming, adversarial testing, and simulation under edge cases.
  • Transparency: Ensuring explainability for AI decisions, especially those tied to lethal outcomes.
  • Oversight mechanisms: Establishing audit trails and human-in-the-loop controls.

These principles are already widely practiced in enterprise environments and should be adapted for military applications. For instance:

  • AI filters used in customer support to prevent hallucinations can be modified for autonomous targeting.
  • Human-in-the-loop review systems in financial services can guide operational oversight in combat zones.
  • Prompt auditing tools from platforms like GoML ensure prompt inputs do not trigger unsafe outputs.

How will LLM capabilities evolve in defense contexts?

As frontier AI systems become a cornerstone of national defense strategies, large language models (LLMs) will evolve and transform faster and in different ways.  

Defense use cases, such as real-time threat analysis, autonomous mission planning, and secure communications, demand a fundamentally different class of large language models (LLMs), ones that far exceed conventional benchmarks. These models must be highly robust in adversarial environments, where challenges like signal interference, misinformation campaigns, or deceptive inputs are common in AI in defense.

They need deep contextual understanding to interpret strategic nuances, geopolitical dynamics, and mission-specific instructions with precision. In addition, such models must be resilient to issues like model drift and hallucinations, with rigorous safeguards in place to ensure operational accuracy and reliability.

How will AI use in defense impact enterprises?

The cascading effects of defense-grade AI innovation will shape enterprise-grade AI in two distinct ways:

  1. Stricter expectations around safety and control: As the public and regulators observe how AI in defense is managed in defense contexts, enterprises will face growing pressure to adopt similar safeguards in civilian domains like healthcare, finance, and critical infrastructure.  
  1. Accelerated adoption of policy frameworks and tooling: Enterprise teams will increasingly look to the standards developed for national security as templates to guide their own implementations.

Responsible AI use hinges on five pillars: model alignment, human oversight, adversarial testing, explainability, and red teaming. These AI safety principles, born in the context of real-world enterprise applications, are now proving essential in high-stakes defense use cases.

Embedding AI guardrails into military and enterprise systems

The use of AI in defense, and its spillover into enterprise systems, requires more than model improvements. It demands an ecosystem of AI guardrails, safety, and governance structures. Here’s how these guardrails are being adapted for both military and enterprise domains:

Ethical alignment

Models must reflect democratic values, human rights norms, and proportional use of force in conflict scenarios. This requires codifying ethical boundaries during training and deployment phases.

Strategic stability

Especially in defense, there’s a risk of escalation if AI models are deployed without international coordination. Guardrails must account for long-term peace and deterrence goals, not just immediate tactical gains.

Explainability and auditability

Black-box behavior is unacceptable in defense AI. Enterprises have already been demanding transparent outputs, traceable reasoning chains, and the ability to audit decisions post hoc. This has been particularly evident in highly regulated industries till now. We expect AI explainability to become table stakes for all AI implementations.

Human-in-the-loop systems

Across both sectors, final decision-making is increasingly routed through human oversight, with AI serving as a co-pilot rather than a commander. This mitigates risks and maintains accountability.

Balancing innovation with caution

AI in defense pushes the boundaries of what LLMs can do, but it also reveals the dangers of unchecked AI development. As agencies and enterprises race to adopt these technologies, a balance must be struck:

  • Innovation drives competitive and tactical advantage, enabling faster, smarter, and more scalable decision-making.
  • Caution ensures that such innovation doesn’t outpace our ability to control, govern, and understand it.

Ultimately, the lessons from national security AI deployments will inform the next generation of responsible enterprise AI. By embedding safety, transparency, and ethical principles from the start, both public and private sectors can leverage LLMs for transformative outcomes without compromising control or accountability.

Conclusion

The U.S. Department of Defense’s $800 million investment into frontier AI is more than just a technology upgrade; it’s a defining moment for the future of global security. By aligning with the world’s most advanced AI labs, the DoD has taken a strategic lead in redefining how wars are fought, how intelligence is processed, and how decisions are made.

But with this power comes a responsibility to ensure AI in defense is safe, ethical, and aligned with democratic values. AI will define 21st-century power. The question is not whether we build powerful models, but whether we build them responsibly. To safeguard national security, global stability, and democratic values, we must prioritize safety, ethics, and transparency alongside innovation in AI in defense.