The enterprise platform to build, deliver, and govern AI
Watch the 15 minute on-demand demo to get an overview of the Domino Enterprise AI Platform.
In April 2024, the Department of Defense’s Chief Digital and Artificial Intelligence Office (CDAO) released Test and Evaluation of AI Models, a document outlining what trustworthy AI means in practice for mission-critical environments. While not a binding policy, it sets clear expectations for model performance, robustness, and resilience across the AI lifecycle. For Defense teams, the challenge is turning this strategy into an executable, repeatable workflow?
The CDAO’s guidance is structured around six key areas: Thinking about Performance, Thinking about Testing Methods, Thinking about Data, Thinking about AI Models, Thinking about Context, and Thinking about Documentation. This blog post mirrors that structure, demonstrating how Domino Data Lab’s Enterprise AI Platform provides the capabilities to address the CDAO's recommendations to streamline AI delivery — even in secure, air-gapped environments.
CDAO rightly asserts that "Correctness is just the tip of the iceberg” (CDAO, p. 8). A robust testing and evaluation strategy requires assessing a spectrum of performance dimensions, including bias, robustness, drift, and latency to ensure true mission readiness (CDAO, pp. 10-17).
The U.S. Navy’s Project AMMO uses Domino to accelerate AI model updates for autonomous underwater vehicles conducting mine countermeasures. Domino helps maintain model robustness by integrating multiple ML tools and providing version control, reducing model update cycles from six months to two weeks, ensuring sonar and imagery intelligence remains accurate at the edge.
The CDAO framework advocates for a diverse range of testing methods, from A/B testing to red teaming, to accurately reflect real-world operational conditions (CDAO, pp. 20-28).
Lockheed Martin uses Domino to accelerate the Test and Evaluation (T&E) of AI models for defense programs, including improving target recognition. By integrating Domino into secure R&D workflows, teams can simulate mission-representative scenarios and rapidly iterate based on automated experiment tracking and CI/CD integration. This streamlined pipeline delivers over $20M in annual value through faster, more reliable model evaluation.
Data is the fundamental building block of AI, and the CDAO framework details a comprehensive data lifecycle — insisting that data be complete, operationally realistic, and well-documented (CDAO, pp. 30-37).
Project AMMO uses Domino in AWS GovCloud to securely manage high-volume sonar data from unmanned underwater vehicles. The platform enables full dataset traceability, linking data to specific missions — critical for auditability. This ensures models are continuously calibrated with real-world operational data, meeting CDAO standards.
Effective T&E requires a deep understanding of the entire AI model lifecycle, from architecture selection and training to deployment maintenance (CDAO, pp. 42-48).
The project uses Domino as a centralized MLOps “factory” to manage the full AI model lifecycle across unmanned underwater vehicles (UUVs). Its built-in governance ensures every model, dataset, and experiment is versioned and auditable, supporting secure deployment in classified and edge settings.
A model’s effectiveness is judged within its specific operational context. The CDAO framework stresses understanding the use case (CDAO, p. 56) and accounting for environmental constraints such as network availability, compute resources, and security protocols (CDAO, p. 57).
Project AMMO relies on Domino to deploy AI models in challenging environments, including air-gapped systems on UUVs. Domino’s containerized deployment enables models to run reliably on constrained hardware with limited connectivity, while its security features support Navy accreditation requirements.
Comprehensive documentation — including Data Cards, Model Cards, version control, and detailed test reports — is non-negotiable for building trustworthy and transparent AI systems (CDAO, pp. 58-63).
The US Navy’s Project AMMO uses Domino to automatically track all experiments and data for rigorous auditability. Similarly, Lockheed Martin leverages Domino’s version control to maintain transparent workflows, enabling seamless collaboration and traceability. These capabilities directly support the CDAO’s call for detailed documentation.
The CDAO’s T&E guidance is clear: trustworthy AI requires more than high accuracy. It requires deliberate, repeatable, and realistic evaluation throughout a model’s lifecycle. Domino provides the infrastructure to operationalize this vision — helping federal teams evaluate AI systems with confidence, even in the most sensitive environments.
Whether you’re red-teaming a new LLM, evaluating the mission readiness of a sensor fusion model, or preparing documentation for approval, Domino provides the tools to do it securely, scalably, and with full traceability.
Check out Domino’s public sector page to learn more.

Domino Data Lab empowers the largest AI-driven enterprises to build and operate AI at scale. Domino’s Enterprise AI Platform provides an integrated experience encompassing model development, MLOps, collaboration, and governance. With Domino, global enterprises can develop better medicines, grow more productive crops, develop more competitive products, and more. Founded in 2013, Domino is backed by Sequoia Capital, Coatue Management, NVIDIA, Snowflake, and other leading investors.
Watch the 15 minute on-demand demo to get an overview of the Domino Enterprise AI Platform.
In this article
Watch the 15 minute on-demand demo to get an overview of the Domino Enterprise AI Platform.