Under active development Content is continuously updated and improved · Last updated Feb 18, 2026, 2:55 AM UTC

MEASURE-2.3AI system performance or assurance criteria are measured qualitatively or quantitatively and demonstrated for conditions similar to deployment setting(s). Measures are documented.

>Control Description

AI system performance or assurance criteria are measured qualitatively or quantitatively and demonstrated for conditions similar to deployment setting(s). Measures are documented.

>About

The current risk and impact environment suggests AI system performance estimates are insufficient and require a deeper understanding of deployment context of use. Computationally focused performance testing and evaluation schemes are restricted to test data sets and in silico techniques. These approaches do not directly evaluate risks and impacts in real world environments and can only predict what might create impact based on an approximation of expected AI use. To properly manage risks, more direct information is necessary to understand how and under what conditions deployed AI creates impacts, who is most likely to be impacted, and what that experience is like.

>Suggested Actions

  • Conduct regular and sustained engagement with potentially impacted communities
  • Maintain a demographically diverse and multidisciplinary and collaborative internal team
  • Regularly test and evaluate systems in non-optimized conditions, and in collaboration with AI actors in user interaction and user experience (UI/UX) roles.
  • Evaluate feedback from stakeholder engagement activities, in collaboration with human factors and socio-technical experts.
  • Collaborate with socio-technical, human factors, and UI/UX experts to identify notable characteristics in context of use that can be translated into system testing scenarios.
  • Measure AI systems prior to deployment in conditions similar to expected scenarios.
  • Measure and document performance criteria such as validity (false positive rate, false negative rate, etc.) and efficiency (training times, prediction latency, etc.) related to ground truth within the deployment context of use.
  • Measure assurance criteria such as AI actor competency and experience.
  • Document differences between measurement setting and the deployment environment(s).

>Documentation Guidance

Organizations can document the following

  • What experiments were initially run on this dataset? To what extent have experiments on the AI system been documented?
  • To what extent does the system/entity consistently measure progress towards stated goals and objectives?
  • How will the appropriate performance metrics, such as accuracy, of the AI be monitored after the AI is deployed? How much distributional shift or model drift from baseline performance is acceptable?
  • As time passes and conditions change, is the training data still representative of the operational environment?
  • What testing, if any, has the entity conducted on theAI system to identify errors and limitations (i.e.adversarial or stress testing)?

AI Transparency Resources

  • Artificial Intelligence Ethics Framework For The Intelligence Community.
  • WEF Companion to the Model AI Governance Framework- WEF - Companion to the Model AI Governance Framework, 2020.
  • Datasheets for Datasets.

>References

Trevor Hastie, Robert Tibshirani, and Jerome Friedman. The Elements of Statistical Learning: Data Mining, Inference, and Prediction. 2nd ed. Springer-Verlag, 2009.

Jessica Zosa Forde, A. Feder Cooper, Kweku Kwegyir-Aggrey, Chris De Sa, and Michael Littman. "Model Selection's Disparate Impact in Real-World Deep Learning Applications." arXiv preprint, submitted September 7, 2021.

Inioluwa Deborah Raji, I. Elizabeth Kumar, Aaron Horowitz, and Andrew Selbst. “The Fallacy of AI Functionality.” FAccT '22: 2022 ACM Conference on Fairness, Accountability, and Transparency, June 2022, 959–72.

Amandalynne Paullada, Inioluwa Deborah Raji, Emily M. Bender, Emily Denton, and Alex Hanna. “Data and Its (Dis)Contents: A Survey of Dataset Development and Use in Machine Learning Research.” Patterns 2, no. 11 (2021): 100336.

Christopher M. Bishop. Pattern Recognition and Machine Learning. New York: Springer, 2006.

Md Johirul Islam, Giang Nguyen, Rangeet Pan, and Hridesh Rajan. "A Comprehensive Study on Deep Learning Bug Characteristics." arXiv preprint, submitted June 3, 2019.

Swaroop Mishra, Anjana Arunkumar, Bhavdeep Sachdeva, Chris Bryan, and Chitta Baral. "DQI: Measuring Data Quality in NLP." arXiv preprint, submitted May 2, 2020.

Doug Wielenga. "Paper 073-2007: Identifying and Overcoming Common Data Mining Mistakes." SAS Global Forum 2007: Data Mining and Predictive Modeling, SAS Institute, 2007.

Software Resources

  • Drifter library (performance assessment)
  • Manifold library (performance assessment)
  • MLextend library (performance assessment)
  • PiML library (explainable models, performance assessment)
  • SALib library (performance assessment)
  • What-If Tool (performance assessment)

>AI Actors

TEVV
AI Deployment

>Topics

TEVV
Impact Assessment

>Cross-Framework Mappings

Ask AI

Configure your API key to use AI features.