Under active development Content is continuously updated and improved

MEASURE-2.13Effectiveness of the employed TEVV metrics and processes in the MEASURE function are evaluated and documented.

>Control Description

Effectiveness of the employed TEVV metrics and processes in the MEASURE function are evaluated and documented.

>About

The development of metrics is a process often considered to be objective but, as a human and organization driven endeavor, can reflect implicit and systemic biases, and may inadvertently reflect factors unrelated to the target function. Measurement approaches can be oversimplified, gamed, lack critical nuance, become used and relied upon in unexpected ways, fail to account for differences in affected groups and contexts.

Revisiting the metrics chosen in Measure 2.1 through 2.12 in a process of continual improvement can help AI actors to evaluate and document metric effectiveness and make necessary course corrections.

>Suggested Actions

  • Review selected system metrics and associated TEVV processes to determine if they are able to sustain system improvements, including the identification and removal of errors.
  • Regularly evaluate system metrics for utility, and consider descriptive approaches in place of overly complex methods.
  • Review selected system metrics for acceptability within the end user and impacted community of interest.
  • Assess effectiveness of metrics for identifying and measuring risks.

>Documentation Guidance

Organizations can document the following

  • To what extent does the system/entity consistently measure progress towards stated goals and objectives?
  • Given the purpose of this AI, what is an appropriate interval for checking whether it is still accurate, unbiased, explainable, etc.? What are the checks for this model?
  • What corrective actions has the entity taken to enhance the quality, accuracy, reliability, and representativeness of the data?
  • To what extent are the model outputs consistent with the entity’s values and principles to foster public trust and equity?
  • How will the accuracy or appropriate performance metrics be assessed?

AI Transparency Resources

  • GAO-21-519SP - Artificial Intelligence: An Accountability Framework for Federal Agencies & Other Entities.
  • Artificial Intelligence Ethics Framework For The Intelligence Community.

>References

Arvind Narayanan. "The limits of the quantitative approach to discrimination." 2022 James Baldwin lecture, Princeton University, October 11, 2022.

Devansh Saxena, Karla Badillo-Urquiola, Pamela J. Wisniewski, and Shion Guha. “A Human-Centered Review of Algorithms Used within the U.S. Child Welfare System.” CHI ‘20: Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems, April 23, 2020, 1–15.

Rachel Thomas and David Uminsky. “Reliance on Metrics Is a Fundamental Challenge for AI.” Patterns 3, no. 5 (May 13, 2022): 100476.

Momin M. Malik. "A Hierarchy of Limitations in Machine Learning." arXiv preprint, submitted February 29, 2020. [URL](https://arxiv.org/abs/2002.05193

>AI Actors

TEVV
AI Deployment
Operation and Monitoring

>Topics

TEVV
Effectiveness

>Cross-Framework Mappings

Ask AI

Configure your API key to use AI features.