Featured Product
This Week in Quality Digest Live
Customer Care Features
Michael King
Augmenting and empowering life-science professionals
Meg Sinclair
100% real, 100% anonymized, 100% scary
Mike Figliuolo
The customer isn’t always right
Alonso Diaz
Consulting the FDA’s Case for Quality program
Mike Figliuolo
Focus on solving defined problems and increase your chances of market adoption

More Features

Customer Care News
For companies using TLS 1.3 while performing required audits on incoming internet traffic
Accelerates service and drives manufacturing profitability
Providing pipeline integrity with remote monitoring
Integrated energy services enhance customer experience
Empowers real-time remote monitoring and technical support
Latest features enhance complex BOM generation and customer experiences
New video in the NIST ‘Heroes’ series
A tool to help detect sinister email

More News


Customer Care

NIST Seeks Collaborators to Support Artificial Intelligence Safety

Developing tools to measure and improve trustworthiness

Published: Monday, November 27, 2023 - 12:00

(NIST: Gaithersburg, MD) -- The U.S. Department of Commerce’s National Institute of Standards and Technology (NIST) is calling for participants in a new consortium supporting development of innovative methods for evaluating artificial intelligence (AI) systems to improve the rapidly growing technology’s safety and trustworthiness. This consortium is a core element of the new NIST-led U.S. AI Safety Institute announced recently at the U.K.’s AI Safety Summit 2023, in which U.S. Secretary of Commerce Gina Raimondo participated.

The institute and its consortium are part of NIST’s response to the recently released Executive Order on Safe, Secure, and Trustworthy Development and Use of AI. The order tasks NIST with a number of responsibilities, including developing a companion resource to the AI Risk Management Framework (AI RMF) focused on generative AI; guidance on authenticating content created by humans and watermarking AI-generated content; a new initiative to create guidance and benchmarks for evaluating and auditing AI capabilities; and creating test environments for AI systems. NIST will rely heavily on engagement with industry and relevant stakeholders in carrying out these assignments. The new institute and consortium are central to those efforts.

“The U.S. AI Safety Institute Consortium will enable close collaboration among government agencies, companies, and impacted communities to help ensure that AI systems are safe and trustworthy,” says Undersecretary of Commerce for Standards and Technology and NIST director Laurie E. Locascio. “Together we can develop ways to test and evaluate AI systems so that we can benefit from AI’s potential while also protecting safety and privacy.”

The U.S. AI Safety Institute will harness work already underway by NIST and others to build the foundation for trustworthy AI systems, supporting use of the AI RMF, which NIST released in January 2023. The framework offers a voluntary resource to help organizations manage the risks of their AI systems and make them more trustworthy and responsible. The institute aims to measurably improve organizations’ ability to evaluate and validate AI systems, as detailed in the AI RMF Road Map.

“The institute’s collaborative research will strengthen the scientific underpinnings of AI measurement so that extraordinary innovations in artificial intelligence can benefit all people in a safe and equitable way,” says NIST’s Elham Tabassi, federal AI standards coordinator and a member of the National AI Research Resource Task Force.

Building on its long track record of working with the private and public sectors, as well as its history of measurement and standards-oriented solutions, NIST is seeking collaborators from across society to join the consortium. The consortium will function as a convening space for an informed dialogue, sharing information and insights. It will be a vehicle to support collaborative research and development through shared projects, and will promote the assessment and evaluation of test systems and prototypes to inform future AI measurement efforts.

“Participation in the consortium is open to all organizations interested in AI safety that can contribute through combinations of expertise, products, data, and models,” says Jacob Taylor, NIST’s senior advisor for critical and emerging technologies. “NIST is responsible for helping industry understand how to manage the risks inherent in AI products. To do so, NIST intends to work with stakeholders at the intersection of the technical and the applied. We want the U.S. AI Safety Institute to be highly interactive because the technology is emerging so quickly, and the consortium can help ensure that the community’s approach to safety evolves alongside.”

In particular, NIST is soliciting responses from all organizations with relevant expertise and capabilities to enter into a consortium cooperative research and development agreement (CRADA) to support and demonstrate pathways to enable safe and trustworthy AI. Members would be expected to contribute:
• Expertise in one or more of several specific areas, including AI metrology, responsible AI, AI system design and development, human-AI teaming and interaction, socio-technical methodologies, AI explainability and interpretability, and economic analysis
• Models, data, and/or products to support and demonstrate pathways to enable safe and trustworthy AI systems through the AI RMF
• Infrastructure support for consortium projects
• Facility space and handling of hosting consortium researchers, workshops, and conferences

Interested organizations with relevant technical capabilities should submit a letter of interest by Dec. 2, 2023. More details on NIST’s request for collaborators are available in the Federal Register. NIST hosted a workshop on Nov. 17, 2023, for those interested in learning more about the consortium and engaging in the conversation about AI safety.

The U.S. AI Safety Institute will partner with other U.S. government agencies on evaluating AI capabilities, limitations, risks, and impacts, and coordinate on building testbeds. The institute will also work with organizations in allied and partner countries to share best practices, align capability evaluation, and provide red-team guidance and benchmarks.


About The Author

NIST’s picture


Founded in 1901, the National Institute of Standards and Technology (NIST) is a nonregulatory federal agency within the U.S. Department of Commerce. Headquartered in Gaithersburg, Maryland, NIST’s mission is to promote U.S. innovation and industrial competitiveness by advancing measurement science, standards, and technology in ways that enhance economic security and improve our quality of life.