Virtual Library

Our virtual library is an online repository of all of the reports, papers, and briefings that IST has produced, as well as works that have influenced our thinking.

Submit your Content

Op-ed

ROOST Reminds Us Why Open Source Tools Matter

view

Reports

Navigating AI Compliance, Part 2: Risk Mitigation Strategies for Safeguarding Against Future Failures

Mariami Tkeshelashvili, Tiffany Saade

viewpdf

Reports

Deterring the Abuse of U.S. IaaS Products: Recommendations for a Consortium Approach

Steve Kelly, Tiffany Saade

viewpdf

Podcasts

TechnologIST Talks: Looking Back and Looking Ahead: Deep Dive on the New Cybersecurity Executive Order

Carole House, Megan Stifel, and Steve Kelly

view

Podcasts

TechnologIST Talks: The Offense-Defense Balance

Philip Reiner and Heather Adkins

view

Reports

The Generative Identity Initiative: Exploring Generative AI’s Impact on Cognition, Society, and the Future

Gabrielle Tran, Eric Davis

viewpdf

Podcasts

TechnologIST Talks: A Transatlantic Perspective on Quantum Tech

Megan Stifel and Markus Pflitsch

view

Contribute to our Library!

We also welcome additional suggestions from readers, and will consider adding further resources as so much of our work has come through crowd-sourced collaboration already. If, for any chance you are an author whose work is listed here and you do not wish it to be listed in our repository, please, let us know.

SUBMIT CONTENT

Navigating AI Compliance, Part 2: Risk Mitigation Strategies for Safeguarding Against Future Failures

Mariami Tkeshelashvili, Tiffany Saade

SUMMARY

Historical trends do not wholly dictate the future of AI. While the first installment of this report acknowledged the importance of historic lessons, we can make deliberate choices to shape what comes next. We hope that “Navigating AI Compliance, Part 2: Safeguarding Against Future Failures” will guide decision-makers in fostering societal trust in AI systems, all while preventing the repetition of past mistakes.

This report, the second in a two-part series, presents 39 risk mitigation strategies for avoiding institutional, procedural, and performance failures of AI systems (see Risk Mitigation Strategies for Safeguarding Against Future Failures). These strategies aim to enhance user trust in AI systems and maximize product utilization. AI builders and users, including AI labs, enterprises deploying AI systems, as well as state and local governments, can use and implement a selection of the 22 technical and 17 policy-oriented risk mitigation strategies presented in this report according to their needs and risk thresholds. 

Through implementing these practices, organizations building and utilizing AI systems not only reduce regulatory risk exposure and build user trust for their product, but they could also attract top talent, gain a competitive edge, enhance their financial performance, and increase the lifetime value of their solutions. Based on our research and the results of stakeholder engagement, we emphasize to AI builders and users the following nine recommendations from the complete list of 39: 

  • Implement proportional compliance measures for high-impact AI applications. AI builders and users should consider which compliance measures are most appropriate for their work, especially when building or deploying AI systems in sensitive or high-impact areas. This consideration should be proportional to factors such as the intended use, potential risks, and application domain—ranging from entertainment to critical sectors like national security, healthcare, and finance. 
  • Acknowledge and address acceptable risks in AI development and deployment. Unintended consequences are not to be confused with compliance failure. Still, these unplanned effects should be acknowledged by developers, builders, and regulators as they consider thresholds of acceptable tolerance for the enhanced risks associated with exposed attack surfaces and features or functionalities of AI that are not yet thoroughly understood or anticipated.
  • Prioritize data management and privacy practices to maintain user trust. Implementing proper data management and privacy-enhancing practices will protect user rights, maintain trust, and comply with data protection regulations. Measures such as privacy-preserving technologies, content provenance features, and user consent mechanisms can alleviate procedural failures.
  • Implement robust cybersecurity controls for AI infrastructure protection and enhanced reliability. Cybersecurity controls, red-teaming, fail-safe mechanisms, and other techniques protect AI systems from attacks and strengthen their reliability in various scenarios. Security guardrails may alleviate or preempt both performance and procedural failures. 
  • Utilize safety and risk assessments to proactively mitigate AI harms. Safety and risk assessment procedures, such as incident reporting frameworks and AI safety benchmarks at different stages of the lifecycle, identify and mitigate possible harms before they occur–potentially mitigating both procedural and performance failures. 
  • Design and implement compliance and AI literacy training for staff. Training should be mandatory for all staff members involved in the AI supply chain, from data providers to model developers and deployers. All staff members utilizing AI tools in some manner should also obtain a minimum set of AI literacy skills through the training. 
  • Build trust by implementing transparency mechanisms. Transparency and interpretability mechanisms such as model cards, data cards, and disclosure frameworks are necessary to build user and stakeholder trust, facilitate accountability, and enable informed decision-making.
  • Enhance AI explainability and disclosure frameworks to improve understanding of system behavior. Efforts to increase the explainability of AI systems, supplemented with disclosure frameworks for model evaluation, allow both builders and users to better understand the behavior patterns and outputs of these systems and potentially safeguard against performance failures.
  • Employ strategies for non-discriminatory AI. Bias mitigation strategies across model training, data collection, and ongoing monitoring and maintenance, in addition to adversarial debiasing, can prevent performance failures and help to ensure fairness while preventing discriminatory outcomes in AI systems.

Learn more about the AI Risk Reduction Initiative

See Part 1: Tracing Failure Patterns in History

download pdf