Skip to content

Latest commit

 

History

History

responsible-ai

Folders and files

NameName
Last commit message
Last commit date

parent directory

..
 
 

Guidance: Responsible AI (RAI)

Responsible AI (RAI)

Introduction

There is a lot to be excited about with recent advances in AI and LLM technology, but every day there are examples in the media about where and how it has gone wrong. As AI is integrated into more of our daily work and personal lives it can cause minor inconveniences, such as mistakenly canceled appointments, to more serious issues, such as potential job displacement and privacy compromises — and may even compound already existing social or economic inequities. All of us who design, develop, and deploy AI have a responsibility to confront the risks that the technology introduces.

If your team makes use of AI APIs or AI systems or designs, develops, or deploys AI, please consider joining us in the commitment to innovate responsibly. This page contains recommendations and guidance for how to get started.

Get Started

RAI sounds promising. Who wouldn't want their product to be responsible or trustworthy? But there can be challenges:

  • AI systems are complex and require a diversity of teams, skills, and tools
  • Potential harms and risks of AI systems are different from those of traditional software systems
  • Leadership support is crucial for a sustainable RAI practice

The recommendations that follow are based on lessons learned from rolling out a responsible AI practice across one of the organizations within Microsoft. As you review the recommendations, keep in mind they should be adapted to fit your organization, circumstances, and product plans.

The approach used was grounded in principles, standards, practices that were used internally and are published here:

  • Microsoft AI principles of fairness, reliability and safety, privacy and security, Inclusiveness, Transparency, and Accountability. However your organization defines and prioritizes the principles, they become the foundation for RAI standards and practices.
  • Microsoft Responsible AI Standard, v2: The RAI standard is the first step in putting the principles into practice. It answers the question... How are we going to execute on a responsible AI practice...?. For each principle, the standard sets out concrete steps and outcomes.
  • Microsoft Responsible AI Impact Assessment Template and Impact Assessment Guide provides documentation and guidance for how a team can capture and document the benefits, potential harms and mitigations for an AI system.
  • RAI tools and guidance such as the Responsible AI Maturity Model, HAX Toolkit, RAI Toolbox, and more (link here).

Complete an impact assessment

A Responsible AI impact assessment is the process a product team follows to identify and assess the potential risks and harms of an AI system. It is a new process, and some organizations may be reluctant to consider it, giving reasons such as:

  • It is still too early in AI lifecycle to do impact assessments. RAI is mostly academic.
  • AI is so new. How can we expect product teams to know about potential risks and harms?
  • Discussions around RAI will only devolve into messy disagreements and take time away from design/development/deployment timelines.

An RAI impact assessment is the primary method for guiding a team through the process of examining an AI system and aligning it to responsible AI principles and standards. The questions it examines include: What are the use cases for the AI system? Who are the stakeholders? How do we monitor and measure AI? Who might be harmed and how? How do we prevent these harms?

The purpose of a good RAI impact assessment is to identify the potential risks and harms of an AI system and introduce mitigations to reduce negative consequences. The templates and guidance introduced above can help a team put responsible AI principles to work. Consider adjustments to better align with your organizational requirements and product team processes.

Plan to update an impact assessment on an ongoing basis as use cases change, new functionality or technology is introduced, or even as the statistical property of the training data changes over time (data drift).

Strategies to help

The pace of AI is moving quickly and a variety of methods are needed to help measure and advance the responsible design, development, and deployment of AI systems.

Adopt a layered mitigation approach for RAI

At Microsoft, we recommend a layered mitigation approach that combines technical, operational, and governance measures to reduce the potential harms of LLMs. A layered approach applies different measures at different stages of development and deployment as documented in the article Deploy large language models responsibly with Azure AI and depicted below: Responsible AI (RAI) See Harms mitigation strategies with Azure AI for a walk-through of the mitigation layers.

Get started with RAI metrics

Measuring RAI involves both technical attributes, like accuracy and security, and socio-technical attributes such as fairness, representational harms, or safety and reliability. A starter set of metrics could include platform, usage, and operational level metrics like: * resource consumption (GPU, CPU, memory utilization, response times), * response/request tracking (number of unique/active users, requests, prompts, token usage), and * operational metrics (quota, capacity, latency, throttling). See Monitoring Azure OpenAI Service.

Add to this a layer of metrics to measure the potential for RAI harms identified through the impact assessment process. Consider measuring the number of input prompts blocked, number of responses blocked, number of jailbreak attempts, or metrics related to quality of generated output such as groundedness, relevance, and similarity.

Tools to support RAI

Tools can be a particular challenge because the RAI tools landscape is relatively new and rapidly evolving. While there are libraries and building blocks, the overall landscape can seem fragmented, making it difficult to establish an end-to-end workflow. Some of the challenges include:

  • There is no single tool that can holistically assess potential AI harms.
  • Tools often do not map directly to an RAI principle or standard, such as fairness or transparency, making it difficult to determine an appropriate tool.
  • Components and libraries built to support distinct aspects of RAI practice often require additional effort to use them in a complementary way.

Tools to support an RAI practice for GenAI applications focus on model selection, prompt engineering, and model output. Recommendations include:

Resources & References

Here are additional RAI resources: