Last year saw huge advances in generative AI, as people experienced the ability to generate lifelike visuals with words and Microsoft Copilot tools that can summarize missed meetings, help write business proposals or suggest a dinner menu based on what’s in your fridge. While Microsoft has long established principles and processes for building AI applications in ways that seek to minimize unexpected harm and give people the experiences they’re looking for, deploying generative AI products on such a large scale has introduced new challenges and opportunities.
That’s why Microsoft recently released its first annual Responsible AI Transparency Report to help people understand how we approach responsible AI (RAI). The company has also rolled out new tools available in Azure AI for enterprise customers and developers to help safeguard the quality of their AI outputs and protect against malicious or unexpected uses of the systems.
It’s been a momentous year of stress-testing exciting new technology and safeguards at scale. Here are some key takeaways from Natasha Crampton, Microsoft’s Chief Responsible AI Officer, who leads the team defining and governing the company’s approach to RAI, and Sarah Bird, Microsoft’s Chief Product Officer for Responsible AI, who drives RAI implementation across the product portfolio:
#1: Make responsible AI a foundation, not an afterthought
Microsoft has also invested in mandatory training to build awareness and advocacy across the company – at the end of last year 99 percent of employees had completed a training module on responsible AI in our annual standards of business conduct training.
“It’s not possible to do responsible AI work as some sort of after-thought bolt on checklist immediately prior to shipping a product,” says Natasha Crampton. “It needs to be integrated into the way in which we build products from the very beginning. We need everyone across the company to be thinking about responsible AI considerations from the very get go.”
#2: Be ready to evolve and move quickly
It was through this process that Microsoft decided to offer different conversational styles – more creative, more balanced or more precise mode – as part of Copilot on its Bing search engine.
“We need to have an experimentation cycle with them where they try things on,” says Sarah Bird. “We learn from that and adapt the product accordingly.”
#3: Centralize to get to scale faster
Microsoft’s responsible AI experts also developed a new approach that centralizes how product releases are evaluated and approved. The team reviews the steps product teams have taken to map, measure and manage potential risks from generative AI, based on a consensus-driven framework, at every layer of the technology stack and before, during and after a product launch. They also consider data collected from testing, threat modeling and “red-teaming,” a technique to pressure-test new generative AI technology by attempting to undo or manipulate safety features.
Centralizing this review process made it easier to detect and mitigate potential vulnerabilities across the portfolio, develop best practices, and ensure timely information-sharing across the company and with customers and developers outside Microsoft.
“The technology is changing, superfast,” says Sarah Bird. “We’ve had to really focus on getting it right once, and then reuse (those lessons) maximally.”
#4: Tell people where things come from
In February, Microsoft joined with 19 other companies in agreeing to a set of voluntary commitments aimed at combating deceptive use of AI and the potential misuse of “deepfakes” in the 2024 elections. This includes encouraging features to block abusive prompts aimed at creating false images meant to mislead the public, embedding metadata to identify the origins of an image and providing mechanisms for political candidates to report deepfakes of themselves.
Microsoft has developed and deployed media provenance capabilities – or “Content Credentials” – that enable users to verify whether an image or video was generated by AI, using cryptographic methods to mark and sign AI-generated content with metadata about its source and history, following an open technical standard developed by the Coalition for Content Provenance and Authenticity (C2PA), which we co-founded in 2021. Microsoft’s AI for Good Lab has also directed more of its focus on identifying deepfakes, tracking bad actors and analyzing their tactics.
“These issues aren’t just a challenge for technology companies, it’s a broader societal challenge as well,” says Natasha Crampton.
#5: Put RAI tools in the hands of customers
Last year, Microsoft released Azure AI Content Safety, a tool that helps customers identify and filter out unwanted outputs from AI models such as hate, violence, sexual or self-harm content. More recently, the company has added new tools that are now available or coming soon in Azure AI Studio to help developers and customers improve the safety and reliability of their own generative AI systems.
These include new features that allow customers to conduct safety evaluations of their applications that help developers to identify and address vulnerabilities quickly, perform additional risk and safety monitoring and detect instances where a model is “hallucinating” or generating data that is false or fictional.
“The point is, we want to make it easy to be safe by default,” says Sarah Bird.
#6: Expect people to break things
In addition to probing for potential vulnerabilities before it releases updates of new AI products, Microsoft works with customers to ensure they also have the latest tools to protect their own custom AI applications built on Azure.
For instance, Microsoft has recently made new models available that use pattern recognition to detect and block malicious jailbreaks, helping to safeguard the integrity of large language models (LLM) and user interactions. Another seeks to prevent a new type of attack that attempts to insert instructions allowing someone to take control of the AI system.
“These are uses that we certainly didn’t design for, but that’s what naturally happens when you are pushing on the edges of the technology,” says Natasha Crampton.
#7: Help inform users about the limits of AI
Since 2019, Microsoft has been releasing “transparency notes” providing customers of the company’s platform services with detailed information about capabilities, limitations, intended uses and details for responsible integration and use of AI. The company also includes user-friendly notices in products aimed at consumers, such as Copilot, to provide important disclosures around topics like risk identification, the potential for AI to make errors or generate unexpected content, and to remind people they are interacting with AI.
As generative AI technology and its uses continue to expand, it will be critical to continue to strengthen systems, adapt to new regulation, update processes and keep striving to create AI systems that deliver the experiences that people want.
“We need to be really humble in saying we don’t know how people will use this new technology, so we need to hear from them,” says Sarah Bird. “We have to keep innovating, learning and listening.”
Images for this story were created using Microsoft Designer, an AI-powered graphic design application.