The Principles in Action Playbook contains guidance, examples, and tips to help those building products that leverage AI to do so responsibly.
Guided by the Vector Institute’s AI Trust and Safety Principles, we put principles in action.
We encourage you to start by exploring the product development process and reflecting on the key questions at each of the five stages. Use it to assess where you’re at and where you still need to go.
Our worksheets were created to help put the guidance into action in your product. They can be used at any stage of the product development process to facilitate discussions across stakeholders.
We’re excited that you’re looking to build something that leverages AI. As you go through the product development process, focus on being people-first over technology-first.
For example, instead of approaching your brainstorming with “How can we use AI to ___?”
‘How might we’ statements force you to start with a human need and steer you away from suggesting a solution, so that you can be open to generating further possibilities.
How do you decide what problem you should tackle?
Identify a need humans have.
Collect evidence that this is a problem.
Don’t use AI just because you can. Think hard about whether you really need AI in your product to solve your problem:
AI is good at: | Avoid AI when: |
---|---|
|
|
AI is good at:
Avoid AI when:
If leveraging AI makes sense for your product, then consider this:
Evaluate your capacity before you start building to decide whether your team is able to implement your idea.
How will you know when your AI system is good enough for people to use?
Start by identifying the action or behaviour you are trying to optimize and the possible outcomes. If your AI product will make predictions, you can expect successes: true positives and true negatives, and errors: false positives and false negatives.
Think through the consequences of false positive and false negative predictions and weigh the cost of these errors.
Next, consider how your model metrics translate to your product metrics. When choosing high-level product metrics, such as engagement, speed, or cost savings, consider the following:
AI is not perfect; it’s probabilistic, so you should expect your product to give users incorrect or unforeseen output at some point, and those consequences can have their own consequences, also known as second-order effects.
Plan to design your user experience around these error possibilities.
The default reaction to poor AI system output doesn’t always have to be to fix your AI model to get better results; you can make design changes to the user experience, too.
Building AI solutions differs from traditional software development where there are often defined product milestones, requirements, and estimates.
Whether or not you decide to build your own AI depends on various factors such as your specific goals, resources, expertise, and the availability of suitable AI solutions in the market.
Here are some reasons to build your own model: | Here are some reasons not to build your own model: |
---|---|
|
|
Here are some reasons to build your own model:
Here are some reasons not to build your own model:
If your team has decided to build its own AI model or fine-tune an off-the-shelf solution you will need data for training and testing. Your AI model, and thus your product, will only be as good as the data and labels that feed it, so think through your data needs carefully.
Remember, it is your responsibility to minimize unfair bias in your dataset. You are not absolved of your human rights obligations just because you don’t have access to good data.
Consulting people with subject matter expertise will greatly help you in this process. Domain experts don’t need to be data experts, they just need to be willing to share insights and highlight implications about your data’s subject matter.
Now that you have good quality data that reflects your users and use case, you can start thinking about how to develop the model that will output predictions or content that will help address your users’ needs.
Here are some tips:
The idea of achieving an optimally responsible model is a fallacy; developing models is a constant balance of making trade-offs for your unique use case.
Test, test, test!
This is true in traditional software engineering, and is especially true for model development.
We can say with confidence that your AI system will give wrong and unexpected outputs at some point. To limit bad predictions users might encounter, make a plan for testing early on in your product life cycle.
Here are some tips:
Model development is an iterative process - make sure your product leaders and stakeholders understand that. Quick, positive outcomes achieved by implementing the happy path is an effective way to gain the favour of both your product manager and user.
Focus on initially getting qualitative feedback from a variety of users instead of obsessively tracking metrics. Your users will quickly tell you if it’s not working as expected.
Perform retrospective testing on your model. Have multiple individuals with the appropriate expertise examine cases where there is a mismatch between the model output and the ground truth. Measure the agreement between humans and the model, and between humans.
Next, you can test using a prototype within your team, then within your company. Avoid relying on the development team for testing as they may only test particular aspects of the system or look to confirm something they already know. The test cases they write are typically limited to how they think the system will be used. By opening up testing to a variety of people in your organization, you can observe the variety of alternative ways the system is used by real people.
Eventually, bring in actual people to observe how they interact with your product. Test with a small group of representative users in a staging environment and give users the option to use or ignore your model’s predictions. Pre- and post-surveys work well to understand how users felt about a product change. Quick, optional pop-ups presented at the time of output are ideal.
There are many metrics you can use to evaluate your AI model. But at the end of the day, what you really care about is assessing whether you’ve addressed your target user’s needs in a responsible way. Therefore, the performance of your model should be measured against product success metrics and bias and fairness metrics. Choose metrics that are simple to measure.
For product metrics:
First, measure a user behaviour that is directly observed and attributable to an action of the system: | During A/B testing and launch decisions, measure indirect effects: |
---|---|
|
|
First, measure a user behaviour that is directly observed and attributable to an action of the system:
During A/B testing and launch decisions, measure indirect effects:
Finally, don’t expect your new AI to tell you:
These are all important, but also incredibly hard to measure. Instead, use indirect indicators: if the user is happy, they will stay on the site longer. If the user is satisfied, they will visit again tomorrow.
Remember, there is no one metric that will tell you that your product is a good one and a responsible one. Your team should care about engagement, daily active users, retention, and revenue during A/B testing. But while A/B tests help us optimize specific elements, they're not the end goal. Your true focus is on creating a product that delights users, attracts more customers, fosters strong partnerships, and drives sustainable growth in a safe, trustworthy, and ethical way.
As such, always assess your product metrics with fairness and bias in mind:
One of the challenging parts of model development may be using your own judgement to assess trade-offs while optimizing for certain metrics, e.g. fairness vs accuracy or false positives vs false negatives.
Work on building trust from day one.
Be cognizant of the language you use to describe your AI-powered product. Messaging such as “magical” and “human-like” can leave users with the wrong impression of what your system is capable of.
Phase rollout.
Prepare to onboard users to the new product or feature.
Set expectations about what your AI-powered product can do, cannot do, its risks, and how to improve it.
Use hedging language when appropriate, e.g. “We think you’ll like…”.
Have a plan for handling errors and failures so users can move forward with completing their task.
AI is not perfect, and you need to remind users of this. Avoid suggesting that your product is perfect and can fully replace a specific task, especially if your system’s outputs are not yet reliable.
When helpful or necessary, tell users how confident you are in a specific prediction or recommendation, especially in high stakes situations, to help them gauge trust in the system and guide their decision-making.
Confidence can be communicated in a number of ways, including visual bar charts, percentages, rankings, attributions, or categorically (e.g. “best match” or “this price is likely to increase” or “because you read fantasy”). You will have to define the threshold for how to group recommendations, which may require user testing.
A released model should be accompanied by documentation detailing its dataset and performance metrics in aggregate and across different demographic groups. Though disclosing this information may put companies in an uncomfortable position, it is good practice. The documentation should capture the trade-offs across different metrics.
Consider open-sourcing your model to increase transparency and trust. It's important to note, however, that open-sourcing also involves potential risks, such as intellectual property concerns and the possibility of misuse. Carefully consider these factors before making a decision.
Your model has been trained; your product is being used by real people. What now?
Maintain user trust.
Revisit your feedback mechanisms.
Observe how users are engaging with your AI-powered product.
Evaluate whether your success metrics should change as users use your product more.
Stay updated on new regulations and the dynamic AI landscape to maintain compliance and adapt to evolving legal and ethical standards.
The Principles in Action Playbook incorporates insights from Vector Institute user interviews, alongside existing academic and industry research.
This playbook is for informational purposes only and Vector Institute is not, by means of this playbook, rendering professional advice or services or providing an opinion of any kind on any subject. No one should act upon or refrain from acting upon the information contained in this playbook without obtaining the advice of a qualified professional advisor. You are urged to contact a qualified professional advisor for guidance. Vector Institute does not warrant or guarantee the accuracy, currency, usefulness, or completeness of the information contained in this playbook, which may include information obtained from third party sources. Vector Institute shall not have any responsibility or owe any duty to any person in respect of this playbook or be responsible for any loss whatsoever sustained by any person who relies on this playbook.
All content in this playbook is either created by Vector Institute, or otherwise used with permission and is protected by copyright. All rights are reserved. No part of this report may be reproduced without the prior written permission of Vector Institute.
Launched in 2017, the Vector Institute works with industry, institutions, startups, and governments to build AI talent and drive research excellence in AI to develop and sustain AI-based innovation to foster economic growth and improve the lives of Canadians. Vector aims to advance AI research, increase adoption in industry and health through programs for talent, commercialization, and application, and lead Canada towards the responsible use of AI. Programs for industry, led by top AI practitioners, offer foundations for applications in products and processes, company-specific guidance, training for professionals, and connections to workforce-ready talent. Vector is funded by the Province of Ontario, the Government of Canada through the Pan-Canadian AI Strategy, and leading industry sponsors from across multiple sectors of Canadian Industry. For further information or media enquiries, please contact: media@vectorinstitute.ai
© The Vector Institute for Artificial Intelligence.