being ethical: how to beat bias in ai and be...

12
Being Ethical: How to Beat Bias in AI and Be Transparent IN ASSOCIATION WITH:

Upload: phunganh

Post on 26-Apr-2019

212 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Being Ethical: How to Beat Bias in AI and Be Transparentinfo.forbes.com/rs/790-SNV-353/images/Intel-AI#4_REPORT-FINAL-WEB.pdf · Forum. “If we don’t ... 1 Vulnerability of Female

Being Ethical: How to Beat Bias in AI and Be Transparent

IN ASSOCIATION WITH:

Page 2: Being Ethical: How to Beat Bias in AI and Be Transparentinfo.forbes.com/rs/790-SNV-353/images/Intel-AI#4_REPORT-FINAL-WEB.pdf · Forum. “If we don’t ... 1 Vulnerability of Female
Page 3: Being Ethical: How to Beat Bias in AI and Be Transparentinfo.forbes.com/rs/790-SNV-353/images/Intel-AI#4_REPORT-FINAL-WEB.pdf · Forum. “If we don’t ... 1 Vulnerability of Female

COPYRIGHT © 2019 FORBES INSIGHTS | 3

TABLE OF CONTENTS

4 INTRODUCTION

5 THE IMPORTANCE OF DATA

7 THE ROAD TOWARD ETHICAL AI

8 THE FUTURE IS ETHICAL

10 METHODOLOGY

10 ACKNOWLEDGMENTS

Page 4: Being Ethical: How to Beat Bias in AI and Be Transparentinfo.forbes.com/rs/790-SNV-353/images/Intel-AI#4_REPORT-FINAL-WEB.pdf · Forum. “If we don’t ... 1 Vulnerability of Female

4 | BEING ETHICAL: HOW TO BEAT BIAS IN AI AND BE TRANSPARENT

INTRODUCTION

Few leaders in business and government dispute the power of artificial intelligence (AI) to change the world. Transformation and disruption are, after all, already happening in ways large and small, visible and invisible, across the business world and society at large, at unprecedented speed. The

question is now: How do we manage AI for the common good?

The issues at the heart of ethics in AI have been percolating over the past few years as companies across industries have begun to deploy and make decisions off machine learning (ML) applications

like image recognition and natural language processing (NLP). The benefits and potential to be unlocked are staggering, but it’s becoming increasingly apparent that many machine-driven

decisions—and the collection and use of data to train and fuel them—is happening outside of full human control or without established best practices. Among tech-forward brands and leaders,

there is a growing call to arms toward fairness, accountability and transparency.

The urgency around these issues runs deep. After all, any business wants its decisions—especially the ones that have a significant impact on individual customers—to be accurate and free of bias.

Business leaders and regulators need to know the why behind a machine-driven decision, and end-customers want to see transparency in this process and accountability for the results. AI can be a

boon for this if it is explainable.

“We’re seeing a gradual recognition that, with the deployment of more AI, ethical issues should be considered,” says Kay Firth-Butterfield, head of AI and machine learning at the World Economic

Forum. “If we don’t establish firm foundations around the technology, we will just build more problems for ourselves. We have to tackle the core issues around the deployment of AI for the

benefit of humanity.”

AI is now at a crossroads—and best practices and guidelines are needed to guide these powerful technologies down the right path. While many organizations are already taking action—79% said they’re employing tactics to ensure algorithms are built without bias, like having diverse teams working on their development, according to a Forbes Insights survey of more than 500

executives—more needs to be done. Less than half have revised their data policies to be proactive on transparency and even fewer have appointed a data steward and/or ethics engineer to oversee

the results of their AI models.

This report is about the challenges business leaders face in building ethical systems and how to go about solving them.

Page 5: Being Ethical: How to Beat Bias in AI and Be Transparentinfo.forbes.com/rs/790-SNV-353/images/Intel-AI#4_REPORT-FINAL-WEB.pdf · Forum. “If we don’t ... 1 Vulnerability of Female

COPYRIGHT © 2019 FORBES INSIGHTS | 5

THE IMPORTANCE OF DATA

Kathy Baxter is a prominent voice in the development of ethical technology. As the architect of the ethical AI practice at Salesforce, she develops research-informed best practices to educate Salesforce employees, customers and the industry on the development of ethical AI. Like Firth-Butterfield, she sees plenty of discussions and awareness around ethical issues. “But we’re not seeing as much action as I might hope, and part of the problem,” she says, “is that building ethical AI is hard—it’s a pretty tough nut to crack.”

Data is at the heart of ethics in AI for obvious reasons—it’s the fuel that makes the engine run. Solving for the challenges around data, which every business should do, will go a long way toward solving for ethics, notably the core concerns around bias and privacy. As seen in Figure 1, the Forbes Insights survey reflects the challenge behind this priority. Forty-four percent say that managing data in a privacy-compliant environment is a challenge, and almost half report that determining the right data to collect and analyze is another barrier.

AI technologies aren’t smart enough (yet) to weed out the biases either programed into them or fed into their systems with poor training data—data that isn’t accurate or inclusive (or exclusive, if relevant) for the model being designed. This bias that finds its way into algorithms is an issue for humans and society to solve. “We can’t expect AI, as amazing as it is, to remove bias that exists in our society. We have hundreds of years of decisions being made either implicitly or explicitly against certain groups,” Baxter says. “As long as that exists, it would be simply naive to think that any technology, including AI, could solve the bias that exists in our technology.”

Baxter cites vehicle crash test dummies as an example of an unwitting societal bias that ends up baked into product design—more specifically, it is sampling bias that excludes an entire gender. The standard in accident

Figure 1: Data Challenges

With respect to data, which of these statements reflect your view of your organization’s state of readiness for effective AI? (Select all that apply)

What specific challenges does your organization have with extracting value from the data you already have? (Select all that apply)

47%

47%

44%

39%

43%

37%

42%

33%

45%

29%

23%

21%

Ethical issues come into play when you’re capturing and using personal and product data, which must be addressed and solved for in order to see the continued advancement of AI technologies and IoT

Determining the right data to collect and analyze

Managing data in a privacy-compliant environment

Prioritizing data to generate insights

Measuring and benchmarking data quality and accuracy

Obtaining fresh, relevant data

Integrating multiple sources of data

Characterizing data segments

There’s a gap between the goals or vision for AI technologies at our organization and our ability to capture the data we want

We must buy most of the data we use to gain insights and/or train algorithms

Low-quality, incomplete or inaccurate data is holding us back from effective AI deployments

We don’t have access to quality data from the source (e.g., sensors on devices, machinery, real-time consumer behavior)

Page 6: Being Ethical: How to Beat Bias in AI and Be Transparentinfo.forbes.com/rs/790-SNV-353/images/Intel-AI#4_REPORT-FINAL-WEB.pdf · Forum. “If we don’t ... 1 Vulnerability of Female

6 | BEING ETHICAL: HOW TO BEAT BIAS IN AI AND BE TRANSPARENT

testing for decades was a dummy modeled after the average male body. But a study in 2011 by the University of Virginia’s Center for Applied Biomechanics found that, in one car model, women had a 47% greater chance of being injured in a 35-mph collision.1 The design of the vehicle had not factored in the physiology of the female body. Because the crash test dummies were all built for a male, the data was built around an assumption that proved to be wrong and unfair—and dangerous.

“They didn’t have a similar female weight and height and dimension crash test dummy,” Baxter says. “Once they started using a female crash test dummy, the results looked very different in evaluations of cars.” Bias, she notes, doesn’t have to be explicit. “It comes as default thinking that has existed in different industries for many decades. If we want to come up with a solution that creates a world we want to see, not the world as it exists, we’re going to have to put a lot of resources into making that happen.”

Here are two big steps to improve the data you’re working with and limit the possibility of introducing bias in your AI systems.

1. Clean training data. The design of the model is important, “and then, of course, the training data is a big source of bias in AI,” Baxter says. “That’s where history creates that bias that we see in the outcome. If you train a model to make recommendations for loan applications based on how you’ve always approved loans, it’s just going to replicate and magnify what humans already did.”

2. Monitor the impact. Although Salesforce doesn’t collect data from consumers, it works with its clients to help them understand the implications of the data they feed into Einstein, the company’s flagship software program. “They may not have thought about the data they’ve been collecting over years or decades in terms of bias,” Baxter says. “Having to go through and look at their data in a very different light is something new for them. We will raise a flag if we believe a variable might lead to bias, and we find ways to provide in-app education to empower customers in the moment to understand their models and the implications of their decisions.”

Yet the first step—clean your training data—is easier said than done. How do you actually go about it so it doesn’t bias the outcomes? Baxter points to the big W’s around data quality and accuracy as the road to finding and eliminating bias. For each of the questions below, consider why there is a challenge and get to the root of the problem.

• WHERE IS YOUR DATA COMING FROM? That question is just the tip of the ethics iceberg. Ask who was involved in getting your data and who interacted with it. Look at the sources—is this first-party data or third-party data from a vendor?—and who collected it, categorized it or labeled it. All of this information may affect the credibility of the data you’re feeding into your algorithms.

1 Vulnerability of Female Driver Involved in Motor Vehicle Crashes, https://www.ncbi.nlm.nih.gov/pmc/articles/PMC3222446/

[Bias] comes as default thinking that has existed in different industries for many decades. If we want to come up with a solution that creates a world we want to see, not the world as it exists, we’re going to have to put a lot of resources into making that happen.”KATHY BAXTER, ARCHITECT, ETHICAL AI PRACTICE, SALESFORCE

Page 7: Being Ethical: How to Beat Bias in AI and Be Transparentinfo.forbes.com/rs/790-SNV-353/images/Intel-AI#4_REPORT-FINAL-WEB.pdf · Forum. “If we don’t ... 1 Vulnerability of Female

COPYRIGHT © 2019 FORBES INSIGHTS | 7

• WHEN IS YOUR DATA COMING? This is about data freshness (which, back to the previous figure, 37% of executives say is challenging to obtain). Digging deeper, consider the impact of an organization’s decision if the data used in its AI is old—hours or days old, even weeks or months. Decisions based on the AI might be inaccurate; worse, they might be unfair to the customer. At scale, the bias from such errors can have a big impact on results and even brand reputation.

• WHO IS THE DATA FROM OR ABOUT? This is about representation and being sure your data sets reflect the full and accurate panorama of data about your target—customers or other users. Capturing or feeding training data missing a subset or specific type of target will produce skewed results that, again, may have unintended consequences.

• WHAT VARIABLES ARE YOU USING? This is where using or failing to use categories like race, ethnicity, gender, age, sexual orientation and other data can produce biased results. Think back to the crash test dummy example that produced bias by excluding gender data. Additionally, any data that’s not directly relevant to the end goal but could influence outcomes should be weeded out.

THE ROAD TOWARD ETHICAL AI

Talk to top technology thinkers and leaders and you’ll hear that algorithmic bias is solvable—and it is. But as Baxter and Firth-Butterfield reiterate, it requires action. Black-box insights and actions won’t work for business leaders. And they won’t work for regulators in many industries. Decisions with real-world importance—a loan decision for a family or a business; a medical decision on surgery for a patient—have to be understandable and traceable. The answer won’t be found in the millions of lines of code in a neural network.

“We could open up the hood of an AI and take a look at the code, the model, but it really isn’t going to tell you anything,” Baxter says. “There’s a lot you have to do for it to be understandable. It isn’t a one size fits all. You need to have different solutions for different types of AI.”

Enter explainable AI, a topic that inspires different responses and some confusion. One camp of scientists maintains that humans can never understand these algorithms because they’re far too complex. A solution on that side of the argument calls for the creation of guardian algorithms to look after standard

Page 8: Being Ethical: How to Beat Bias in AI and Be Transparentinfo.forbes.com/rs/790-SNV-353/images/Intel-AI#4_REPORT-FINAL-WEB.pdf · Forum. “If we don’t ... 1 Vulnerability of Female

8 | BEING ETHICAL: HOW TO BEAT BIAS IN AI AND BE TRANSPARENT

algorithms. The other camp says that black-box stance is wrong and that humans need to be in the loop. “It is possible to create transparency,” Baxter says. “That camp seems to be gaining ascendency at the moment—and that’s probably just because we’ve encountered the problems. Now more people have actually started to think about how we can do this.”

Explainability means a leader or team can see what went right—or, more important, what went wrong—and take ownership for it. A brand can own the results and reassure customers, even attract them to its brand by its positions on ethical issues. Many are hiring for new positions created around the imperative to understand and mold algorithms.

Look at the new jobs coming online for proof of the urgency around transparency—explainability analysts who determine the impact of outcomes, trainers who monitor the inputs, controllers who mold bots and virtual agents (Figure 2). This is the future. And it only shows how important the human hand is—and will continue to be—as AI advances. Ultimately, it will be the human involvement in these technologies that ensures trust in these systems.

THE FUTURE IS ETHICAL

The full value to be unlocked by AI can emerge only when business leaders and all stakeholders—consumers and society at large—place their faith in the results produced by machines. But that will happen only when they see action from the top, at the board level in organizations and from the governments that represent them. And, for the time being, it may depend on active human management of these deep, mysterious networks.

As you continue your AI journey, be sure to consider the following:

• LEADERSHIP DEFINES ETHICS: The touchstone of a work culture created around ethical AI is leadership, the same quality that is essential in the adoption and development of innovative technology in the first place. Diversity and a rigorous approach to data across an organization are de rigueur, but a truly ethical culture is one where values and approaches permeate an organization. This in turn will attract talented people with the right mindset. You end up with a virtuous circle.

How do you structure a team to develop great, ethical AI? “It’s about culture,” Baxter says. “We have an agile-meets-AI approach, in which we make sure that teams are thinking through the potential implications of AI and

Figure 2: What new roles do you see being formed now or in the near future as a result of AI technologies? (Select all that apply)

55% 54%

44%

36%

Explainability analysts to

determine the impact of outputs from AI systems

Trainers to teach AI

systems how to perform

Algorithm forensics analysts

Bot controllers

Page 9: Being Ethical: How to Beat Bias in AI and Be Transparentinfo.forbes.com/rs/790-SNV-353/images/Intel-AI#4_REPORT-FINAL-WEB.pdf · Forum. “If we don’t ... 1 Vulnerability of Female

COPYRIGHT © 2019 FORBES INSIGHTS | 9

the potential unintended use cases. Then we go through the results and develop a maturity model that allows us to evaluate the results and see how we’re getting better over time. We want to keep ourselves honest and to hold ourselves accountable; we have to find ways to be able to say we are doing better today than yesterday. Having that maturity model is one way of being able to measure improvement or measure change over time.”

• DIVERSITY IS IMPERATIVE. “The most important action is diversity. We need to be training more people from minorities and more women to create algorithms because the actual training of the algorithms carries the biases of the people who create them,” says Firth-Butterfield, adding that it’s important for “the teams who create algorithms to look like our society and thus reflect it. This is also true internationally, which means that countries and regions need to be training their own AI developers.”

Baxter agrees: “Diversity is huge, and it’s not just diversity of gender and race or age. It’s also about experience and background. When I’m in meetings and thinking about products and features and impact, I feel a responsibility to make sure that I’m always representing the people who are not in the room. We have to be able to think about everybody, and you can only do that if there’s representation in the room.”

• MODELS CAN ALWAYS BE IMPROVED. Ethics is reflected in how an AI is built, the parameters that define the problem to be solved. The models that are the foundation of machine learning systems have to be free of implicit or explicit bias—and that’s not easy to do. Baxter talks about the potential presence of proxy variables—data that might not be directly relevant to the goal but influences outcomes. In the U.S., she notes, zip code and income can be a proxy for race, for example.

• ETHICS CAN BE TAUGHT. The World Economic Forum is creating a repository of curricula from professors who are teaching their AI students about the social implications of AI, including bias, at the undergraduate and graduate level. “We’re inviting professors to upload their curricula into the forum repository so that we can then offer it to other professors around the world who want to set up curricula like this,” Firth-Butterfield says. Yet ethics can be taught outside the classroom as well. Consider hosting workshops or forums at your organization where ethics becomes a focal point.

“I want to make sure that we think about AI for good,” Baxter says. “But we also have to make sure that we don’t end up with tech worship, a belief that because it’s technology, it’s neutral and beneficent and that we can take our hands off the wheels and everything will go well. There’s tremendous power to create a much more fair and equitable world. But it takes effort, and a lot of thought, to be able to make sure that happens.”

To learn more about how companies are leading with AI today, visit Forbes AI.

I want to make sure that we think about AI for good . But we also have to make sure that we don’t end up with tech worship, a belief that because it’s technology, it’s neutral and beneficent and that we can take our hands off the wheels and everything will go well. There’s tremendous power to create a much more fair and equitable world. But it takes effort, and a lot of thought, to be able to make sure that happens.”

KATHY BAXTER, ARCHITECT, ETHICAL AI PRACTICE, SALESFORCE

Page 10: Being Ethical: How to Beat Bias in AI and Be Transparentinfo.forbes.com/rs/790-SNV-353/images/Intel-AI#4_REPORT-FINAL-WEB.pdf · Forum. “If we don’t ... 1 Vulnerability of Female

10 | BEING ETHICAL: HOW TO BEAT BIAS IN AI AND BE TRANSPARENT

ACKNOWLEDGMENTS

METHODOLOGY

Forbes Insights and Intel would like to thank the following individuals for their time and expertise:

• Kathy Baxter, Architect, Ethical AI Practice, Salesforce

• Kay Firth-Butterfield, Head of Artificial Intelligence and Machine Learning, World Economic Forum

The findings in this report are based on a Forbes Insights survey of 510 executives who lead organizations in the United States. Nearly half (48%) are C-suite executives, representing CEOs (15%), CIOs (10%), CTOs (9%), CISOs (3%) and other C-suite positions (11%). All executives are involved in the development and implementation of their organization’s AI strategy. The executives work in a variety of sectors, including energy, financial services, retail, healthcare, communications, manufacturing, transportation, automotive and distribution. All represent organizations with $500 million or more in annual revenue, with 74% coming from organizations with more than $1 billion in annual revenue.

EXECUTIVES

3%

9%

10%

11% CISOs

CTOs

CIOs

Other C-suite

positions

15%CEOs

48%C-SUITE EXECUTIVES

EXECUTIVES

Page 11: Being Ethical: How to Beat Bias in AI and Be Transparentinfo.forbes.com/rs/790-SNV-353/images/Intel-AI#4_REPORT-FINAL-WEB.pdf · Forum. “If we don’t ... 1 Vulnerability of Female

C-SUITE EXECUTIVES

Page 12: Being Ethical: How to Beat Bias in AI and Be Transparentinfo.forbes.com/rs/790-SNV-353/images/Intel-AI#4_REPORT-FINAL-WEB.pdf · Forum. “If we don’t ... 1 Vulnerability of Female

12 | BEING ETHICAL: HOW TO BEAT BIAS IN AI AND BE TRANSPARENT

ABOUT FORBES INSIGHTS

Forbes Insights is the strategic research and thought leadership practice of Forbes Media, a global media, branding and technology company whose combined platforms reach nearly 94 million business decision makers worldwide on a monthly basis. By leveraging proprietary databases of senior-level executives in the Forbes community, Forbes Insights conducts research on a wide range of topics to position brands as thought leaders and drive stakeholder engagement. Research findings are delivered through a variety of digital, print and live executions, and amplified across Forbes’ social and media platforms.

499 Washington Blvd. Jersey City, NJ 07310 | 212.367.2662 | https://forbes.com/forbes-insights/

EDITORIAL & RESEARCH

Erika Maguire EXECUTIVE EDITORIAL DIRECTOR

Kasia Wandycz Moreno EDITORIAL DIRECTOR

Hugo S. Moreno EDITORIAL DIRECTOR

Ross Gagnon RESEARCH DIRECTOR

Scott McGrath RESEARCH ANALYST

Nick Lansing REPORT AUTHOR

Zehava Pasternak DESIGNER

PROJECT MANAGEMENT

Casey Zonfrilli DIRECTOR, ACCOUNT MANAGEMENT

Tori Kreher PROJECT MANAGER

Brian Lee PROJECT MANAGER

Todd Della Rocca PROJECT MANAGER

SALES

North America Brian McLeod VICE PRESIDENT

[email protected]

Matthew Muszala EXECUTIVE DIRECTOR

[email protected]

William Thompson DIRECTOR

[email protected]

Kimberly Kurata MANAGER

[email protected]

Europe Charles Yardley SVP MANAGING DIRECTOR

[email protected]

Asia Will Adamopoulos PRESIDENT & PUBLISHER, FORBES ASIA [email protected]