8 minute read 17 Jul 2020
Reframe your future ai car

How can building trust in AI accelerate your transformation?

Authors
Ben Falk

Director, Global Insights, Research Institute | Global Markets – EY Knowledge

Dad. Gardener. Foodie. Wildlife lover. Social scientist and reformer of capitalism. Studying interactions between technology, markets, and public policy, focusing on AI, data, intangibles, and crypto.

Gautam Jaggi

Director, Global Insights, Research Institute | Global Markets – EY Knowledge

Focus areas include disruption, megatrends, behavioral economics, health, future of work, AI. Passionate about photography, travel and music.

Prianka Srinivasan

Associate Director, Technology Insights, Research Institute | Global Markets – EY Knowledge

Student of the future. Focused on technology disruption, innovation and impact on the human experience. Wife, mother, runner and dog lover with a strong case of wanderlust.

Contributors
8 minute read 17 Jul 2020
Related topics AI Technology Trust

Show resources

  • Bridging AI’s trust gaps report 2020 (pdf)

Ethical concerns around the use of AI are growing as COVID-19 stimulates deployment, but poor governance erodes trust.

Artificial intelligence (AI) offers a transformational potential for both companies and public health authorities. But the pandemic has crystallized risks which seemed distant only several months ago. Imagine an AI algorithm scanning you for signs of COVID-19. Would you trust the AI’s diagnosis or have a trained physician review the results? Are you willing to share personal data so that public health authorities can use advanced algorithms to manage the pandemic? And do you trust industry to do the right thing with that information after the crisis has passed?

Around the world, people are asked to sacrifice privacy to facilitate machine learning-powered track-and-trace public health strategies. AI techniques are used to diagnose patients over the phone, scan medical lung images, estimate the likelihood a patient requires respiratory support, identify public hotspots, forecast the spread of the virus, and design useful diagnostics and treatments. The growing use of this nascently regulated technology in responding to the crisis illustrates its enormous opportunities. It also exposes profound emerging risks.

The trust gap

Bridging AI’s trust gaps: Aligning policymakers and companies (pdf) is a new global EY survey conducted in collaboration with The Future Society (TFS) and reveals ethical gaps that may hinder technological responses to the crisis and beyond.

Show resources

We asked policymakers and companies to identify the most important ethical principles when regulating a range of AI applications, and we found divergent priorities between them across use cases. Disagreements generally focused on four areas: fairness and avoiding bias, innovation, data access, and privacy and data rights.

Relative importance of ethical importance

The biggest absolute gap in ethical priorities concerns “fairness and avoiding bias.” In 7 of 12 use cases — ranging from social credit to facial recognition — policymakers placed this among the three most important principles, while companies failed to do so. The European Commission’s 2020 AI whitepaper (pdf) reinforces this principle as a top policymaker concern globally, reflecting widespread public anxiety.

AI should engender not endanger an individual’s autonomy, defend not desecrate human rights and promote rather than imperil social well-being.

Significant misalignments around fairness and avoiding bias generate substantial market risks, as companies may be deploying products and services poorly aligned to emerging social values and regulatory guidance. To illustrate, the recent social upheaval in the US amid the pandemic demonstrates intense public anger directed against discrimination. Brands deploying biased algorithms risk bearing that wrath.

The second-largest gap was around the principle of innovation, which industry prioritized more strongly in all 12 use cases when compared with policymakers. While perhaps unsurprising, innovating with AI is inextricably linked to the risks created and the risks vary depending on the use case. If personal data is analyzed to build a new product, privacy is usually given up (e.g. “training data”) before the algorithm gets released. If an AI inappropriately modifies someone’s behavior, they may not be able to regain any lost autonomy.

Companies’ stronger relative emphasis on the principle of innovation indicates they may not be fully assessing ethical risks when embarking on transformation projects or during the research and development of AI products and services. This finding is supported by the recent EY global board risk survey which shows just 39% of companies believe their enterprise risk practices adequately manage non-traditional or emergent risks, such as those created by building and deploying new AI applications. Improvements in risk management techniques are needed to balance these risks with the significant benefits offered by AI.

The COVID-19 pandemic throws the third- and fourth-largest areas of divergence in our survey into particularly stark relief; inevitably, there is a trade-off between data access on the one hand, and privacy and data rights on the other. Data is a fundamental input into the production of innovative and new AI algorithms. Training an AI to scan human lung tissue requires people to submit medical records. Identifying clusters to enable track-and-trace means people must share some private location data and contact information.

Policymakers surveyed place privacy and data rights as a primary ethical consideration in the development of AI while many industry professionals, instead, emphasize data access. The misuse of personal data has been a key driver of the recent backlash against technology companies and the pandemic may be aggravating those concerns as invasive applications proliferate. Where a recent survey shows 53% of consumers are willing to share personal data to help monitor and track COVID-19, those preferences may not translate more broadly or continue once the immediate crisis has subsided.

Abrogating principles of privacy and fairness, whether intentional or by accident, imposes significant costs on people. Discriminatory algorithms may further threaten some of society’s most vulnerable populations, including ethnic minorities and women — many of who are disproportionately impacted by the pandemic and related economic fallout. Privacy risks are perhaps most acute for children, who lack the experience needed to protect it themselves, yet are now spending considerably more time online to continue distance learning amid widespread school closures. Policymaker consensus around the principles of privacy and fairness suggests resolving conflicting company ethical priorities is critical to soothe public concerns.

Transformation planning thus needs to consider emerging ethical risks. Trust in new technologies will vary depending on local cultural and political conditions, as well as regulatory and governance regimes. For example, the principle of “purpose limitation” is enshrined in Europe’s General Data Protection Regulation. This rule restricts how personal data collected for one use can be reused for other applications in the future and may encourage greater acceptance of invasive data sharing than in places lacking such restrictions. While responding to the pandemic is an urgent need, gaining efficiency at the expense of weakened freedoms and human rights will not build an enduring public trust. The current low-trust environment translates into weaker adoption, in turn, undermining the vast potential of AI-powered public health applications.

The expectations gap

In addition to ethical misalignments, the survey reveals a large expectations gap. The industry has promised self-regulation for some time, but verifiable progress on the key challenges facing the technology sector, such as disinformation and privacy abuses, remains lacking. Thirty-eight percent of companies continue to believe they will lead any rule-making process. But 84% of policymakers expect state actors to dominate. Such a large divergence in expectations is concerning and indicates some companies will be surprised by the direction and scope of upcoming legislative changes. Legal and compliance risks ensue when the industry isn’t well aligned with policymakers. These risks are heightened when products are rushed to market to meet emergency requirements or when multi-year transformation plans are implemented in a matter of weeks.

Policymakers and companies chart

With the deliberative phase of AI regulation coming to an end, policymakers are now clearly moving beyond principles and toward practice. The COVID-19 pandemic has not slowed this legislative agenda, rather it is providing fresh impetus as AI-powered public health applications multiply. This means the window allowing companies to voluntarily align with the emerging governance framework is closing quickly. Business-as-normal risks a larger adjustment burden when stricter governance arrives soon.

Mitigating these risks before they materialize requires leadership commitment to building trust; AI applications deployed must reflect company values. AI should engender not endanger an individual’s autonomy, defend not desecrate human rights and promote rather than imperil social well-being.

Embracing ethical AI

As firms realize the benefits of AI by bringing more products and services to the market, firmer governance will aid in risk management. Appropriate regulations can keep AI safe and secure without violating individuals’ privacy or fundamental freedoms. Better rules will keep businesses fully accountable for the results of their AI innovations. And stronger safeguards will help ensure Trusted AI deployed in the real world is fair, transparent and explainable. This will increase public trust and, therefore, consumer adoption, unlocking new products and strategies more quickly, and turbocharging the use of AI in transformation agendas.

The COVID-19 crisis puts ethical AI in focus, making it a central feature of any digital transformation strategy. The first step toward realigning with the emerging ethics is to foster an internal consensus around key principles across divisions, including technology, data, risk, compliance, legal, sales, HR and management. Engaging a broader set of stakeholders, such as customers who may be discriminated against by AI products and services, promotes trust in a volatile social climate. For example, to minimize the risk of algorithmic bias, businesses should actively build diverse and inclusive training data sets that fairly represent vulnerable groups. In the current social climate, the careless deployment of potentially discriminatory algorithms represents a substantial and growing brand risk. Companies might also consider validating their software with external audits of their production processes, data inputs and algorithmic outputs.

Transparency also helps build the trust needed to drive adoption. Companies should convert ethical principles into clear, published guidelines for supplying AI products and services. Lines of accountability should be formalized. Corporate policies and procedures are required to facilitate regular reviews and ongoing risk assessments, and update systems and products accordingly. Management must provide employees with the resources and training required to reinforce these crucial foundations and to continue learning as the environment evolves. Consulting with policymakers to understand how emerging ethical principles will influence AI regulatory developments in their sector helps ensure ongoing alignment.

Weak consumer trust risks slowing the adoption of transformational, even lifesaving, technologies. As the health crisis expands into a social and economic crisis, ethical AI is becoming a prerequisite to building a better working world.

Summary

A powerful consensus is emerging among policymakers regarding principles for the ethical application of AI. Governments are set to convert the ethics into enforceable rules and regulations, but industry and policymakers disagree on governance approaches and diverge on key principles. Poor alignment diminishes public trust in AI, slowing the adoption of critical applications. New risks are emerging as accelerated digital transformation plans encounter severe public health, social and economic conditions. Closing the AI trust gap is a top priority.

About this article

Authors
Ben Falk

Director, Global Insights, Research Institute | Global Markets – EY Knowledge

Dad. Gardener. Foodie. Wildlife lover. Social scientist and reformer of capitalism. Studying interactions between technology, markets, and public policy, focusing on AI, data, intangibles, and crypto.

Gautam Jaggi

Director, Global Insights, Research Institute | Global Markets – EY Knowledge

Focus areas include disruption, megatrends, behavioral economics, health, future of work, AI. Passionate about photography, travel and music.

Prianka Srinivasan

Associate Director, Technology Insights, Research Institute | Global Markets – EY Knowledge

Student of the future. Focused on technology disruption, innovation and impact on the human experience. Wife, mother, runner and dog lover with a strong case of wanderlust.

Contributors
Related topics AI Technology Trust