Business is in fast pursuit of artificial intelligence (AI) as the next frontier in creating new products and services as well as boosting profits. And this shift reaches far beyond technology firms: companies in sectors from banking to retail to energy, security and healthcare recognize the allure of AI, with new applications emerging every day.
But are companies thinking about the ethical pitfalls around privacy, unintended bias and misuse of data that could stop their innovation in its tracks?
Recent headlines would indicate otherwise. A voice recognition platform was found to have gender bias. A crime prediction algorithm targeted black neighborhoods. And an online ad platform was more likely to show men highly paid executive jobs.
Ethical concerns around bias, fairness, safety, privacy, transparency and accountability associated with AI are growing. With an estimated $1.2 trillion value in 2018 according to Gartner, the AI industry isn’t about to slow down any time soon. In the meantime, a number of concerned voices from within industry, academia and civil society are calling for a more thoughtful approach that balances innovation and responsibility.
Artificial intelligence is capable of a speed and capacity of processing that extends far beyond that of humans, but it cannot be trusted to be fair or neutral. That has led to what AISC described as unintended effects, such as misuse of data leading to privacy intrusion, data and algorithm biases leading to discrimination.
Companies, she said, need to not only look not at how digitalization will improve the bottom line, but also examine the sustainability gains and risks inherent in the technology. “Our mission is to help companies be fair and inclusive and act proactively as these AI systems mature,” Weidman-Grunewald told TriplePundit.
The center aims to develop operational frameworks that identify and address pitfalls and the broader ethical implications of AI and conduct multidisciplinary research. AISC says it wants to guide companies, tech startups, regulators and policymakers to make “human-centric decisions in the AI area.”
AISC brings together companies, academic institutions, public agencies and civil society. Partners include technology investment firm Atomico, media company Bonnier, mobile network operator Telia Company, Microsoft and the KTH Royal Institute of Technology in Sweden.
“We’re at a crossroads where we need to collectively chose a purpose-driven and responsible approach to AI,” AISC co-founder Anna Felländer, a digital economist and former chief economist for Swedbank, told TriplePundit.
"Short-term profits are seductive, yet the unintended pitfalls are costly—both from a financial and societal perspective," she said. "We are convinced that responsible and purpose-driven AI can be combined with profitable business models."
To address these issues, a number of companies are introducing tools to detect bias or developing policies to avoid unintended bias in AI. Last fall IBM launched the Fairness 360 Kit to scan for signs of bias, recommend adjustments, and analyze how and why algorithms make decisions in real time.
Microsoft published an extensive report on “the future computed” last year and has stated the need for principles, policies and laws for the responsible use of AI.
Vahé Torossian, corporate vice president for Microsoft and president of Microsoft Western Europe, says that while the benefits of AI are “significant,” companies need to show “accountability to ensure it is used responsibly and ethically, to the benefit of all. Cooperation across industries has never been more important as this technology becomes increasingly pervasive.”
A group of companies including Google, Microsoft, Amazon, Facebook, Apple and IBM formed the Partnership on AI in 2016 to develop industry best practices to guide AI development.
But are these efforts enough? AISC’s Weidman-Grunewald thinks that what is needed is a robust risk-assessment framework to guide decision-making around AI. That will be one of the first tasks of the center, she said.
“There’s nothing really out there to help companies navigate the sustainability elements of future technology. And that’s especially concerning when you think about how fast the technology is evolving,” she told TriplePundit.
“The ‘move fast and break things’ approach to product development has resulted in too many examples of AI causing harm,” Kristofer Ågren, head of data insights for Division X for Telia Company, told TriplePundit. “We think it is imperative to work proactively with identification and mitigation of risks as we increase the use of AI in our data-driven products, such as Crowd Insights for smart cities."
Crowd Insights are insights from anonymized data designed to help cities better understand their citizens as they seek to improve public transport and urban planning, among other things, according to Telia Company. While Telia says it puts privacy first by anonymizing and aggregating the data based on groups, not individuals, a number of studies have pointed out the privacy risks of compiling mobility data.
“In the end,” Ågren says, “customers and consumers will favor companies that act in their long-term best interests. Working with experts and other partners in AISC will support the development of our skills, processes and tools and allow us to continuously apply the latest knowledge.”
Image credit: GLAS-8/Flickr
Based in southwest Florida, Amy has written about sustainability and the Triple Bottom Line for over 20 years, specializing in sustainability reporting, policy papers and research reports for multinational clients in pharmaceuticals, consumer goods, ICT, tourism and other sectors. She also writes for Ethical Corporation and is a contributor to Creating a Culture of Integrity: Business Ethics for the 21st Century. Connect with Amy on LinkedIn.