6 tips for facilitating ethical AI in the enterprise

Artificial intelligence (AI) is potentially the single most disruptive technology of the digital era, as enterprises explore ways to harness machine learning (ML) and other AI tools to mine customer insights, identify talent and corporate networks. And while IT departments can quickly roll out and benefit from most technologies, evidence suggests that CIOs should exercise extreme caution when implementing AI, including employing technologies with strong ethical considerations.

The reason? AI suffers from a big bias problem. In one example, Amazon.com scrapped a recruiting tool after it failed to fairly rate women for developer jobs and other technical posts. In another example, MIT and University of Toronto researchers found that the company’s facial mistakes women, especially those with dark skin, for men.

Biases abound in AI

Amazon.com is not alone, as AI issues have surfaced  in other companies and across other high-stakes domains. A Facebook program manager encountered algorithmic discrimination while testing the company’s Portal video chat device. ProPublica showed that software used across the U.S. to predict future criminals was biased against African-Americans. A study of fintechs by UC Berkeley found that both face-to-face decisions and algorithms used in mortgage lending charged Latinx/African-American borrowers higher interest rates.

You might also like More from author

Comments are closed.