As artificial intelligence and machine learning permeate every facet of our lives, machines are coming closer and closer to acquiring human-like language capabilities. Research has uncovered that, without intention, business and moral decisions are being made based on deeply ingrained biases that are obscured within AI/ML learning and language patterns. Credit: Thinkstock Machines are fed mounds and mounds of data to extrapolate, interpret and learn. Unlike humans, algorithms are ill-equipped to consciously counteract learned biases because although we would like to believe AI/ML correlates to human thinking, it really doesn’t. AI/ML has created what we have determined to be the newest industrial revolution by giving computers the ability to interpret human language and without intention, it has learned human biases as well. So, where does the data being used by AI/ML systems come from? Most of this historical data comes from the same type of people who created the algorithms and the programs using the algorithms which until recently has been those socio-economically above average and male. So “without thinking” or intent, gender and racial biases have dominated the AI/ML learning process. An AI/ML system is not capable of “thinking on its feet” or reversing this bias once it makes a decision. The point is AI/ML systems are bias because humans are innately biased and AI/ML systems are not capable of moral decisions only humans are; at least not yet any way. Research has shown recruiting (HR) software is biased Much research shows that as machines are acquiring human-like language capabilities, they are also absorbing deeply ingrained human biases concealed within language patterns. Within recruiting (HR) selection software, this means a resume may not make the “first cut” based on the language and pattern recognition of the resume not based on the skills. As time passes, writing resumes has become an art and science; this alone is a skill belonging to a data scientist coupled with a professional writer; above all someone highly language educated with an analytical mind. How many professional writers are capable of being data scientists? Our educational system needs to address this because I believe that everyone will need to be a highly skilled data scientist or have access to one quickly and easily. Recent research has shown through implicit mathematical word association tests that categorize pleasant word versus unpleasant word associations, human psychological biases in AI/ML systems can be exposed. Words associated to “flowers” versus “insects” have been determined as psychologically more pleasant. Professional results for women are seen with gender biases through the words “female” and “woman” as associated with humanities professions and with the home. On the other hand, “male” and “man” algorithms result in associations with math, science and engineering professions. European American names perceiving to be more Anglo-Saxon were heavily associated with the words “gift” or “happy” while African American names were associated with unpleasant words. Statistically, research shows that even with an identical coefficient of variation (CV) of 50% a European American is still more likely to be interviewed over an African American. “The coefficient of variation (CV) represents the ratio of the standard deviation to the mean, and it is a useful statistic for comparing the degree of variation from one data series to another, even if the means are drastically different from one another.” Because algorithms can potentially show when the algorithm is biased, it suggests that algorithms, explicitly inherit the same social prejudices as the humans who programmed them. It is believed that although a complicated task, it is possible that AI/ML systems can be programmed to address this mathematical bias. Correction is taking place already within companies like Google and Amazon search engines on the web. Machine translations of web searches construct mathematical representations of language in which the meaning of a word is refined into a series of numbers (word vector) based on which other words most frequently appear in correlation. This mathematical approach seems to capture the deep cultural societal language context more accurately over any possible dictionary definition. Can bias in AI/ML be eliminated? How to eliminate inappropriate bias by modifying interpretation is not as easy as you can imagine. Language inference and interpretation is a subtle human trait. It’s typically based on such influences as socio-economic background, gender, education and race…all of which contribute to human biases. How to program algorithms designed to “understand” language, without weakening their interpretive powers, is extremely challenging. To select “only one” most appropriate interpretation and adding it to the decision tree leading to the next “only one” most appropriate interpretation and so on down the decision tree causes algorithms to mimic thinking. What if the first interpretation by AI/ML goes down what we humans believe is the wrong path based on human intellect, cultural and moral laws? Immediate course correction input would be necessary as data accumulates along the decision tree as very minute behavior steps are executed. How do we program in moral and cultural acceptable laws into AI/ML systems? Who decides what those moral and cultural laws are? Amazon, Google, IBM, Microsoft and many others have been evaluating bias within its AI/ML platforms trying to understand both the problem and the solution. Amazon has even stopped using AI/ML software as a recruiting and employment tool. After many years of research what has been determined is that since AI/ML replicate patterns of male engineers who build the AI/ML software and systems, the patterns simulated are of their making. Most major companies are beginning to look at the biases their AI/ML systems have created and are trying to “find a cure.” One suggestion is to have extreme diversity in the AI/ML development team with constant diversity oversight from within. Also suggested is to create an AI/ML supervisory and compliance body to police the systems applying AI/ML diverse course corrections. This AI/ML body of humans would be extremely powerful when empowered and ultimately become our AI/ML moral authority. Are we entering into a science fiction novel similar to Orwell’s “1984”? Don’t you see evidence of a global race for a one-world economic and possible moral authority through AI/ML domination? Whoever dominates the creation and deployment of AI/ML platforms could affect not only small global decisions but major ones as well. Related content opinion AI/ML in education closes the gap on income inequality Education for all is the most profound productive and precise answer to income inequality. History has shown time and time again that education in reading, writing and math is the great economic leveler for every human being throughout the globe. Art By Robin Austin Aug 06, 2019 8 mins Education Industry Technology Industry Machine Learning opinion Smart devices may be hazardous to your health As mobile devices and digital assistants continue to proliferate in all aspects of our lives, the World Health Organization is starting to urge caution, as testing is beginning to show alarming results. By Robin Austin Jul 31, 2019 5 mins Laptops Smartphones Technology Industry opinion Why AI/ML is top of mind with corporate executives From around every corner, from behind every door and through every window our lives are being disrupted by artificial intelligence with embedded machine learning. The escalation and widespread global availability of AI/ML solutions in every possible By Robin Austin Jul 23, 2019 10 mins Healthcare Industry Retail Industry Technology Industry opinion Can private online communities drive corporate cultural change? u201cMany of the things that seem impossible now will become realities of tomorrow,u201d said Walt Disney. We are in the 4th Industrial Revolution, able to invent what our imagination dreams up. But it takes communities working together to achieve wh By Robin Austin Jul 18, 2019 7 mins Technology Industry IT Strategy Social Networking Apps Podcasts Videos Resources Events SUBSCRIBE TO OUR NEWSLETTER From our editors straight to your inbox Get started by entering your email address below. Please enter a valid email address Subscribe