TL;DR Science Artificial Intelligence Bias

By Thomas P.
February 03, 2021 · 3 minute read

Despite the term 'artificial intelligence,' machines are actually quite primitive. While a computer can process and store information at astounding rates, it turns out that common everyday situations like reading emotions or noticing sarcasm are complicated to code into a machine. Machines and the artificial intelligence programs we create are only as smart as we make them. In this sense, we can consider AI to be similar to children in that we have to teach the computer everything about the world.

Google's AI BERT

Despite the misperception that AI is neutral, it turns out that because AI is made by humans, it is just like them. Recent Artificial Intelligence made by Alphabet (Google's parent company) has come under scrutiny. The company’s AI technology named BERT had apparently learned by reading many Wikipedia articles, century-old books, and newspaper articles. All sources could be bound to have some bias. For example, century-old books could use terms that are not considered politically correct to refer to women or minorities; likewise, Wikipedia articles contain user-generated content. Finally, newspaper articles could have political leanings based on the source. In an article in the New York Times, Robert Munro, a computer scientist, found that when BERT was given 100 random words, 99 were associated with Men instead of Women. It is possible that the aforementioned biased sources, of which many may have been written by men (such as the century-old books), skewed BERT's responses to associate these words with men instead of women.


Another example of racial bias is a program called C.O.M.P.A.S., which was created to evaluate the probability of selected criminals becoming subsequent offenders. Examples of the program’s identification include the following: 

Taking a look at the left picture, the system identified the African-American male as a “high risk” individual, although he had no prior offenses as an adult. On the other hand, the system identified his white counterpart as lower risk, although he had more criminal activity. So what are the implications of this? Given C.O.M.P.A.S. was used to rate selected criminals based on their threat (risk) to their community, it is possible that this could be used to determine either the bail amount in court and possible sentence outcomes. Both of these could determine someone’s time in jail and prison.    


Another example of racially biased facial recognition AI includes a technology called “ImageNet,” which tended to misrecognize images when fed an image of a person from a certain race. For example, ImageNet, when fed pictures of programmers, were more likely to assume that they were white because of the images being racially skewed. This trend of Artificial Intelligence being biased towards people of lighter complexion is even more noticeable when we look at how AI identifies those with different skin colors. Often, the pictures used to train artificial intelligence programs are dominantly white. This leads to the misidentification of other ethnicities. For example, in an extreme case, an AI program accidentally identified an African-American as a Gorilla, which evidently shows the serious consequences of the overuse of white faces. This caused scientists to undertake the task of diversifying the image set to make the software less biased.

And there you have it. Three examples of where AI was not trained effectively and ended up with significant biases in the final products. It is important for scientists and programmers that work on these machines in the future to understand the source of possible biases and work to eliminate them. While AI holds much potential for growth, handling the technology in a responsible way will be key to using the technology for the future.

TL;DR AI has similar biases to those that its creators (humans) have. Developers of this new technology need to be cognizant of the possible limitations of AI when using it to help make decisions.



Did you enjoy this article?

About The Author

Thomas is a high school student at Eastside High.


More on this topic...

TL;DR Science: Carbon Cycle

Carbon is found everywhere; it’s the backbone of life. It’s in plants, animals, the oceans, rocks, the air, and even inside you. So how has carbon made it around the Earth to become part of the deepest rocks and highest points of the atmosphere? In this week’s article, we’ll be covering the carbon cycle as we trace this crucial element’s path around the world.

Every Drop Counts: Installing Smart Showers

As our scarce water supplies are being depleted at faster rates, a new generation of scientists are challenged with coming up with more efficient ways of conserving our remaining water resources. These days, new technologies like Smart Showers are helping people all around the world limit and reduce their water usage. Find out more about these technologies in this week’s article.

Bioethics - Unethical human experimentation 

Science is meant to improve our lives, right; or is it possible that not all scientists may not have the best intentions? Throughout scientific history, there have been an unfortunate number of cases in which the scientific method has been carried out with the best intentions or ethics. In this article, historical examples of unethical human experiments are going to be discussed, and how they are avoided in the modern day.

TL;DR Science: Classification of Animals as it Relates to Humans

Ever wonder why humans are classified the way we are? Check out this week's article for a brief overview of the classification system within the animal kingdom.

Today, 48 Years Ago

In this week’s article: Which properties of space were utilized for human needs in the vacuum? What is the purpose of the Mariner 10 project? What planet has a longer day than a year? What discoveries did the Mariner 10 program make? How does our Solar system look? and much more