We are at the cusp of a new age of AI. The adoption of AI in consumer applications has become more widespread. Yet business adoption of AI has been slower due to two major areas of concern:

  1. The black box problem: AI’s decisions are uninterpretable.
  2. The AI bias problem: AI’s decisions may be biased.

The black box problem occurs when it’s unclear how AI is coming to a decision. This causes fear for decision-makers because they cannot understand how AI is coming up with recommendations. However, not all business AI solutions are black boxes. Moreover, business leaders have used many other technologies in the past that they couldn't explain. Therefore, the black box “problem” is just a convenient scapegoat for decision-makers who are afraid of losing control. Today, the explainable AI (XAI) community are developing many technologies that provide much greater transparency to black box algorithms. So the black box problem will soon become a problem of the past.

However, the bias problem is legitimate and deserves closer examination. It’s only a matter of time until AI applications will become more prevalent in high-stake industries. In these industries, businesses may be relying on AI to help review resumes and automate job interviews, or even determine one’s credit worthiness. Biased decisions from AI could lead to discriminative business practices that ultimately result in a PR crisis and/or huge fines. This could prevent business from even considering the use of such AI, as these industries may have compliance requirements.

How Does AI Become Biased?

To address the AI bias problem, we must first understand where these biases came from. Here, we will use the simple definition that AI is a machine mimicry of human behaviors with two important characteristics:

  1. The ability to automate decisions and subsequent actions.
  2. The ability to learn and improve future decision-making with usage.

Since AI must automate “human” decisions, it must mimic our decision-making processes. AI does this through machine learning (ML), which attempts to recreate our mental model of how the world operates, so biases in AI are a result of the biased models created by ML.

Now, because all ML models must be trained with data to fix their model parameters, an important reason why ML creates biased models is because the data used to train the model is biased. Most of the data used in ML training are generated by humans, from previous decisions they made. Therefore, AI bias is simply a reflection of the inherent biases in our own decision-making processes.

Related Article: What Is Explainable AI?

How Can We Combat Bias in AI?

Knowing this sequence of causality is important, because it provides us with multiple points of attack to address this AI bias problem. Today, most AI practitioners in the industry are treating AI bias as a data problem. To a large extent, if we can fix the biased data, we fix the AI bias problem. In fact, many technologies and startups have been created to discover, monitor, and potentially correct for the biases in the data. However, fixing the biased data is not the only way to address the problem.

There are many ways to deal with biases in AI, depending on the context and use case of the AI in question. Here are three general approaches to the AI bias problem:

  1. Simply acknowledging the biases.
  2. Correcting the biased data.
  3. Fixing the root cause of biases.

Since a detailed exposition of all three would make this article too long, we will cover the first approach today and save the others for future entries of this mini-series on the AI bias problem.

Related Article: 4 Reasons Why Explainable AI Is the Future of AI

Learning Opportunities

What Does Acknowledging the Biases Mean?

The first step to acknowledging the biases is to understand what types of biases exist in the AI under scrutiny. This requires a detailed analysis of both the input and the output of the training process before the AI is used in a deployed production environment. It is important to note that these are not the input data and the output decisions of the AI under normal operation.

The main input to the training process is the training data that is used to determine the parameters of the model in the AI. Training data may be biased because it’s typically collected from a much smaller sample of the target population. The target population includes all potential users or anyone who might be affected by the AI’s decision. It’s crucial to ensure the target population is well represented in the training data. Otherwise, the AI will inherit the bias from the training data and not function properly for groups who are not well represented in the training data.

The output of the training process is the model, so we must establish continuous monitoring of the model in operation. Analytics on the AI’s decision must be parsed to see if all the biased AI decisions can be explained by the biased training data. Emergent biases that are not expected from the training data must be further analyzed to understand how they arose.

Related Article: Choosing the Right AI for Your Business Goals

Prove Your Acknowledgement With Actions

Once we have a good understanding of both the inherited and emergent biases, the question next is what do we do about it. The simplest thing we can do is to do nothing. So how can we know if someone is really acknowledging the biases? They must prove it with actions.

Since all the known biases will still be in the AI, the best way to acknowledge their existence is to ensure this AI is used appropriately and responsibly. That means we must not over-generalize the AI’s applicability beyond the training data. If the AI exhibits biases in certain groups, then it should not be used among those groups. This will essentially prevent the biased decisions from manifesting and will limit the negative impacts of those decisions.

Therefore, acknowledging the biases is more than just passively admitting the existence of biases. It takes efforts to identify the biases, and it requires action to ensure the AI is not used outside of its generalization boundary.


Bias in AI is an emerging issue that is hindering AI adoption in certain high-stake industries. The first step to combatting bias in AI is to acknowledge that there are biases in AI, and take steps to identify them and understand their scope and impact. Finally, to truly acknowledge these biases, we must use AI carefully and make sure its usage is not over-generalized. In my next piece, we will discuss how we can correct the biases in the data.

fa-solid fa-hand-paper Learn how you can join our contributor community.