Back to Glossary
Ethics and Governance

Bias in AI

Bias in AI refers to systematic and repeatable errors in an artificial intelligence system that result in unfair outcomes or skewed predictions. These biases typically emerge when a model reflects existing human prejudices, historical inequities, or flaws in the data used to train it.

Explanation

AI bias is a multifaceted challenge that can be introduced at various stages of the machine learning lifecycle. It often begins with 'Data Bias,' where the training dataset is unrepresentative of the real world or contains historical prejudices (e.g., a hiring tool trained on historical data that favored one gender). It can also manifest as 'Algorithmic Bias' through the design of the model’s objective function or 'Deployment Bias' when a model is used in a context it wasn't designed for. Addressing bias is critical because AI is increasingly used for high-stakes decision-making in sectors like healthcare, criminal justice, and finance. Mitigating these issues involves implementing 'Fairness-aware Machine Learning,' which includes auditing datasets for diversity, applying mathematical fairness constraints during model training, and performing post-hoc evaluations to ensure equitable performance across different demographic groups.

Related Terms