[Access original publication on EDPB website]
SUPPORT POOL OF EXPERTS PROGRAMME
AI-Complex Algorithms and effective Data Protection Supervision - Bias evaluation by Dr. Kris SHRISHAK
Artificial intelligence (AI) systems are socio-technical systems whose behaviour and outputs can harm people. Bias in AI systems can harm people in various ways. Bias can result from interconnected factors that may together amplify harms such as discrimination (European Union Agency for Fundamental Rights, 2022; Weerts et al., 2023). Mitigating bias in AI systems is important and identifying the sources of bias is the first step in any bias mitigation strategy.
The AI pipeline involves many choices and practices that contribute to biased AI systems. Biased data is just one of the sources of biased AI systems, and understanding its various forms can help to detect and to mitigate the bias. In one application, the lack of representative data might be the source of bias, e.g., medical AI where data from women with heart attacks is less represented than men in the dataset. In another, the proxy variables that embed gender bias might be the problem, e.g., in résumé screening. Increasing the dataset size for women could help in the former case, but not in the latter case.
In addition to bias from data, AI systems can also be biased due to the algorithm and the evaluation. These three sources of bias are discussed next.
When defining the target population, the subgroups with sensitive characteristics should be considered. An AI system built using a dataset collected from a city will only have a small percentage of certain minority groups, say 5%. If the dataset is used as- is, then the outputs of this AI system will be biased against this minority group because they only make up 5% of the dataset and the AI system has relatively less data to learn from about them.
Although much of the discussion around bias focusses on the bias from data, other sources of bias that contribute to discriminatory decisions should not be overlooked. In fact, AI models reflect biased outputs not only due to the datasets but also due to the model itself (Hooker, 2021). Even when the datasets are not biased and are properly sampled, the algorithmic choices can contribute to biased decisions. This includes the choice of objective functions, regularisations, how long the model is trained, and even the choice of statistically biased estimators (Danks & London, 2017).
The various trade-offs made during the design and development process could result in discriminatory outputs. Such trade-offs can include model size and the choice of privacy protection mechanisms (Ferry et al., 2023; Fioretto et al., 2022; Kulynych et al., 2022). Even with Diversity in Faces (DiF) dataset that has broad coverage of facial images, an AI model trained with certain differential privacy techniques disproportionately degrades performance for darker-skinned faces (Bagdasaryan et al., 2019). Furthermore, techniques to compress AI models can disproportionally affect the performance of AI models for people with underrepresented sensitive attributes (Hooker et al., 2020).
The performance of AI systems is evaluated based on many metrics, from accuracy to “fairness”. Such assessments are usually performed against a benchmark, or a test dataset. Evaluation bias arises at this stage because the benchmark itself could contribute to bias.
AI systems can perform extremely well against a specific test dataset, and this test performance may fail to translate into real-world performance due to “overfitting” to the test dataset. This is especially a problem if the test dataset carries over historical, representation or measurement bias.