A systematic review of socio-technical gender bias in AI algorithms
PurposeGender bias in artificial intelligence (AI) should be solved as a priority before AI algorithms become ubiquitous, perpetuating and accentuating the bias. While the problem has been identified as an established research and policy agenda, a cohesive review of existing research specifically ad...
Saved in:
| Published in | Online information review Vol. 47; no. 7; pp. 1264 - 1279 |
|---|---|
| Main Authors | , |
| Format | Journal Article |
| Language | English |
| Published |
Bradford
Emerald Publishing Limited
08.11.2023
Emerald Group Publishing Limited |
| Subjects | |
| Online Access | Get full text |
| ISSN | 1468-4527 1468-4535 |
| DOI | 10.1108/OIR-08-2021-0452 |
Cover
| Summary: | PurposeGender bias in artificial intelligence (AI) should be solved as a priority before AI algorithms become ubiquitous, perpetuating and accentuating the bias. While the problem has been identified as an established research and policy agenda, a cohesive review of existing research specifically addressing gender bias from a socio-technical viewpoint is lacking. Thus, the purpose of this study is to determine the social causes and consequences of, and proposed solutions to, gender bias in AI algorithms.Design/methodology/approachA comprehensive systematic review followed established protocols to ensure accurate and verifiable identification of suitable articles. The process revealed 177 articles in the socio-technical framework, with 64 articles selected for in-depth analysis.FindingsMost previous research has focused on technical rather than social causes, consequences and solutions to AI bias. From a social perspective, gender bias in AI algorithms can be attributed equally to algorithmic design and training datasets. Social consequences are wide-ranging, with amplification of existing bias the most common at 28%. Social solutions were concentrated on algorithmic design, specifically improving diversity in AI development teams (30%), increasing awareness (23%), human-in-the-loop (23%) and integrating ethics into the design process (21%).Originality/valueThis systematic review is the first of its kind to focus on gender bias in AI algorithms from a social perspective within a socio-technical framework. Identification of key causes and consequences of bias and the breakdown of potential solutions provides direction for future research and policy within the growing field of AI ethics.Peer reviewThe peer review history for this article is available at https://publons.com/publon/10.1108/OIR-08-2021-0452 |
|---|---|
| Bibliography: | ObjectType-Article-2 SourceType-Scholarly Journals-1 content type line 14 ObjectType-Feature-3 ObjectType-Evidence Based Healthcare-1 |
| ISSN: | 1468-4527 1468-4535 |
| DOI: | 10.1108/OIR-08-2021-0452 |