Researchers discover method to put an end to gender biases on the internet
[Dec 27, 2022: Anna Sánchez-Juárez P, Universitat Oberta de Catalunya (UOC)]
A well-known recruitment tool that preferred male over female applicants because of algorithms (Credit: Creative Commons)
Endless screeds have been penned on whether the internet algorithms with which we constantly interact suffer from gender bias, and all you need to do is carry out a simple search to see this for yourself. However, according to the researchers behind a new study that seeks to reach a conclusion on this matter, "until now, the debate has not included any scientific analysis".
This new article, by an interdisciplinary team, puts forward a new way of tackling the question and suggests some solutions for preventing these deviances in the data and the discrimination they entail.
Algorithms are being used more and more to decide whether to grant a loan or to accept applications. As the range of uses for artificial intelligence (AI) increases, as do its capabilities and importance, it becomes increasingly vital to assess any possible prejudices associated with these operations.
"Although it's not a new concept, there are many cases in which this problem has not been examined, thus ignoring the potential consequences," stated the researchers, whose study, published open-access in the Algorithms journal, focused mainly on gender bias in the different fields of AI.
Such prejudices can have a huge impact upon society: "Biases affect everything that is discriminated against, excluded or associated with a stereotype. For example, a gender or a race may be excluded in a decision-making process or, simply, certain behaviour may be assumed because of one's gender or the colour of one's skin," explained the principal investigator of the research, Juliana Castañeda Jiménez, an industrial doctorate student at the Universitat Oberta de Catalunya (UOC) under the supervision of Ángel A. Juan, of the Universitat Politècnica de València, and Javier Panadero, of the Universitat Politècnica de Catalunya.
According to Castañeda, "it is possible for algorithmic processes to discriminate by reason of gender, even when programmed to be 'blind' to this variable".
The research team –which also includes researchers Milagros Sáinz and Sergi Yanes, both of the Gender and ICT (GenTIC) research group of the Internet Interdisciplinary Institute (IN3), Laura Calvet, of the Salesian University School of Sarrià, Assumpta Jover, of the Universitat de València, and Ángel A. Juan– illustrate this with a number of examples: the case of a well-known recruitment tool that preferred male over female applicants, or that of some credit services that offered less favourable terms to women than to men.
Figure 1. Scopus-indexed articles for different gender-related terms.
"If old, unbalanced data are used, you're likely to see negative conditioning with regard to black, gay and even female demographics, depending upon when and where the data are from," explained Castañeda.
The sciences are for boys and the arts are for girls
To understand how these patterns are affecting the different algorithms we deal with, the researchers analysed previous works that identified gender biases in data processes in four kinds of AI: those that describe applications in natural language processing and generation, decision management, speech recognition and facial recognition.
Figure 2. Decision-making process and decomposition of algorithms into their characteristics and components. (CREDIT: MDPI)
In general, they found that all the algorithms identified and classified white men better. They also found that they reproduce false beliefs about the physical attributes that should define someone depending upon their biological sex, ethnic or cultural background or sexual orientation, and also that they made stereotypical associations linking men with the sciences and women with the arts.
Many of the procedures used in image and voice recognition are also based on these stereotypes: cameras find it easier to recognize white faces and audio analysis has problems with higher-pitched voices, mainly affecting women.
The cases most likely to suffer from these issues are those whose algorithms are built on the basis of analysing real-life data associated with a specific social context. "Some of the main causes are the under-representation of women in the design and development of AI products and services, and the use of datasets with gender biases," noted the researcher, who argued that the problem stems from the cultural environment in which they are developed.
Figure 3. Socio-technical definition of the algorithm concept. (CREDIT: MDPI)
"An algorithm, when trained with biased data, can detect hidden patterns in society and, when operating, reproduce them. So if, in society, men and women have unequal representation, the design and development of AI products and services will show gender biases."
How can we put an end to this?
The many sources of gender bias, as well as the peculiarities of each given type of algorithm and dataset, mean that doing away with this deviation is a very tough – though not impossible – challenge.
Biased model for P(Yes) vs Score by group. (CREDIT: MDPI)
"Designers and everyone else involved in their design need to be informed of the possibility of the existence of biases associated with an algorithm's logic. What's more, they need to understand the measures available for minimizing, as far as possible, potential biases, and implement them so that they don't occur, because if they are aware of the types of discriminations occurring in society, they will be able to identify when the solutions they develop reproduce them," suggested Castañeda.
This work is innovative because it has been carried out by specialists in different areas, including a sociologist, an anthropologist and experts in gender and statistics. "The team's members provided a perspective that went beyond the autonomous mathematics associated with algorithms, thereby helping us to view them as complex socio-technical systems," said the study's principal investigator.
"If you compare this work with others, I think it is one of only a few that present the issue of biases in algorithms from a neutral standpoint, highlighting both social and technical aspects to identify why an algorithm might make a biased decision," she concluded.
For more science news stories check out our New Discoveries section at The Brighter Side of News.
Note: Materials provided above by Universitat Oberta de Catalunya (UOC). Content may be edited for style and length.
Like these kind of feel good stories? Get the Brighter Side of News' newsletter.