19,123 research outputs found
Evaluating the Underlying Gender Bias in Contextualized Word Embeddings
Gender bias is highly impacting natural language processing applications.
Word embeddings have clearly been proven both to keep and amplify gender biases
that are present in current data sources. Recently, contextualized word
embeddings have enhanced previous word embedding techniques by computing word
vector representations dependent on the sentence they appear in.
In this paper, we study the impact of this conceptual change in the word
embedding computation in relation with gender bias. Our analysis includes
different measures previously applied in the literature to standard word
embeddings. Our findings suggest that contextualized word embeddings are less
biased than standard ones even when the latter are debiased
Measuring and Comparing Social Bias in Static and Contextual Word Embeddings
Word embeddings have been considered one of the biggest breakthroughs of deep learning for natural language processing. They are learned numerical vector representations of words where similar words have similar representations. Contextual word embeddings are the promising second-generation of word embeddings assigning a representation to a word based on its context. This can result in different representations for the same word depending on the context (e.g. river bank and commercial bank). There is evidence of social bias (human-like implicit biases based on gender, race, and other social constructs) in word embeddings. While detecting bias in static (classical or non-contextual) word embeddings is a well-researched topic, there has been limited work in detecting bias in contextual word embeddings, mostly focussed on using the Word Embedding Association Test (WEAT). This paper explores measuring social bias (gender, ethnicity, and religion) in contextual word embeddings using a number of fairness metrics, including the Relative Norm Distance (RND), the Relative Negative Sentiment Bias (RNSB) and the already mentioned WEAT. It extends the Word Embeddings Fairness Evaluation (WEFE) framework to facilitate measuring social biases in contextual embeddings and compares these with biases in static word embeddings. The results show when ranking performance over a number of fairness metrics that contextual word embedding pre-trained models BERT and RoBERTa have more social bias than static word embedding pre-trained models GloVe and Word2Vec
Exploring the Linear Subspace Hypothesis in Gender Bias Mitigation
Bolukbasi et al. (2016) presents one of the first gender bias mitigation
techniques for word embeddings. Their method takes pre-trained word embeddings
as input and attempts to isolate a linear subspace that captures most of the
gender bias in the embeddings. As judged by an analogical evaluation task,
their method virtually eliminates gender bias in the embeddings. However, an
implicit and untested assumption of their method is that the bias sub-space is
actually linear. In this work, we generalize their method to a kernelized,
non-linear version. We take inspiration from kernel principal component
analysis and derive a non-linear bias isolation technique. We discuss and
overcome some of the practical drawbacks of our method for non-linear gender
bias mitigation in word embeddings and analyze empirically whether the bias
subspace is actually linear. Our analysis shows that gender bias is in fact
well captured by a linear subspace, justifying the assumption of Bolukbasi et
al. (2016)
A Causal Inference Method for Reducing Gender Bias in Word Embedding Relations
Word embedding has become essential for natural language processing as it
boosts empirical performances of various tasks. However, recent research
discovers that gender bias is incorporated in neural word embeddings, and
downstream tasks that rely on these biased word vectors also produce
gender-biased results. While some word-embedding gender-debiasing methods have
been developed, these methods mainly focus on reducing gender bias associated
with gender direction and fail to reduce the gender bias presented in word
embedding relations. In this paper, we design a causal and simple approach for
mitigating gender bias in word vector relation by utilizing the statistical
dependency between gender-definition word embeddings and gender-biased word
embeddings. Our method attains state-of-the-art results on gender-debiasing
tasks, lexical- and sentence-level evaluation tasks, and downstream coreference
resolution tasks.Comment: Accepted by AAAI 202
Cultural Differences in Bias? Origin and Gender Bias in Pre-Trained German and French Word Embeddings
Smart applications often rely on training data in form of text. If there is a bias in that training data, the decision of the applications might not be fair. Common training data has been shown to be biased towards different groups of minorities. However, there is no generic algorithm to determine the fairness of training data. One existing approach is to measure gender bias using word embeddings. Most research in this field has been dedicated to the English language. In this work, we identified that there is a bias towards gender and origin in both German and French word embeddings. In particular, we found that real-world bias and stereotypes from the 18th century are still included in today’s word embeddings. Furthermore, we show that the gender bias in German has a different form from English and there is indication that bias has cultural differences that need to be considered when analyzing texts and word embeddings in different languages
- …