Louis-Alexandre Dit Petit-Frere, J. (2022). Visual exploration of indirect biases in natural language processing transformer models [Diploma Thesis, Technische Universität Wien]. reposiTUm. https://doi.org/10.34726/hss.2022.99921
E193 - Institut für Visual Computing and Human-Centered Technology
-
Datum (veröffentlicht):
2022
-
Umfang:
134
-
Keywords:
visual exploration; visual analytics; natural language processing; bias; transformer models
en
Abstract:
In recent years, the importance of Natural Language Processing has been increasing with more and more fields of application. The word representations, such as word embedding or transformer models, used to transcribe the language are trained using large text corpora that may include stereotypes. These stereotypes may be learned by Natural Language Processing algorithms and lead to biases in their results. Extensive research has been performed on the detection, repair and visualization of the biases in the field of Natural Language Processing. Nevertheless, the methods developed so far mostly focus on word embeddings, or direct and binary biases.To fill the research gap regarding multi-class indirect biases learned by transformer models, this thesis proposes new visualisation interfaces to explore indirect and multi-class biases learned by BERT and XLNet models. These visualisations are based on an indirect quantitative method to measure the potential biases encapsulated in transformer models, the Indirect Logarithmic Probability Bias Score. This metric is adapted from an existing one, to enable the investigation of indirect biases. The evaluation of our new indirect method shows that it enables to reveal known biases and to discover new insights which could not be found using the direct method. Moreover, the user study performed on our visualization interfaces demonstrates that the visualizations supports the exploration of multi-class indirect biases, even though improvements may be needed to fully assist the investigation of the sources of the biases.