Racial Audit of Langiage Models

When trained on public data sources the language model can also learn the biases of the society. I wanted to explore a simple way of auditing a language model for racism. (Read Full Report)

Neural Language Models are an important part of virtual speech based assistants. Once trained, it can assign a calculate which sentences are more likely to be uttered by assigning probabilities to them. I trained a Language Model on publicly available news data and performed a racial audit using a simple technique. Dataset I used consisted of news articles that had word crime in the heading.

Data Set

Results

I trained a simple off the shelf language model on a very small dataset for a few epochs. Subsequently,I performed a racial audit of my model by comparing probabilities of negative sentences associated with different race groups. R1 R2

It can be observed that there is huge discrimination among racial identities, because model kind of learned the crime statistics from the news data.