Disambiguation of morpho-syntactic features of African American English – the case of habitual "be"

Abstract

Recent research has highlighted that natural language processing (NLP) systems exhibit a bias against African American speakers. The bias errors are often caused by poor representation of linguistic features unique to African American English (AAE), due to the relatively low probability of occurrence of many such features in training data. We present a workflow to overcome such bias in the case of habitual “be”. Habitual “be” is isomorphic, and therefore ambiguous, with other forms of “be” found in both AAE and other varieties of English. This creates a clear challenge for bias in NLP technologies. To overcome the scarcity, we employ a combination of rule-based filters and data augmentation that generate a corpus balanced between habitual and non-habitual instances. With this balanced corpus, we train unbiased machine learning classifiers, as demonstrated on a corpus of AAE transcribed texts, achieving .65 F1 score disambiguating habitual “be”.

Publication
Proceedings of the Second Workshop on Language Technology for Equality, Diversity and Inclusion