Debunking biases in attention

Shijing Chen, Usman Naseem, Imran Razzak

Research output: Chapter in Book/Report/Conference proceedingConference proceeding contribution

3 Citations (Scopus)

Abstract

Despite the remarkable performances in various applications, machine learning (ML) models could potentially discriminate. They may result in biasness in decision-making, leading to an impact negatively on individuals and society. Recently, various methods have been developed to mitigate biasness and achieve significant performance. Attention mechanisms are a fundamental component of many stateof-the-art ML models and may potentially impact the fairness of ML models. However, how they explicitly influence fairness has yet to be thoroughly explored. In this paper, we investigate how different attention mechanisms affect the fairness of ML models, focusing on models used in Natural Language Processing (NLP) models. We evaluate the performance of fairness of several models with and without different attention mechanisms on widely used benchmark datasets. Our results indicate that the majority of attention mechanisms that have been assessed can improve the fairness performance of Bidirectional Gated Recurrent Unit (BiGRU) and Bidirectional Long ShortTerm Memory (BiLSTM) in all three datasets regarding religious and gender-sensitive groups, however, with varying degrees of trade-offs in accuracy measures. Our findings highlight the possibility of fairness being affected by adopting specific attention mechanisms in machine learning models for certain datasets.

Original languageEnglish
Title of host publicationProceedings of the 3rd Workshop on Trustworthy Natural Language Processing (TrustNLP 2023)
Place of PublicationStroudsburg, PA
PublisherAssociation for Computational Linguistics
Pages141-150
Number of pages10
ISBN (Electronic)9781959429869
DOIs
Publication statusPublished - 2023
Externally publishedYes
Event3rd Workshop on Trustworthy Natural Language Processing, TrustNLP 2023, co-located with ACL 2023 - Toronto, Canada
Duration: 14 Jul 202314 Jul 2023

Conference

Conference3rd Workshop on Trustworthy Natural Language Processing, TrustNLP 2023, co-located with ACL 2023
Country/TerritoryCanada
CityToronto
Period14/07/2314/07/23

Fingerprint

Dive into the research topics of 'Debunking biases in attention'. Together they form a unique fingerprint.

Cite this