Literary Canonicity and Algorithmic Fairness: The Effect of Author Gender on Classification Models

Research output: Contribution to journal/Conference contribution in journal/Contribution to newspaperConference articleResearchpeer-review

4 Downloads (Pure)

Abstract

This study examines gender biases in machine learning models that predict literary canonicity. Using algorithmic fairness metrics like equality of opportunity, equalised odds, and calibration within groups, we show that models violate the fairness metrics, especially by misclassifying non-canonical books by men as canonical. Feature importance analysis shows that text-intrinsic differences between books by men and women authors contribute to these biases. Men have historically dominated canonical literature, which may bias models towards associating men-authored writing styles with literary canonicity. Our study highlights how these biased models can lead to skewed interpretations of literary history and canonicity, potentially reinforcing and perpetuating existing gender disparities in our understanding of literature. This underscores the need to integrate algorithmic fairness in computational literary studies and digital humanities more broadly to foster equitable computational practices.

Original languageEnglish
Article number76
JournalCEUR Workshop Proceedings
Volume3834
Pages (from-to)153-171
Number of pages19
ISSN1613-0073
Publication statusPublished - 2024
Event2024 Computational Humanities Research Conference, CHR 2024 - Aarhus University, Aarhus, Denmark
Duration: 4 Dec 20246 Dec 2024
Conference number: 5
https://2024.computational-humanities-research.org

Conference

Conference2024 Computational Humanities Research Conference, CHR 2024
Number5
LocationAarhus University
Country/TerritoryDenmark
CityAarhus
Period04/12/202406/12/2024
Internet address

Keywords

  • algorithmic fairness
  • bias
  • canonicity
  • computational literary studies
  • gender bias

Fingerprint

Dive into the research topics of 'Literary Canonicity and Algorithmic Fairness: The Effect of Author Gender on Classification Models'. Together they form a unique fingerprint.

Cite this