Epistemic consequences of unfair tools

Publikation: Bidrag til tidsskrift/Konferencebidrag i tidsskrift /Bidrag til avisTidsskriftartikelForskningpeer review

Abstract

This paper examines the epistemic consequences of unfair technologies used in digital humanities. We connect bias analysis informed by the field of algorithmic fairness with perspectives on knowledge production in digital humanities. We examine the fairness of Danish Named Entity Recognition tools through an innovative experimental method involving data augmentation and evaluate the performance disparities based on two metrics of algorithmic fairness: calibration within groups; and balance for the positive class. Our results show that only two of the ten tested models comply with the fairness criteria.

From an intersectional perspective, we shed light on how unequal performance across groups can lead to the exclusion and marginalization of certain social groups, leading to voices and experiences being disregarded and silenced. We propose incorporating algorithmic fairness in the selection of tools in digital humanities to help alleviate the risk of perpetuating silence and move towards fairer and more inclusive research.
OriginalsprogEngelsk
TidsskriftDigital Scholarship in the Humanities
Sider (fra-til)1-17
ISSN2055-7671
DOI
StatusUdgivet - 2024

Fingeraftryk

Dyk ned i forskningsemnerne om 'Epistemic consequences of unfair tools'. Sammen danner de et unikt fingeraftryk.

Citationsformater