dc.contributor.author | Chowdhry, A | |
dc.contributor.author | Aishwaryaprajna | |
dc.contributor.author | Narayan, A | |
dc.contributor.author | Lewis, PR | |
dc.date.accessioned | 2024-09-23T14:19:52Z | |
dc.date.issued | 2024-10-20 | |
dc.date.updated | 2024-09-23T13:18:05Z | |
dc.description.abstract | In recent years, there has been a shift from focusing exclusively on the accuracy of machine learning systems to a more holistic and human-centered approach that includes privacy, fairness, transparency and more. Many of these dimensions are often considered to conflict with each other. For example, there can be a trade-off between the accuracy and fairness of a predictive model. In fairness analysis, the aim is to establish that machine learning models do not discriminate based on protected or sensitive characteristics such as race, gender, age, or religion. In practice there are many alternative notions of fairness, some of which themselves may not be mutually compatible. In this paper, we explore this relationship between accuracy and different notions of fairness using German Credit dataset, where training a model using standard techniques has been shown to lead to biased predictions. We explore the trade-off between accuracy and six different fairness metrics using a multi-objective training approach, which aims to maximize both accuracy and fairness. Our results show that in certain cases, there exists a trade-off between accuracy and different notions of fairness. In these cases, the multi-objective approach provides a set of models that balance the trade-off in different ways. Further, in other cases, the approach does not lead to a trade-off, instead giving rise to a model that is both accurate and fair simultaneously, when this was not achieved using a single-objective approach. Therefore, we show that by explicitly targeting fairness during training, decision makers can have access to a range of models that might meet their accuracy and fairness requirements. Moreover, we also show that a multi-objective approach identifies situations where an assumed trade-off between fairness and accuracy need not exist. | en_GB |
dc.identifier.citation | MODeM 2024: Multi-Objective Decision Making Workshop at ECAI 2024, Santiago de Compostela, Spain, 20 October 2024 | en_GB |
dc.identifier.uri | http://hdl.handle.net/10871/137514 | |
dc.identifier | ORCID: 0000-0003-4386-9745 (Aishwaryaprajna) | |
dc.language.iso | en | en_GB |
dc.publisher | MODeM 2024 | en_GB |
dc.relation.url | https://modem2024.vub.ac.be/ | en_GB |
dc.rights | © 2024 The author(s) | en_GB |
dc.rights | © 2024 the author(s) | |
dc.title | Discovering Trade-offs in Fairness and Accuracy: A Multi-Objective Approach | en_GB |
dc.type | Conference paper | en_GB |
dc.date.available | 2024-09-23T14:19:52Z | |
exeter.location | Santiago de Compostela, Spain | |
dc.description | This is the author accepted manuscript | en_GB |
dc.description | The workshop is co-located with the 27th European Conference on Artificial Intelligence (ECAI 2024) | |
dc.rights.uri | http://www.rioxx.net/licenses/all-rights-reserved | en_GB |
rioxxterms.version | AM | en_GB |
rioxxterms.licenseref.startdate | 2024-10-20 | |
rioxxterms.type | Conference Paper/Proceeding/Abstract | en_GB |
refterms.dateFCD | 2024-09-23T13:18:08Z | |
refterms.versionFCD | AM | |
refterms.dateFOA | 2024-10-23T11:53:42Z | |
refterms.panel | B | en_GB |
pubs.name-of-conference | MODeM 2024 - Multi-Objective Decision Making Workshop at ECAI 2024 | |