Please use this identifier to cite or link to this item: http://hdl.handle.net/2445/213461
Title: Towards fair machine learning in healthcare: ensuring non-discrimination for disease prediction
Author: Herron Mulet, Claudia
Director/Tutor: Gkontra, Polyxeni
Lekadir, Karim, 1977-
Keywords: Aprenentatge automàtic
Intel·ligència artificial en medicina
Pronòstic mèdic
Treballs de fi de màster
Machine learning
Medical artificial intelligence
Prognosis
Master's thesis
Issue Date: 30-Jun-2023
Abstract: Over the past few years, there has been a rise in the utilization of information and communication technologies (ICTs) and electronic health records (EHRs) within the healthcare system. This increase has led to a substantial gathering of medical data, opening up promising prospects for personalized medicine. Notably, one promising application is the creation of disease risk assessment tools, designed to precisely estimate an individual’s predisposition to developing certain illnessess. These innovative tools empower healthcare professionals to conduct more targeted trials, closely monitor high-risk subjects, and implement timely interventions. However, as these systems start to be tested in real world scenarios, recent studies reveal that they might worsen off the situation of historically underprivileged groups in our society. These discriminatory biases might be caused by many reasons: unequal access to healthcare, false beliefs about biological differences, non-diverse datasets, ma- chine learning (ML) models optimizing for the majority and disregarding underrepresented communities, etc. As a result, it becomes crucial to design and implement metrics and techniques to quantify and mitigate discriminatory biases. In this work, we propose a comprehensive methodology that encompasses data wrangling, model evaluation, and the monitoring of both model performance and potential disparities. Building upon existing research on fairness in machine learning, we aim to adapt the fairness framework specifically for disease prediction, considering that some of the protected features also contribute to increased disease risk. Furthermore, we apply both in-processing and post-processing mitigation techniques to a classifier trained on a large-scale dataset. By experimenting with two diseases of increasing prevalence, Primary Hypertension and Parkinson’s Disease, we seek to assess the effectiveness of these techniques in reducing discriminatory biases and ensuring equitable outcomes.
Note: Treballs finals del Màster de Fonaments de Ciència de Dades, Facultat de matemàtiques, Universitat de Barcelona. Curs: 2022-2023. Tutor: Polyxeni Gkontra i Karim Lekadir
URI: http://hdl.handle.net/2445/213461
Appears in Collections:Màster Oficial - Fonaments de la Ciència de Dades
Programari - Treballs de l'alumnat

Files in This Item:
File Description SizeFormat 
codi_font_herron_mulet_claudia.zipCodi font155.04 kBzipView/Open
tfm_herron_mulet_claudia.pdfMemòria4.11 MBAdobe PDFView/Open


This item is licensed under a Creative Commons License Creative Commons