Please use this identifier to cite or link to this item: http://hdl.handle.net/2445/213461
Full metadata record
DC FieldValueLanguage
dc.contributor.advisorGkontra, Polyxeni-
dc.contributor.advisorLekadir, Karim, 1977--
dc.contributor.authorHerron Mulet, Claudia-
dc.date.accessioned2024-06-20T08:03:09Z-
dc.date.available2024-06-20T08:03:09Z-
dc.date.issued2023-06-30-
dc.identifier.urihttp://hdl.handle.net/2445/213461-
dc.descriptionTreballs finals del Màster de Fonaments de Ciència de Dades, Facultat de matemàtiques, Universitat de Barcelona. Curs: 2022-2023. Tutor: Polyxeni Gkontra i Karim Lekadirca
dc.description.abstractOver the past few years, there has been a rise in the utilization of information and communication technologies (ICTs) and electronic health records (EHRs) within the healthcare system. This increase has led to a substantial gathering of medical data, opening up promising prospects for personalized medicine. Notably, one promising application is the creation of disease risk assessment tools, designed to precisely estimate an individual’s predisposition to developing certain illnessess. These innovative tools empower healthcare professionals to conduct more targeted trials, closely monitor high-risk subjects, and implement timely interventions. However, as these systems start to be tested in real world scenarios, recent studies reveal that they might worsen off the situation of historically underprivileged groups in our society. These discriminatory biases might be caused by many reasons: unequal access to healthcare, false beliefs about biological differences, non-diverse datasets, ma- chine learning (ML) models optimizing for the majority and disregarding underrepresented communities, etc. As a result, it becomes crucial to design and implement metrics and techniques to quantify and mitigate discriminatory biases. In this work, we propose a comprehensive methodology that encompasses data wrangling, model evaluation, and the monitoring of both model performance and potential disparities. Building upon existing research on fairness in machine learning, we aim to adapt the fairness framework specifically for disease prediction, considering that some of the protected features also contribute to increased disease risk. Furthermore, we apply both in-processing and post-processing mitigation techniques to a classifier trained on a large-scale dataset. By experimenting with two diseases of increasing prevalence, Primary Hypertension and Parkinson’s Disease, we seek to assess the effectiveness of these techniques in reducing discriminatory biases and ensuring equitable outcomes.ca
dc.format.extent57 p.-
dc.format.mimetypeapplication/pdf-
dc.language.isoengca
dc.rightscc-by-nc-nd (c) Claudia Herron Mulet, 2023-
dc.rightscodi: MIT (c) Claudia Herron Mulet, 2023-
dc.rights.urihttp://creativecommons.org/licenses/by-nc-nd/3.0/es/*
dc.rights.urihttps://opensource.org/license/mit*
dc.sourceMàster Oficial - Fonaments de la Ciència de Dades-
dc.subject.classificationAprenentatge automàtic-
dc.subject.classificationIntel·ligència artificial en medicina-
dc.subject.classificationPronòstic mèdic-
dc.subject.classificationTreballs de fi de màster-
dc.subject.otherMachine learning-
dc.subject.otherMedical artificial intelligence-
dc.subject.otherPrognosis-
dc.subject.otherMaster's thesis-
dc.titleTowards fair machine learning in healthcare: ensuring non-discrimination for disease predictionca
dc.typeinfo:eu-repo/semantics/masterThesisca
dc.rights.accessRightsinfo:eu-repo/semantics/openAccessca
Appears in Collections:Programari - Treballs de l'alumnat
Màster Oficial - Fonaments de la Ciència de Dades

Files in This Item:
File Description SizeFormat 
codi_font_herron_mulet_claudia.zipCodi font155.04 kBzipView/Open
tfm_herron_mulet_claudia.pdfMemòria4.11 MBAdobe PDFView/Open


This item is licensed under a Creative Commons License Creative Commons