Please use this identifier to cite or link to this item: https://hdl.handle.net/2445/223909
Full metadata record
DC FieldValueLanguage
dc.contributor.advisorStatuto, Nahuel-
dc.contributor.authorMantilla Carreño, Juan Pablo-
dc.date.accessioned2025-10-28T10:38:00Z-
dc.date.available2025-10-28T10:38:00Z-
dc.date.issued2025-06-10-
dc.identifier.urihttps://hdl.handle.net/2445/223909-
dc.descriptionTreballs Finals de Grau d'Enginyeria Informàtica, Facultat de Matemàtiques, Universitat de Barcelona, Any: 2025, Director: Nahuel Statutoca
dc.description.abstractThe increasing use of machine learning poses significant privacy risks, especially when sensitive data is used, and conventional anonymization methods have proven insufficient. Differential privacy is a rigorous framework for data privacy providing strong mathematical guarantees. The possibility of applying this framework to machine learning solves the privacy problem. We will present the fundamental basis of these concepts to empirically investigate, implement, and analyse two techniques for integrating differential privacy into machine learning pipelines. The first technique, dataset perturbation, involves adding calibrated Gaussian noise directly to the training data and then using any standard machine learning pipeline. The second, gradient perturbation, centers on differentially private stochastic gradient descent, is an approach that injects noise into the gradients during the training phase. For the comparative study, we developed a multi-class classification architecture using a real-world, sensitive medical dataset derived from the MIMIC-IV database. Model performance was evaluated against a non-private baseline, using the appropriate metrics considering our class imbalance, such as Macro F1-score and Macro OVO AUC. The results confirm the trade-off between privacy and utility in the models developed, where higher privacy guarantees consistently result in reduced model utility. For the specific context of this study, gradient perturbation provided a slightly more advantageous model in overall balance of utility and privacy. Ultimately, the thesis provides strong evidence for the feasibility of training useful and formally private machine learning models on real-world medical data, successfully demonstrating a practical "sweet spot" between privacy and performance can be found.ca
dc.format.extent49 p.-
dc.format.mimetypeapplication/pdf-
dc.language.isoengca
dc.rightsmemòria: cc-nc-nd (c) Juan Pablo Mantilla Carreño, 2025-
dc.rightscodi: GPL (c) Juan Pablo Mantilla Carreño, 2025-
dc.rights.urihttp://creativecommons.org/licenses/by-nc-nd/3.0/es/-
dc.rights.urihttp://www.gnu.org/licenses/gpl-3.0.ca.html*
dc.sourceTreballs Finals de Grau (TFG) - Enginyeria Informàtica-
dc.subject.classificationAprenentatge automàticca
dc.subject.classificationProtecció de dadesca
dc.subject.classificationDades massivesca
dc.subject.classificationProgramarica
dc.subject.classificationTreballs de fi de grauca
dc.subject.classificationProcessos gaussiansca
dc.subject.otherMachine learningen
dc.subject.otherData protectionen
dc.subject.otherBig dataen
dc.subject.otherComputer softwareen
dc.subject.otherBachelor's thesesen
dc.subject.otherGaussian processesen
dc.titleDifferentially Private Machine Learning: Implementation and Analysis of Gradient and Dataset Perturbation Techniquesca
dc.typeinfo:eu-repo/semantics/bachelorThesisca
dc.rights.accessRightsinfo:eu-repo/semantics/openAccessca
Appears in Collections:Treballs Finals de Grau (TFG) - Enginyeria Informàtica
Treballs Finals de Grau (TFG) - Matemàtiques
Programari - Treballs de l'alumnat

Files in This Item:
File Description SizeFormat 
tfg_Mantilla_Carreño_Juan_Pablo.pdfMemòria1.37 MBAdobe PDFView/Open
codi.zipCodi font38.56 kBzipView/Open


This item is licensed under a Creative Commons License Creative Commons