El Dipòsit Digital ha actualitzat el programari. Contacteu amb dipositdigital@ub.edu per informar de qualsevol incidència.

 

Differentially Private Machine Learning: Implementation and Analysis of Gradient and Dataset Perturbation Techniques

dc.contributor.advisorStatuto, Nahuel
dc.contributor.authorMantilla Carreño, Juan Pablo
dc.date.accessioned2025-10-28T10:38:00Z
dc.date.available2025-10-28T10:38:00Z
dc.date.issued2025-06-10
dc.descriptionTreballs Finals de Grau d'Enginyeria Informàtica, Facultat de Matemàtiques, Universitat de Barcelona, Any: 2025, Director: Nahuel Statutoca
dc.description.abstractThe increasing use of machine learning poses significant privacy risks, especially when sensitive data is used, and conventional anonymization methods have proven insufficient. Differential privacy is a rigorous framework for data privacy providing strong mathematical guarantees. The possibility of applying this framework to machine learning solves the privacy problem. We will present the fundamental basis of these concepts to empirically investigate, implement, and analyse two techniques for integrating differential privacy into machine learning pipelines. The first technique, dataset perturbation, involves adding calibrated Gaussian noise directly to the training data and then using any standard machine learning pipeline. The second, gradient perturbation, centers on differentially private stochastic gradient descent, is an approach that injects noise into the gradients during the training phase. For the comparative study, we developed a multi-class classification architecture using a real-world, sensitive medical dataset derived from the MIMIC-IV database. Model performance was evaluated against a non-private baseline, using the appropriate metrics considering our class imbalance, such as Macro F1-score and Macro OVO AUC. The results confirm the trade-off between privacy and utility in the models developed, where higher privacy guarantees consistently result in reduced model utility. For the specific context of this study, gradient perturbation provided a slightly more advantageous model in overall balance of utility and privacy. Ultimately, the thesis provides strong evidence for the feasibility of training useful and formally private machine learning models on real-world medical data, successfully demonstrating a practical "sweet spot" between privacy and performance can be found.ca
dc.format.extent49 p.
dc.format.mimetypeapplication/pdf
dc.identifier.urihttps://hdl.handle.net/2445/223909
dc.language.isoengca
dc.rightsmemòria: cc-nc-nd (c) Juan Pablo Mantilla Carreño, 2025
dc.rightscodi: GPL (c) Juan Pablo Mantilla Carreño, 2025
dc.rights.accessRightsinfo:eu-repo/semantics/openAccessca
dc.rights.urihttp://creativecommons.org/licenses/by-nc-nd/3.0/es/
dc.rights.urihttp://www.gnu.org/licenses/gpl-3.0.ca.html*
dc.sourceTreballs Finals de Grau (TFG) - Enginyeria Informàtica
dc.subject.classificationAprenentatge automàticca
dc.subject.classificationProtecció de dadesca
dc.subject.classificationDades massivesca
dc.subject.classificationProgramarica
dc.subject.classificationTreballs de fi de grauca
dc.subject.classificationProcessos gaussiansca
dc.subject.otherMachine learningen
dc.subject.otherData protectionen
dc.subject.otherBig dataen
dc.subject.otherComputer softwareen
dc.subject.otherBachelor's thesesen
dc.subject.otherGaussian processesen
dc.titleDifferentially Private Machine Learning: Implementation and Analysis of Gradient and Dataset Perturbation Techniquesca
dc.typeinfo:eu-repo/semantics/bachelorThesisca

Fitxers

Paquet original

Mostrant 1 - 2 de 2
Carregant...
Miniatura
Nom:
tfg_Mantilla_Carreño_Juan_Pablo.pdf
Mida:
1.34 MB
Format:
Adobe Portable Document Format
Descripció:
Memòria
Carregant...
Miniatura
Nom:
codi.zip
Mida:
38.56 KB
Format:
ZIP file
Descripció:
Codi font