Please use this identifier to cite or link to this item: https://hdl.handle.net/2445/215077
Title: Attention mechanisms in transformers: a new formula with mathematical foundations and enhanced interpretability
Author: Conti, Eddie
Director/Tutor: Vieiro Yanes, Arturo
Keywords: Tractament del llenguatge natural (Informàtica)
Processament de dades
Xarxes neuronals (Informàtica)
Treballs de fi de màster
Natural language processing (Computer science)
Data processing
Neural networks (Computer science)
Master's thesis
Issue Date: 9-Jul-2024
Abstract: [en] Large Language Models (LLMs) are AI systems capable of understanding and generating human language by processing vast amounts of text data. In recent years, specifically from 2017, the use of LLMs significantly increased thanks to the introduction of the Transformer architecture.
Note: Treballs finals del Màster de Fonaments de Ciència de Dades, Facultat de matemàtiques, Universitat de Barcelona. Curs: 2023-2024. Tutor: Arturo Vieiro Yanes i Oriol Pujol Vila
URI: https://hdl.handle.net/2445/215077
Appears in Collections:Màster Oficial - Fonaments de la Ciència de Dades
Programari - Treballs de l'alumnat

Files in This Item:
File Description SizeFormat 
tfm_conti_eddie.pdfMemòria1.41 MBAdobe PDFView/Open
Master-Thesis-UB-main.zipCodi font796.03 kBzipView/Open


This item is licensed under a Creative Commons License Creative Commons