Please use this identifier to cite or link to this item: https://hdl.handle.net/2445/179836
Full metadata record
DC FieldValueLanguage
dc.contributor.authorCartas Ayala, Alejandro-
dc.contributor.authorRadeva, Petia-
dc.contributor.authorDimiccoli, Mariella-
dc.date.accessioned2021-09-02T11:05:27Z-
dc.date.available2021-09-02T11:05:27Z-
dc.date.issued2020-04-27-
dc.identifier.issn2169-3536-
dc.identifier.urihttps://hdl.handle.net/2445/179836-
dc.description.abstractActivity recognition from wearable photo-cameras is crucial for lifestyle characterization and health monitoring. However, to enable its wide-spreading use in real-world applications, a high level of generalization needs to be ensured on unseen users. Currently, state-of-the-art methods have been tested only on relatively small datasets consisting of data collected by a few users that are partially seen during training. In this paper, we built a new egocentric dataset acquired by 15 people through a wearable photo-camera and used it to test the generalization capabilities of several state-of-the-art methods for egocentric activity recognition on unseen users and daily image sequences. In addition, we propose several variants to state-of-the-art deep learning architectures, and we show that it is possible to achieve 79.87% accuracy on users unseen during training. Furthermore, to show that the proposed dataset and approach can be useful in real-world applications, where data can be acquired by different wearable cameras and labeled data are scarcely available, we employed a domain adaptation strategy on two egocentric activity recognition benchmark datasets. These experiments show that the model learned with our dataset, can easily be transferred to other domains with a very small amount of labeled data. Taken together, those results show that activity recognition from wearable photo-cameras is mature enough to be tested in real-world applications.-
dc.format.extent20 p.-
dc.format.mimetypeapplication/pdf-
dc.language.isoeng-
dc.publisherInstitute of Electrical and Electronics Engineers (IEEE)-
dc.relation.isformatofReproducció del document publicat a: https://doi.org/10.1109/ACCESS.2020.2990333-
dc.relation.ispartofIEEE Access, 2020, vol. 8, p. 77344-77363-
dc.relation.urihttps://doi.org/10.1109/ACCESS.2020.2990333-
dc.rightscc-by (c) Cartas Ayala, Alejandro et al., 2020-
dc.rights.urihttps://creativecommons.org/licenses/by/4.0/-
dc.sourceArticles publicats en revistes (Matemàtiques i Informàtica)-
dc.subject.classificationAnàlisi de conducta-
dc.subject.classificationSistemes persona-màquina-
dc.subject.otherBehavioral assessment-
dc.subject.otherHuman-machine systems-
dc.titleActivities of Daily Living Monitoring via a WearableCamera: Toward Real-World Applications-
dc.typeinfo:eu-repo/semantics/article-
dc.typeinfo:eu-repo/semantics/publishedVersion-
dc.identifier.idgrec708307-
dc.date.updated2021-09-02T11:05:28Z-
dc.rights.accessRightsinfo:eu-repo/semantics/openAccess-
Appears in Collections:Articles publicats en revistes (Matemàtiques i Informàtica)

Files in This Item:
File Description SizeFormat 
708307.pdf6.42 MBAdobe PDFView/Open


This item is licensed under a Creative Commons License Creative Commons