@mastersthesis {xAssens, title = {The Temporal Dimension of Visual Attention Models}, year = {2017}, abstract = {

Program: Bachelor Degree on Telecommunications Science and Technologies (CITTEL)

Grade: A with honours (10.0/10.0)

This thesis explores methodologies for scanpath prediction on images using deep learning\ frameworks.\ As a preliminary step, we analyze the characteristics of the data provided by di erent datasets.\ We then explore the use of Convolutional Neural Networks (CNN) and Long-Short-Term-Memory\ (LSTM) newtworks for scanpath prediction. We observe that these models fail due to the high\ stochastic nature of the data.\ With the gained insight, we propose a novel time-aware visual saliency representation named\ Saliency Volume, that averages scanpaths over multiple observers.\ Next, we explore the SalNet network and adapt it for saliency volume prediction, and we find\ several ways of generating scanpaths from saliency volumes.\ Finally, we ne-tuned our model for scanpaht prediction on 360-degree images and successfully\ submitted it to the Salient360! Challenge from ICME. The source code and models are publicly\ available at https://github.com/massens/saliency-360salient-2017.

The Temporal Dimension of Visual Attention Models from Xavier Giro-i-Nieto
}, author = {Assens, Marc}, editor = {McGuinness, Kevin and Xavier Gir{\'o}-i-Nieto and Noel E. O{\textquoteright}Connor} } @conference {cReyes, title = {Where is my Phone? Personal Object Retrieval from Egocentric Images}, booktitle = {Lifelogging Tools and Applications Workshop in ACM Multimedia}, year = {2016}, month = {10/2016}, publisher = {ACM}, organization = {ACM}, address = {Amsterdam, The Netherlands}, abstract = {

This work presents a retrieval pipeline and evaluation scheme for the problem of finding the last appearance of personal objects in a large dataset of images captured from a wearable camera. Each personal object is modelled by a small set of images that define a query for a visual search engine.The retrieved results are reranked considering the temporal timestamps of the images to increase the relevance of the later detections. Finally, a temporal interleaving of the results is introduced for robustness against false detections. The Mean Reciprocal Rank is proposed as a metric to evaluate this problem. This application could help into developing personal assistants capable of helping users when they do not remember where they left their personal belongings.

}, doi = {http://dx.doi.org/10.1145/2983576.2983582}, url = {http://arxiv.org/abs/1608.08139}, author = {Reyes, Cristian and Mohedano, Eva and McGuinness, Kevin and Noel E. O{\textquoteright}Connor and Xavier Gir{\'o}-i-Nieto} } @article {xVentura, title = {Improving Spatial Codification in Semantic Segmentation (Supplementary Material)}, year = {2015}, month = {09/2015}, abstract = {

This document contains supplementary material for the paper "Improving Spatial Codification in Semantic Segmentation" submitted to ICIP 2015. First, there is a section dedicated to the results obtained by categories when ideal object candidates (ground truth masks) are used. Then, an analysis of the results using CPMC and MCG object candidates also detailed by categories. Finally, visual results for CPMC and MCG are showed.

}, author = {Ventura, C. and Xavier Gir{\'o}-i-Nieto and Ver{\'o}nica Vilaplana and Kevin McGuinness and Marqu{\'e}s, F. and Noel E. O{\textquoteright}Connor} }