Publications
Here you can find the complete list of my publications. You can use the tag cloud below to select only the papers dealing with specific research topics. You can expand the Abstract, Links and BibTex record of each paper.
2019
Caggianese, Giuseppe; Colonnese, Valerio; Gallo, Luigi
Situated Visualization in Augmented Reality: Exploring Information Seeking Strategies Proceedings Article
In: 2019 15th International Conference on Signal-Image Technology Internet-Based Systems (SITIS), pp. 390–395, 2019.
Abstract | Links | BibTeX | Tags: Augmented Reality, Human computer interaction, Task analysis, Visualization
@inproceedings{caggianeseSituatedVisualizationAugmented2019,
title = {Situated Visualization in Augmented Reality: Exploring Information Seeking Strategies},
author = { Giuseppe Caggianese and Valerio Colonnese and Luigi Gallo},
doi = {10.1109/SITIS.2019.00069},
year = {2019},
date = {2019-11-01},
booktitle = {2019 15th International Conference on Signal-Image Technology Internet-Based Systems (SITIS)},
pages = {390--395},
abstract = {In recent years augmented reality applications have been increasingly demonstrating the requirement for an interaction with information related to and directly shown in the surrounding environment. Situated information is visualized in its semantic and spatial context, building up an environment enhanced by an information level that dynamically adapts to the production of the information and to the actions of the user. The exploration and manipulation of this type of data through see-through augmented reality devices still represents a challenging task. The development of specific interaction strategies capable to mitigating the current limitations of augmented reality devices is essential. In this context, our contribution has been to design possible solutions to address some of these challenges allowing a dynamic interaction with situated information. Following the visual "information-seeking mantra" proposed by Shneiderman and introducing some "superpowers" for the users, in this work we present different strategies aimed at obtaining an overview and filtering, and acquiring details of a collection of situated data.},
keywords = {Augmented Reality, Human computer interaction, Task analysis, Visualization},
pubstate = {published},
tppubtype = {inproceedings}
}
2018
Cascia, Marco La; Vassallo, Giorgio; Gallo, Luigi; Pilato, Giovanni; Vella, Filippo
Automatic Image Annotation Using Random Projection in a Conceptual Space Induced from Data Proceedings Article
In: 2018 14th International Conference on Signal-Image Technology Internet-Based Systems (SITIS), pp. 464–471, 2018.
Abstract | Links | BibTeX | Tags: Feature extraction, Hidden Markov models, Image annotation, Modeling, Semantics, Visualization
@inproceedings{lacasciaAutomaticImageAnnotation2018,
title = {Automatic Image Annotation Using Random Projection in a Conceptual Space Induced from Data},
author = { Marco La Cascia and Giorgio Vassallo and Luigi Gallo and Giovanni Pilato and Filippo Vella},
doi = {10.1109/SITIS.2018.00077},
year = {2018},
date = {2018-11-01},
booktitle = {2018 14th International Conference on Signal-Image Technology Internet-Based Systems (SITIS)},
pages = {464--471},
abstract = {The main drawback of a detailed representation of visual content, whatever is its origin, is that significant features are very high dimensional. To keep the problem tractable while preserving the semantic content, a dimensionality reduction of the data is needed. We propose the Random Projection techniques to reduce the dimensionality. Even though this technique is sub-optimal with respect to Singular Value Decomposition its much lower computational cost make it more suitable for this problem and in particular when computational resources are limited such as in mobile terminals. In this paper we present the use of a ``conceptual'' space, automatically induced from data, to perform automatic image annotation. Images are represented by visual features based on color and texture and arranged as histograms of visual terms and bigrams to partially preserve the spatial information [1]. Using a set of annotated images as training data, the matrix of visual features is built and dimensionality reduction is performed using the Random Projection algorithm. A new unannotated image is then projected into the dimensionally reduced space and the labels of the closest training images are assigned to the unannotated image itself. Experiments on large real collection of images showed that the approach, despite of its low computational cost, is very effective.},
keywords = {Feature extraction, Hidden Markov models, Image annotation, Modeling, Semantics, Visualization},
pubstate = {published},
tppubtype = {inproceedings}
}
2015
Caggianese, Giuseppe; Gallo, Luigi; Neroni, Pietro
User-Driven View Management for Wearable Augmented Reality Systems in the Cultural Heritage Domain Proceedings Article
In: 2015 10th International Conference on P2P, Parallel, Grid, Cloud and Internet Computing (3PGCIC), pp. 545–550, 2015.
Abstract | Links | BibTeX | Tags: Augmented Reality, Context awareness, Cultural heritage, Ego-Vision, Visualization
@inproceedings{caggianeseUserDrivenViewManagement2015,
title = {User-Driven View Management for Wearable Augmented Reality Systems in the Cultural Heritage Domain},
author = { Giuseppe Caggianese and Luigi Gallo and Pietro Neroni},
doi = {10.1109/3PGCIC.2015.90},
year = {2015},
date = {2015-11-01},
booktitle = {2015 10th International Conference on P2P, Parallel, Grid, Cloud and Internet Computing (3PGCIC)},
pages = {545--550},
abstract = {The recent availability of low cost wearable augmented reality (WAR) technologies, is leveraging the design of applications in many different domains in order to support users in their daily activities. For most of these domains, the large amount of information displayable on top of the reality, directly in the user's field of view, represents an important challenge for designers. In this paper we present a view management technique for placing touristic/cultural information, in the form of points of interest (POIs), in an AR system that works in the absence of a priori knowledge of the real environment. The user-driven view management technique, designed as a remote service, improves representation and displacement of the digital information each time the user manifests an interest in a particular area of the real space. The proposed approach includes a layout algorithm, which exploits the user's local position and her/his point of view direction, to correctly set the POI height in the user's view avoiding overlapping and cluttering, together with an adaptive rendering method, using information about the brightness of the area, that computes the visual appearance parameters of each virtual POI in order to improve its readability over the background.},
keywords = {Augmented Reality, Context awareness, Cultural heritage, Ego-Vision, Visualization},
pubstate = {published},
tppubtype = {inproceedings}
}
2014
Gallo, Luigi
Hand Shape Classification Using Depth Data for Unconstrained 3D Interaction Journal Article
In: Journal of Ambient Intelligence and Smart Environments, vol. 6, no. 1, pp. 93–105, 2014, ISSN: 1876-1364.
Abstract | Links | BibTeX | Tags: 3D interaction, Classification, Kinect, Static hand pose recognition, Touchless interaction, Visualization
@article{galloHandShapeClassification2014,
title = {Hand Shape Classification Using Depth Data for Unconstrained 3D Interaction},
author = { Luigi Gallo},
doi = {10.3233/AIS-130239},
issn = {1876-1364},
year = {2014},
date = {2014-01-01},
journal = {Journal of Ambient Intelligence and Smart Environments},
volume = {6},
number = {1},
pages = {93--105},
abstract = {In this paper, we introduce a novel method for view-independent hand pose recognition from depth data. The proposed approach, which does not rely on color information, provides an estimation of the shape and orientation of the user's hand without constraining him/her to maintain a fixed position in the 3D space. We use principal component analysis to estimate the hand orientation in space, Flusser moment invariants as image features and two SVM-RBF classifiers for visual recognition. Moreover, we describe a novel weighting method that takes advantage of the orientation and velocity of the user's hand to assign a score to each hand shape hypothesis. The complete processing chain is described and evaluated in terms of real-time performance and classification accuracy. As a case study, it has also been integrated into a touchless interface for 3D medical visualization, which allows users to manipulate 3D anatomical parts with up to six degrees of freedom. Furthermore, the paper discusses the results of a user study aimed at assessing if using hand velocity as an indicator of the user's intentionality in changing hand posture results in an overall gain in the classification accuracy. The experimental results show that, especially in the presence of out-of-plane rotations of the hand, the introduction of the velocity-based weighting method produces a significant increase in the pose recognition accuracy.},
keywords = {3D interaction, Classification, Kinect, Static hand pose recognition, Touchless interaction, Visualization},
pubstate = {published},
tppubtype = {article}
}
2013
Gallo, Luigi; Placitelli, Alessio Pierluigi
High-Fidelity Visualization of Large Medical Datasets on Commodity Hardware Journal Article
In: ISRN Biomedical Engineering, vol. 2013, no. 892967, pp. 9, 2013, ISSN: 2314-6346.
Abstract | Links | BibTeX | Tags: GPU, Healthcare, Visualization
@article{galloHighfidelityVisualizationLarge2013,
title = {High-Fidelity Visualization of Large Medical Datasets on Commodity Hardware},
author = { Luigi Gallo and Alessio Pierluigi Placitelli},
doi = {10.1155/2013/892967},
issn = {2314-6346},
year = {2013},
date = {2013-01-01},
journal = {ISRN Biomedical Engineering},
volume = {2013},
number = {892967},
pages = {9},
abstract = {Recent advances in CT and MRI static and dynamic scanning techniques have led to great improvements in the resolution and size of volumetric medical datasets, and this trend is still ongoing.However, the explosion of dataset size prevents clinicians fromtaking advantage of an interactive, high-resolution exploration of volumetric medical data on commodity hardware, due to the memory constraints of modern graphics cards.This paper presents a hybrid CPU-GPU volume ray-casting method and some hybrid-based inspection tools aimed at providing interactive, medical-quality visualization using an ordinary desktop PC. Experimental results show that the hybrid method provides a near-interactive high-fidelity visualization of large medical datasets even if only limited hardware resources are available.},
keywords = {GPU, Healthcare, Visualization},
pubstate = {published},
tppubtype = {article}
}
2011
Coronato, Antonio; Gallo, Luigi; Pietro, Giuseppe De
Medical Data Visualization via a Pervasive Multi-Agent Platform Book Section
In: Guah, Matthew (Ed.): Healthcare Delivery Reform and New Technologies: Organizational Initiatives, pp. 331–344, IGI Global, Hershey, PA, USA, 2011.
Abstract | BibTeX | Tags: Healthcare, Visualization
@incollection{coronatoMedicalDataVisualization2011,
title = {Medical Data Visualization via a Pervasive Multi-Agent Platform},
author = { Antonio Coronato and Luigi Gallo and Giuseppe De Pietro},
editor = { Matthew Guah},
year = {2011},
date = {2011-01-01},
booktitle = {Healthcare Delivery Reform and New Technologies: Organizational Initiatives},
pages = {331--344},
publisher = {IGI Global},
address = {Hershey, PA, USA},
abstract = {Pervasive healthcare is the field of application emerging from the combination of healthcare with pervasive computing, which is the computing paradigm that aims to provide users with access to services in a transparent way, wherever they are and whichever their interacting device is. In this paper, a software infrastructure for pervasive healthcare is presented. Such an infrastructure aims at supporting medical practitioners with advanced pervasive access to medical data, which is also context-aware in the sense that the modality to fruit data depends on the device used by the operator and on his physical position within the environment. The paper also describes a service for high quality 3D rendering of medical volume data, which takes advantage of the software infrastructure to distribute the computational load upon the devices available in the environment in a completely transparent way to users.},
keywords = {Healthcare, Visualization},
pubstate = {published},
tppubtype = {incollection}
}
2010
Gallo, Luigi
A Glove-Based Interface for 3D Medical Image Visualization Best Paper Proceedings Article
In: Howlett, Robert J.; Jain, Lakhmi C.; Tsihrintzis, George A.; Damiani, Ernesto; Virvou, Maria; Howlett, Robert J.; Jain, Lakhmi C. (Ed.): Intelligent Interactive Multimedia Systems and Services, pp. 221–230, Springer-Verlag Berlin Heidelberg, Baltimore, USA, 2010, ISBN: 978-3-642-14619-0.
Abstract | Links | BibTeX | Tags: 3D interaction, Data glove, Healthcare, Visualization
@inproceedings{galloGloveBasedInterface3D2010,
title = {A Glove-Based Interface for 3D Medical Image Visualization},
author = { Luigi Gallo},
editor = { Robert J. Howlett and Lakhmi C. Jain and George A. Tsihrintzis and Ernesto Damiani and Maria Virvou and Robert J. Howlett and Lakhmi C. Jain},
doi = {10.1007/978-3-642-14619-0_21},
isbn = {978-3-642-14619-0},
year = {2010},
date = {2010-07-01},
urldate = {2010-07-01},
booktitle = {Intelligent Interactive Multimedia Systems and Services},
volume = {6},
pages = {221--230},
publisher = {Springer-Verlag Berlin Heidelberg},
address = {Baltimore, USA},
series = {Smart Innovation, Systems and Technologies},
abstract = {In this paper, a low cost and portable 3D user interface for exploring medical data is presented. By means of a data glove, equipped with five bend sensors and an accelerometer, and a Wiimote, which tracks additional InfraRed Light Emitting Diodes placed on the glove, 3D imaging data can be visualized and manipulated in a semi-immersive virtual environment. The paper also details the interaction techniques we specifically designed for a medical imaging scenario and provides implementation details of the integration of the interface in an open-source medical image viewer.},
keywords = {3D interaction, Data glove, Healthcare, Visualization},
pubstate = {published},
tppubtype = {inproceedings}
}
2009
Gallo, Luigi; Ciampi, Mario
Wii Remote-enhanced Hand-Computer Interaction for 3D Medical Image Analysis Proceedings Article
In: CTIT '09: Proceedings of International Conference on the Current Trends in Information Technology, pp. 85–90, IEEE Computer Society, Dubai, United Arab Emirates, 2009, ISBN: 978-1-4244-5755-7.
Abstract | Links | BibTeX | Tags: 3D interaction, Data glove, Healthcare, Visualization, Wiimote
@inproceedings{galloWiiRemoteenhancedHandComputer2009,
title = {Wii Remote-enhanced Hand-Computer Interaction for 3D Medical Image Analysis},
author = { Luigi Gallo and Mario Ciampi},
doi = {10.1109/CTIT.2009.5423137},
isbn = {978-1-4244-5755-7},
year = {2009},
date = {2009-12-01},
booktitle = {CTIT '09: Proceedings of International Conference on the Current Trends in Information Technology},
pages = {85--90},
publisher = {IEEE Computer Society},
address = {Dubai, United Arab Emirates},
abstract = {In this paper, we introduce a low cost, not cumbersome, 3D user interface for exploring medical data. 3D imaging data can be manipulated in a semi-immersive virtual environment by means of an off-the-shelf wireless data glove equipped with an additional infrared LED. The LED is tracked by a Wii Remote, which is able to provide accurate positional information with no need for further six degrees of freedom position trackers. We also describe the interaction techniques we have designed specifically for a medical imaging scenario and provide some implementation details of the integration of the interface in an open-source image processing software for medical research.},
keywords = {3D interaction, Data glove, Healthcare, Visualization, Wiimote},
pubstate = {published},
tppubtype = {inproceedings}
}
Other
Talks
Keynote Talks
- 2023, October 10 – Artificial Intelligence and Virtual Reality Advances and Applications in Research Oncology and Clinical Oncology, Next Oncology 2023, Next Oncology – Supporting Oncology through innovation, Milano, Italy. Website
- 2022, March 1 – Touchless interactions in Surgery, ICCI 2022, International Conference on Cybernetics and Innovations, Ratchaburi, Thailand. Website
- 2019, June 25 – Interactive Virtual Environments: From the Laboratory to the Field, SalentoAVR 2019, 6th International Conference on Augmented and Virtual Reality, Lecce, Italy. Website
- 2016, October 2 – Vision-based human-computer interaction in the operating theatre, PRIP 2016, 13th International Conference on Pattern Recognition and Information Processing, Minsk, Republic of Belarus. Website
- 2015, September 1 – Touchless Interaction in Surgery: the Medical Imaging Toolkit experience, SalentoAVR 2015, 2nd International Conference on Augmented and Virtual Reality, Lecce, Italy. Website
Invited Talks
- 2023, November 14 – Realtà Aumentata e Virtuale, La Scienza che non c’era: L’informatica e i prossimi 100 anni del CNR, Area di Ricerca CNR Pisa, Pisa, Italy. Website
- 2023, September 30 – Lessons Learnt from the SMART BEAR Project, STRESS Congress, Palatul Parlamentului, Bucarest, Romania. Host: Prof. Luiza Spiru.
- 2023, May 2 – Enhancing and promoting tangible and intangible cultural heritage: an HCI perspective, PhD Course in “Umanesimo e Tecnologie”, Università di Macerata, Italy. Host: Prof. Roberto Lambertini.
- 2022, December 19 – Sperimentare la realtà estesa e l’interazione uomo-macchina nella riabilitazione personalizzata, Personalized Rehabilitation: Combining Mind, Body and Genetics 2022, Università degli Studi della Campania Luigi Vanvitelli, Napoli, Italia.
- 2022, October 25 – Home Care and Assisted Living for the Elderly: The SMART BEAR Approach, Digital Transformation Summit 2022, Funchal, Madeira, Portugal. Website
- 2022, September 9 – Enhancing and promoting tangible and intangible cultural heritage: an HCI perspective, eXtended Reality and Artificial Intelligence, International Summer School 2022 on “XR and AI for enhancing cultural and territorial heritage”, Matera, Italy. Host: Prof. Ugo Erra. Website
- 2018, December 14 – The potential of virtual reality in various health care settings: promises and challenges, Artificial Intelligence and Health, Rome, Italy. Host: Prof. Clara Balsano. Website
- 2018, November 15 – Piattaforma DatabencArt – progetto per le scuole della Campania, XXII Edizione della Borsa Mediterranea del Turismo Archeologico, Paestum (SA), Italy. Host: Luisa Franzese. Website
- 2018, September 29 – La realtà virtuale per il training cognitivo, Convention Realtà virtuale in soggetti con impairment cognitivo, Gallarate (VA), Italia. Host: Dr. Marco Predazzi. Website
- 2017, February 15 – Interactive ICT technologies for Cultural Heritage, Museo Archeologico Nazionale di Napoli (MANN), Napoli, Italy. Host: Dr. Paolo Giulierini. Website
- 2014, June 6 – Interfacce gestuali touchless per la visualizzazione di immagini mediche, Chirurgie 2014 – Simulazione nella Formazione, Programmazione e Ricerca in Chirurgia, Napoli, Italy. Host: Prof. Marco De Fazio. Website
- 2013, November 11 – Multi-DOF touchless interaction with 3D medical data, Body Tracking in Healthcare 2013, Microsoft Research, Cambridge, UK. Organisers: Abigail Sellen, Kenton O’Hara, Scarlet Schwiderski-Grosche. Website
- 2008, March – Realistic vs. magic interaction metaphors in virtual environments, Multimedia Techniques for Device and Ambient Intelligence – MTDAI 2008, Mogliano Veneto, Italy. Host: Prof. E. Damiani.
Awards
Best Research Paper Award @ AIxPAC
Best Research Paper Award @ ICT4AWE
Second Place Award - Gesture Demonstration Competition @ ICPR
L. Gallo, A.P. Placitelli, G. De Pietro, “A Kinect NUI for 3D Medical Visualization”, demonstrated at the CHALEARN Gesture Demonstration Competition, in conjunction with the 21st International Conference on Pattern Recognition (ICPR 2012), Tsukuba International Congress Center, Tsukuba Science City, Japan, November 10-11, 2012. Organizers: Isabelle Guyon, Vassilis Athitsos. Judges: Alex Balan, Hugo Jair Escalante, Paul Doliotis, Jeffrey Margolis. ChaLearn Gesture Demonstration Competition website
Best Research Paper Award @ IIMSS
L. Gallo, “A Glove-Based Interface for 3D Medical Image Visualization”, presented at the 3rd International Symposium on Intelligent and Interactive Multimedia: Systems and Services, Baltimore, USA, 28-30 July 2010.
Master Degree & PhD Theses
L. Gallo, “Semi-immersive interactive virtual environments for 3D medical imaging,” University of Naples “Parthenope”, Philosophiæ Doctor degree in Information Engineering, 2010. Tutor: Prof. Luigi Romano. Ph.D. Thesis
L. Gallo, “Distribuzione ed esecuzione automatica di task in griglie pervasive,” University of Naples “Federico II”, Master of Science in Computer Engineering, 2006. Supervisors: Antonio Coronato, Giuseppe De Pietro. M.D. Thesis