AHCI RESEARCH GROUP
Publications
Papers published in international journals,
proceedings of conferences, workshops and books.
OUR RESEARCH
Scientific Publications
How to
You can use the tag cloud to select only the papers dealing with specific research topics.
You can expand the Abstract, Links and BibTex record for each paper.
2020
Scianna, Andrea; Guardia, Marcello La
Processing of 3D Models for Networking of CH in Geomatics Journal Article
In: Communications in Computer and Information Science, vol. 1246, pp. 267–281, 2020, (Publisher: Springer Science and Business Media Deutschland GmbH).
Abstract | Links | BibTeX | Tags: 3-D environments, 3D Environments, 3D Visualization, Cultural heritage, Cultural heritages, Geo-Spatial Informations, Geomatics, Loading configuration, Quality of information, Semantics, Surveying, Three dimensional computer graphics, Virtual fruitions, Virtual Reality, Visualization
@article{scianna_processing_2020,
title = {Processing of 3D Models for Networking of CH in Geomatics},
author = {Andrea Scianna and Marcello La Guardia},
editor = {Vettore A. Troisi S. Parente C.},
url = {https://www.scopus.com/inward/record.uri?eid=2-s2.0-85097414626&doi=10.1007%2f978-3-030-62800-0_21&partnerID=40&md5=4f0cbed228d7f218580ab8042953049d},
doi = {10.1007/978-3-030-62800-0_21},
year = {2020},
date = {2020-01-01},
journal = {Communications in Computer and Information Science},
volume = {1246},
pages = {267–281},
abstract = {In recent times the possibility of reconstruction of complex 3D Cultural Heritage (CH) environments has opened new scenarios for touristic and scientific aims. The different needs for networking or conservation purposes of CH lead to study proper structuring of 3D models. In light of this, a scientific approach has been developed in order to test the networking capabilities, comparing different loading configurations of 3D environments with multiple combinations of 3D models inside them, considering different solutions. This experimentation has been based on WebGL-HTML5 technologies and allowed to discover the true balance between performances of proposed system, the quality of visualization, and the quality of information (geometric and semantic ones) characterizing the 3D visualization of the virtual environment. The present work analyzes all of these parameters in order to find the best combination for the implementation of these models into a virtual 3D Geographic Information System (GIS) environment, based on WebGL technologies and accessible via web. This study could be considered a basic step for the development of interactive geospatial information platforms for the virtual fruition of CH. © 2020, Springer Nature Switzerland AG.},
note = {Publisher: Springer Science and Business Media Deutschland GmbH},
keywords = {3-D environments, 3D Environments, 3D Visualization, Cultural heritage, Cultural heritages, Geo-Spatial Informations, Geomatics, Loading configuration, Quality of information, Semantics, Surveying, Three dimensional computer graphics, Virtual fruitions, Virtual Reality, Visualization},
pubstate = {published},
tppubtype = {article}
}
2019
Caggianese, Giuseppe; Colonnese, Valerio; Gallo, Luigi
Situated Visualization in Augmented Reality: Exploring Information Seeking Strategies Proceedings Article
In: 2019 15th International Conference on Signal-Image Technology Internet-Based Systems (SITIS), pp. 390–395, 2019.
Abstract | Links | BibTeX | Tags: Augmented Reality, Human computer interaction, Task analysis, Visualization
@inproceedings{caggianeseSituatedVisualizationAugmented2019,
title = {Situated Visualization in Augmented Reality: Exploring Information Seeking Strategies},
author = { Giuseppe Caggianese and Valerio Colonnese and Luigi Gallo},
doi = {10.1109/SITIS.2019.00069},
year = {2019},
date = {2019-11-01},
booktitle = {2019 15th International Conference on Signal-Image Technology Internet-Based Systems (SITIS)},
pages = {390--395},
abstract = {In recent years augmented reality applications have been increasingly demonstrating the requirement for an interaction with information related to and directly shown in the surrounding environment. Situated information is visualized in its semantic and spatial context, building up an environment enhanced by an information level that dynamically adapts to the production of the information and to the actions of the user. The exploration and manipulation of this type of data through see-through augmented reality devices still represents a challenging task. The development of specific interaction strategies capable to mitigating the current limitations of augmented reality devices is essential. In this context, our contribution has been to design possible solutions to address some of these challenges allowing a dynamic interaction with situated information. Following the visual "information-seeking mantra" proposed by Shneiderman and introducing some "superpowers" for the users, in this work we present different strategies aimed at obtaining an overview and filtering, and acquiring details of a collection of situated data.},
keywords = {Augmented Reality, Human computer interaction, Task analysis, Visualization},
pubstate = {published},
tppubtype = {inproceedings}
}
Caggianese, Giuseppe; Colonnese, Valerio; Gallo, Luigi
Situated Visualization in Augmented Reality: Exploring Information Seeking Strategies Proceedings Article
In: 2019 15th International Conference on Signal-Image Technology Internet-Based Systems (SITIS), pp. 390–395, 2019.
Abstract | Links | BibTeX | Tags: Augmented Reality, Human computer interaction, Task analysis, Visualization
@inproceedings{caggianese_situated_2019,
title = {Situated Visualization in Augmented Reality: Exploring Information Seeking Strategies},
author = {Giuseppe Caggianese and Valerio Colonnese and Luigi Gallo},
doi = {10.1109/SITIS.2019.00069},
year = {2019},
date = {2019-11-01},
booktitle = {2019 15th International Conference on Signal-Image Technology Internet-Based Systems (SITIS)},
pages = {390–395},
abstract = {In recent years augmented reality applications have been increasingly demonstrating the requirement for an interaction with information related to and directly shown in the surrounding environment. Situated information is visualized in its semantic and spatial context, building up an environment enhanced by an information level that dynamically adapts to the production of the information and to the actions of the user. The exploration and manipulation of this type of data through see-through augmented reality devices still represents a challenging task. The development of specific interaction strategies capable to mitigating the current limitations of augmented reality devices is essential. In this context, our contribution has been to design possible solutions to address some of these challenges allowing a dynamic interaction with situated information. Following the visual "information-seeking mantra" proposed by Shneiderman and introducing some "superpowers" for the users, in this work we present different strategies aimed at obtaining an overview and filtering, and acquiring details of a collection of situated data.},
keywords = {Augmented Reality, Human computer interaction, Task analysis, Visualization},
pubstate = {published},
tppubtype = {inproceedings}
}
Scianna, Andrea; Gaglio, Giuseppe Fulvio; Guardia, Marcello
Augmented reality for cultural heritage: The rebirth of a historical square Proceedings Article
In: P., Macher H. Murtiyoso A. Grussenmeyer (Ed.): International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences - ISPRS Archives, pp. 303–308, International Society for Photogrammetry and Remote Sensing, 2019, (Issue: 2/W17).
Abstract | Links | BibTeX | Tags: 3-D printing, 3D model reconstruction, 3D modeling, 3D Modelling, 3D printers, 3D Printing, Augmented Reality, Cultural heritage, Cultural heritages, Data acquisition, Data handling, Image Reconstruction, Mobile Applications, Photogrammetry, Rapid prototyping, Rapid prototyping technology, Repair, Reverse engineering, Reverse engineering techniques, Surveying instruments, Terrestrial Laser Scanners, Thallium, Three dimensional computer graphics, Virtual heritage, Visualization
@inproceedings{scianna_augmented_2019,
title = {Augmented reality for cultural heritage: The rebirth of a historical square},
author = {Andrea Scianna and Giuseppe Fulvio Gaglio and Marcello Guardia},
editor = {Macher H. Murtiyoso A. Grussenmeyer P.},
url = {https://www.scopus.com/inward/record.uri?eid=2-s2.0-85078130517&doi=10.5194%2fisprs-archives-XLII-2-W17-303-2019&partnerID=40&md5=7c28c9b8bf88d7a1af810aeca9c2cff9},
doi = {10.5194/isprs-archives-XLII-2-W17-303-2019},
year = {2019},
date = {2019-01-01},
booktitle = {International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences - ISPRS Archives},
volume = {42},
pages = {303–308},
publisher = {International Society for Photogrammetry and Remote Sensing},
abstract = {The case study, faced in this paper, arises in the context of Interreg Italia-Malta European project named I-Access, dedicated to the improvement of accessibility to Cultural Heritage (CH). Accessibility considered not only as the demolition of physical architectural barriers, but also as the possibility of fruition of CH through technological tools that can increase its perception and knowledge. Last achievements in photogrammetry and terrestrial laser scanner (TLS) technology offered new methods of data acquisition in the field of CH, giving the possibility of monitoring and processing big data, in the form of point clouds. Ever in this field, reverse engineering techniques and computer graphics are even more used for involving visitors to discover CH, with navigation into 3D reconstructions, empowering the real visualization adding further 3D information through the Augmented Reality (AR). At the same time, recent advances on rapid prototyping technologies grant the automated 3D printing of scaled 3D model reconstructions of real CH elements allowing the tactile fruition of visitors that suffer from visual defects and the connection with 3D AR visualizations. The presented work shows how these technologies could revive an historical square, the Piazza Garraffo in Palermo (Italy), with the virtual insertion of its baroque fountain, originally placed there. The final products of this work are an indoor and an outdoor AR mobile application, that allow the visualization of the historical original asset of the square. This study case shows how the mixing of AR and the rapid prototyping technologies could be useful for the improvement of the fruition of CH. This work could be considered a multidisciplinary experimentation, where different technologies, today still in development, contribute to the same goal aimed at improving the accessibility of the monument for enhancing the fruition of CH. © Authors 2019. CC BY 4.0 License},
note = {Issue: 2/W17},
keywords = {3-D printing, 3D model reconstruction, 3D modeling, 3D Modelling, 3D printers, 3D Printing, Augmented Reality, Cultural heritage, Cultural heritages, Data acquisition, Data handling, Image Reconstruction, Mobile Applications, Photogrammetry, Rapid prototyping, Rapid prototyping technology, Repair, Reverse engineering, Reverse engineering techniques, Surveying instruments, Terrestrial Laser Scanners, Thallium, Three dimensional computer graphics, Virtual heritage, Visualization},
pubstate = {published},
tppubtype = {inproceedings}
}
2018
Cascia, Marco La; Vassallo, Giorgio; Gallo, Luigi; Pilato, Giovanni; Vella, Filippo
Automatic Image Annotation Using Random Projection in a Conceptual Space Induced from Data Proceedings Article
In: 2018 14th International Conference on Signal-Image Technology Internet-Based Systems (SITIS), pp. 464–471, 2018.
Abstract | Links | BibTeX | Tags: Feature extraction, Hidden Markov models, Image annotation, Modeling, Semantics, Visualization
@inproceedings{lacasciaAutomaticImageAnnotation2018,
title = {Automatic Image Annotation Using Random Projection in a Conceptual Space Induced from Data},
author = { Marco La Cascia and Giorgio Vassallo and Luigi Gallo and Giovanni Pilato and Filippo Vella},
doi = {10.1109/SITIS.2018.00077},
year = {2018},
date = {2018-11-01},
booktitle = {2018 14th International Conference on Signal-Image Technology Internet-Based Systems (SITIS)},
pages = {464--471},
abstract = {The main drawback of a detailed representation of visual content, whatever is its origin, is that significant features are very high dimensional. To keep the problem tractable while preserving the semantic content, a dimensionality reduction of the data is needed. We propose the Random Projection techniques to reduce the dimensionality. Even though this technique is sub-optimal with respect to Singular Value Decomposition its much lower computational cost make it more suitable for this problem and in particular when computational resources are limited such as in mobile terminals. In this paper we present the use of a ``conceptual'' space, automatically induced from data, to perform automatic image annotation. Images are represented by visual features based on color and texture and arranged as histograms of visual terms and bigrams to partially preserve the spatial information [1]. Using a set of annotated images as training data, the matrix of visual features is built and dimensionality reduction is performed using the Random Projection algorithm. A new unannotated image is then projected into the dimensionally reduced space and the labels of the closest training images are assigned to the unannotated image itself. Experiments on large real collection of images showed that the approach, despite of its low computational cost, is very effective.},
keywords = {Feature extraction, Hidden Markov models, Image annotation, Modeling, Semantics, Visualization},
pubstate = {published},
tppubtype = {inproceedings}
}
Cascia, Marco La; Vassallo, Giorgio; Gallo, Luigi; Pilato, Giovanni; Vella, Filippo
Automatic Image Annotation Using Random Projection in a Conceptual Space Induced from Data Proceedings Article
In: 2018 14th International Conference on Signal-Image Technology Internet-Based Systems (SITIS), pp. 464–471, 2018.
Abstract | Links | BibTeX | Tags: Feature extraction, Hidden Markov models, Image annotation, Modeling, Semantics, Visualization
@inproceedings{la_cascia_automatic_2018,
title = {Automatic Image Annotation Using Random Projection in a Conceptual Space Induced from Data},
author = {Marco La Cascia and Giorgio Vassallo and Luigi Gallo and Giovanni Pilato and Filippo Vella},
doi = {10.1109/SITIS.2018.00077},
year = {2018},
date = {2018-11-01},
booktitle = {2018 14th International Conference on Signal-Image Technology Internet-Based Systems (SITIS)},
pages = {464–471},
abstract = {The main drawback of a detailed representation of visual content, whatever is its origin, is that significant features are very high dimensional. To keep the problem tractable while preserving the semantic content, a dimensionality reduction of the data is needed. We propose the Random Projection techniques to reduce the dimensionality. Even though this technique is sub-optimal with respect to Singular Value Decomposition its much lower computational cost make it more suitable for this problem and in particular when computational resources are limited such as in mobile terminals. In this paper we present the use of a “conceptual” space, automatically induced from data, to perform automatic image annotation. Images are represented by visual features based on color and texture and arranged as histograms of visual terms and bigrams to partially preserve the spatial information [1]. Using a set of annotated images as training data, the matrix of visual features is built and dimensionality reduction is performed using the Random Projection algorithm. A new unannotated image is then projected into the dimensionally reduced space and the labels of the closest training images are assigned to the unannotated image itself. Experiments on large real collection of images showed that the approach, despite of its low computational cost, is very effective.},
keywords = {Feature extraction, Hidden Markov models, Image annotation, Modeling, Semantics, Visualization},
pubstate = {published},
tppubtype = {inproceedings}
}
2016
Bovier, Fabrizia; Caggianese, Giuseppe; Pietro, Giuseppe De; Gallo, Luigi; Neroni, Pietro
An Interactive 3D Holographic Pyramid for Museum Exhibition Proceedings Article
In: 2016 12th International Conference on Signal-Image Technology Internet-Based Systems (SITIS), pp. 428–434, 2016.
Abstract | Links | BibTeX | Tags: Cultural heritage, Education, Holograms, Prototypes, Three-dimensional displays, Touchless interaction, User interface, Virtual heritage, Virtual Reality, Visualization
@inproceedings{bovierInteractive3DHolographic2016,
title = {An Interactive 3D Holographic Pyramid for Museum Exhibition},
author = { Fabrizia Bovier and Giuseppe Caggianese and Giuseppe De Pietro and Luigi Gallo and Pietro Neroni},
doi = {10.1109/SITIS.2016.73},
year = {2016},
date = {2016-11-01},
booktitle = {2016 12th International Conference on Signal-Image Technology Internet-Based Systems (SITIS)},
pages = {428--434},
abstract = {In this paper, an interactive holographic system, realized with the aim of creating, exchanging, discussing and disseminating cultural heritage information, is presented. By using low-cost and off-the-shelf devices, the system provides the visitors with a 'floating' computer generated representation of a virtual cultural artefact that, unlike the real one, can be examined in detail through a touchless natural interface. The proposed system is realized in such a way that it can be easily placed in a cultural exhibition without requiring any structural intervention. As such, it could represent a useful instrument complementary to a museum visit thanks to its capacity both to convey different types of digital cultural information and especially to allow the visitor to become an active actor, able to enjoy different perspectives and all the details of the artefact sharing her/his experience with other visitors. The paper describes the system modules and the hardware design to physically realize the pyramid, and details the user interface composed of two main actions designed to obtain a simple exploration of a virtual cultural heritage artefact.},
keywords = {Cultural heritage, Education, Holograms, Prototypes, Three-dimensional displays, Touchless interaction, User interface, Virtual heritage, Virtual Reality, Visualization},
pubstate = {published},
tppubtype = {inproceedings}
}
Bovier, Fabrizia; Caggianese, Giuseppe; Pietro, Giuseppe De; Gallo, Luigi; Neroni, Pietro
An Interactive 3D Holographic Pyramid for Museum Exhibition Proceedings Article
In: 2016 12th International Conference on Signal-Image Technology Internet-Based Systems (SITIS), pp. 428–434, 2016.
Abstract | Links | BibTeX | Tags: Cultural heritage, Education, Holograms, Prototypes, Three-dimensional displays, Touchless interaction, User interface, Virtual heritage, Virtual Reality, Visualization
@inproceedings{bovier_interactive_2016,
title = {An Interactive 3D Holographic Pyramid for Museum Exhibition},
author = {Fabrizia Bovier and Giuseppe Caggianese and Giuseppe De Pietro and Luigi Gallo and Pietro Neroni},
doi = {10.1109/SITIS.2016.73},
year = {2016},
date = {2016-11-01},
booktitle = {2016 12th International Conference on Signal-Image Technology Internet-Based Systems (SITIS)},
pages = {428–434},
abstract = {In this paper, an interactive holographic system, realized with the aim of creating, exchanging, discussing and disseminating cultural heritage information, is presented. By using low-cost and off-the-shelf devices, the system provides the visitors with a 'floating' computer generated representation of a virtual cultural artefact that, unlike the real one, can be examined in detail through a touchless natural interface. The proposed system is realized in such a way that it can be easily placed in a cultural exhibition without requiring any structural intervention. As such, it could represent a useful instrument complementary to a museum visit thanks to its capacity both to convey different types of digital cultural information and especially to allow the visitor to become an active actor, able to enjoy different perspectives and all the details of the artefact sharing her/his experience with other visitors. The paper describes the system modules and the hardware design to physically realize the pyramid, and details the user interface composed of two main actions designed to obtain a simple exploration of a virtual cultural heritage artefact.},
keywords = {Cultural heritage, Education, Holograms, Prototypes, Three-dimensional displays, Touchless interaction, User interface, Virtual heritage, Virtual Reality, Visualization},
pubstate = {published},
tppubtype = {inproceedings}
}
2015
Caggianese, Giuseppe; Gallo, Luigi; Neroni, Pietro
User-Driven View Management for Wearable Augmented Reality Systems in the Cultural Heritage Domain Proceedings Article
In: 2015 10th International Conference on P2P, Parallel, Grid, Cloud and Internet Computing (3PGCIC), pp. 545–550, 2015.
Abstract | Links | BibTeX | Tags: Augmented Reality, Context awareness, Cultural heritage, Ego-Vision, Visualization
@inproceedings{caggianeseUserDrivenViewManagement2015,
title = {User-Driven View Management for Wearable Augmented Reality Systems in the Cultural Heritage Domain},
author = { Giuseppe Caggianese and Luigi Gallo and Pietro Neroni},
doi = {10.1109/3PGCIC.2015.90},
year = {2015},
date = {2015-11-01},
booktitle = {2015 10th International Conference on P2P, Parallel, Grid, Cloud and Internet Computing (3PGCIC)},
pages = {545--550},
abstract = {The recent availability of low cost wearable augmented reality (WAR) technologies, is leveraging the design of applications in many different domains in order to support users in their daily activities. For most of these domains, the large amount of information displayable on top of the reality, directly in the user's field of view, represents an important challenge for designers. In this paper we present a view management technique for placing touristic/cultural information, in the form of points of interest (POIs), in an AR system that works in the absence of a priori knowledge of the real environment. The user-driven view management technique, designed as a remote service, improves representation and displacement of the digital information each time the user manifests an interest in a particular area of the real space. The proposed approach includes a layout algorithm, which exploits the user's local position and her/his point of view direction, to correctly set the POI height in the user's view avoiding overlapping and cluttering, together with an adaptive rendering method, using information about the brightness of the area, that computes the visual appearance parameters of each virtual POI in order to improve its readability over the background.},
keywords = {Augmented Reality, Context awareness, Cultural heritage, Ego-Vision, Visualization},
pubstate = {published},
tppubtype = {inproceedings}
}
Caggianese, Giuseppe; Gallo, Luigi; Neroni, Pietro
User-Driven View Management for Wearable Augmented Reality Systems in the Cultural Heritage Domain Proceedings Article
In: 2015 10th International Conference on P2P, Parallel, Grid, Cloud and Internet Computing (3PGCIC), pp. 545–550, 2015.
Abstract | Links | BibTeX | Tags: Augmented Reality, Context awareness, Cultural heritage, Ego-Vision, Visualization
@inproceedings{caggianese_user-driven_2015,
title = {User-Driven View Management for Wearable Augmented Reality Systems in the Cultural Heritage Domain},
author = {Giuseppe Caggianese and Luigi Gallo and Pietro Neroni},
doi = {10.1109/3PGCIC.2015.90},
year = {2015},
date = {2015-11-01},
booktitle = {2015 10th International Conference on P2P, Parallel, Grid, Cloud and Internet Computing (3PGCIC)},
pages = {545–550},
abstract = {The recent availability of low cost wearable augmented reality (WAR) technologies, is leveraging the design of applications in many different domains in order to support users in their daily activities. For most of these domains, the large amount of information displayable on top of the reality, directly in the user's field of view, represents an important challenge for designers. In this paper we present a view management technique for placing touristic/cultural information, in the form of points of interest (POIs), in an AR system that works in the absence of a priori knowledge of the real environment. The user-driven view management technique, designed as a remote service, improves representation and displacement of the digital information each time the user manifests an interest in a particular area of the real space. The proposed approach includes a layout algorithm, which exploits the user's local position and her/his point of view direction, to correctly set the POI height in the user's view avoiding overlapping and cluttering, together with an adaptive rendering method, using information about the brightness of the area, that computes the visual appearance parameters of each virtual POI in order to improve its readability over the background.},
keywords = {Augmented Reality, Context awareness, Cultural heritage, Ego-Vision, Visualization},
pubstate = {published},
tppubtype = {inproceedings}
}
2014
Gallo, Luigi
Hand Shape Classification Using Depth Data for Unconstrained 3D Interaction Journal Article
In: Journal of Ambient Intelligence and Smart Environments, vol. 6, no. 1, pp. 93–105, 2014, ISSN: 1876-1364.
Abstract | Links | BibTeX | Tags: 3D interaction, Classification, Kinect, Static hand pose recognition, Touchless interaction, Visualization
@article{galloHandShapeClassification2014,
title = {Hand Shape Classification Using Depth Data for Unconstrained 3D Interaction},
author = { Luigi Gallo},
doi = {10.3233/AIS-130239},
issn = {1876-1364},
year = {2014},
date = {2014-01-01},
journal = {Journal of Ambient Intelligence and Smart Environments},
volume = {6},
number = {1},
pages = {93--105},
abstract = {In this paper, we introduce a novel method for view-independent hand pose recognition from depth data. The proposed approach, which does not rely on color information, provides an estimation of the shape and orientation of the user's hand without constraining him/her to maintain a fixed position in the 3D space. We use principal component analysis to estimate the hand orientation in space, Flusser moment invariants as image features and two SVM-RBF classifiers for visual recognition. Moreover, we describe a novel weighting method that takes advantage of the orientation and velocity of the user's hand to assign a score to each hand shape hypothesis. The complete processing chain is described and evaluated in terms of real-time performance and classification accuracy. As a case study, it has also been integrated into a touchless interface for 3D medical visualization, which allows users to manipulate 3D anatomical parts with up to six degrees of freedom. Furthermore, the paper discusses the results of a user study aimed at assessing if using hand velocity as an indicator of the user's intentionality in changing hand posture results in an overall gain in the classification accuracy. The experimental results show that, especially in the presence of out-of-plane rotations of the hand, the introduction of the velocity-based weighting method produces a significant increase in the pose recognition accuracy.},
keywords = {3D interaction, Classification, Kinect, Static hand pose recognition, Touchless interaction, Visualization},
pubstate = {published},
tppubtype = {article}
}
Gallo, Luigi
Hand shape classification using depth data for unconstrained 3D interaction Journal Article
In: Journal of Ambient Intelligence and Smart Environments, vol. 6, no. 1, pp. 93–105, 2014, ISSN: 1876-1364.
Abstract | Links | BibTeX | Tags: 3D interaction, Classification, Kinect, Static hand pose recognition, Touchless interaction, Visualization
@article{gallo_hand_2014,
title = {Hand shape classification using depth data for unconstrained 3D interaction},
author = {Luigi Gallo},
doi = {10.3233/AIS-130239},
issn = {1876-1364},
year = {2014},
date = {2014-01-01},
journal = {Journal of Ambient Intelligence and Smart Environments},
volume = {6},
number = {1},
pages = {93–105},
abstract = {In this paper, we introduce a novel method for view-independent hand pose recognition from depth data. The proposed approach, which does not rely on color information, provides an estimation of the shape and orientation of the user's hand without constraining him/her to maintain a fixed position in the 3D space. We use principal component analysis to estimate the hand orientation in space, Flusser moment invariants as image features and two SVM-RBF classifiers for visual recognition. Moreover, we describe a novel weighting method that takes advantage of the orientation and velocity of the user's hand to assign a score to each hand shape hypothesis. The complete processing chain is described and evaluated in terms of real-time performance and classification accuracy. As a case study, it has also been integrated into a touchless interface for 3D medical visualization, which allows users to manipulate 3D anatomical parts with up to six degrees of freedom. Furthermore, the paper discusses the results of a user study aimed at assessing if using hand velocity as an indicator of the user's intentionality in changing hand posture results in an overall gain in the classification accuracy. The experimental results show that, especially in the presence of out-of-plane rotations of the hand, the introduction of the velocity-based weighting method produces a significant increase in the pose recognition accuracy.},
keywords = {3D interaction, Classification, Kinect, Static hand pose recognition, Touchless interaction, Visualization},
pubstate = {published},
tppubtype = {article}
}
2013
Gallo, Luigi; Placitelli, Alessio Pierluigi
High-Fidelity Visualization of Large Medical Datasets on Commodity Hardware Journal Article
In: ISRN Biomedical Engineering, vol. 2013, no. 892967, pp. 9, 2013, ISSN: 2314-6346.
Abstract | Links | BibTeX | Tags: GPU, Healthcare, Visualization
@article{galloHighfidelityVisualizationLarge2013,
title = {High-Fidelity Visualization of Large Medical Datasets on Commodity Hardware},
author = { Luigi Gallo and Alessio Pierluigi Placitelli},
doi = {10.1155/2013/892967},
issn = {2314-6346},
year = {2013},
date = {2013-01-01},
journal = {ISRN Biomedical Engineering},
volume = {2013},
number = {892967},
pages = {9},
abstract = {Recent advances in CT and MRI static and dynamic scanning techniques have led to great improvements in the resolution and size of volumetric medical datasets, and this trend is still ongoing.However, the explosion of dataset size prevents clinicians fromtaking advantage of an interactive, high-resolution exploration of volumetric medical data on commodity hardware, due to the memory constraints of modern graphics cards.This paper presents a hybrid CPU-GPU volume ray-casting method and some hybrid-based inspection tools aimed at providing interactive, medical-quality visualization using an ordinary desktop PC. Experimental results show that the hybrid method provides a near-interactive high-fidelity visualization of large medical datasets even if only limited hardware resources are available.},
keywords = {GPU, Healthcare, Visualization},
pubstate = {published},
tppubtype = {article}
}
Gallo, Luigi; Placitelli, Alessio Pierluigi
High-fidelity visualization of large medical datasets on commodity hardware Journal Article
In: ISRN Biomedical Engineering, vol. 2013, no. 892967, pp. 9, 2013, ISSN: 2314-6346.
Abstract | Links | BibTeX | Tags: GPU, Healthcare, Visualization
@article{gallo_high-fidelity_2013,
title = {High-fidelity visualization of large medical datasets on commodity hardware},
author = {Luigi Gallo and Alessio Pierluigi Placitelli},
doi = {10.1155/2013/892967},
issn = {2314-6346},
year = {2013},
date = {2013-01-01},
journal = {ISRN Biomedical Engineering},
volume = {2013},
number = {892967},
pages = {9},
abstract = {Recent advances in CT and MRI static and dynamic scanning techniques have led to great improvements in the resolution and size of volumetric medical datasets, and this trend is still ongoing.However, the explosion of dataset size prevents clinicians fromtaking advantage of an interactive, high-resolution exploration of volumetric medical data on commodity hardware, due to the memory constraints of modern graphics cards.This paper presents a hybrid CPU-GPU volume ray-casting method and some hybrid-based inspection tools aimed at providing interactive, medical-quality visualization using an ordinary desktop PC. Experimental results show that the hybrid method provides a near-interactive high-fidelity visualization of large medical datasets even if only limited hardware resources are available.},
keywords = {GPU, Healthcare, Visualization},
pubstate = {published},
tppubtype = {article}
}
2011
Coronato, Antonio; Gallo, Luigi; Pietro, Giuseppe De
Medical Data Visualization via a Pervasive Multi-Agent Platform Book Section
In: Guah, Matthew (Ed.): Healthcare Delivery Reform and New Technologies: Organizational Initiatives, pp. 331–344, IGI Global, Hershey, PA, USA, 2011.
Abstract | BibTeX | Tags: Healthcare, Visualization
@incollection{coronatoMedicalDataVisualization2011,
title = {Medical Data Visualization via a Pervasive Multi-Agent Platform},
author = { Antonio Coronato and Luigi Gallo and Giuseppe De Pietro},
editor = { Matthew Guah},
year = {2011},
date = {2011-01-01},
booktitle = {Healthcare Delivery Reform and New Technologies: Organizational Initiatives},
pages = {331--344},
publisher = {IGI Global},
address = {Hershey, PA, USA},
abstract = {Pervasive healthcare is the field of application emerging from the combination of healthcare with pervasive computing, which is the computing paradigm that aims to provide users with access to services in a transparent way, wherever they are and whichever their interacting device is. In this paper, a software infrastructure for pervasive healthcare is presented. Such an infrastructure aims at supporting medical practitioners with advanced pervasive access to medical data, which is also context-aware in the sense that the modality to fruit data depends on the device used by the operator and on his physical position within the environment. The paper also describes a service for high quality 3D rendering of medical volume data, which takes advantage of the software infrastructure to distribute the computational load upon the devices available in the environment in a completely transparent way to users.},
keywords = {Healthcare, Visualization},
pubstate = {published},
tppubtype = {incollection}
}
Coronato, Antonio; Gallo, Luigi; Pietro, Giuseppe De
Medical Data Visualization via a Pervasive Multi-Agent Platform Book Section
In: Guah, Matthew (Ed.): Healthcare Delivery Reform and New Technologies: Organizational Initiatives, pp. 331–344, IGI Global, Hershey, PA, USA, 2011.
Abstract | Links | BibTeX | Tags: Healthcare, Visualization
@incollection{coronato_medical_2011,
title = {Medical Data Visualization via a Pervasive Multi-Agent Platform},
author = {Antonio Coronato and Luigi Gallo and Giuseppe De Pietro},
editor = {Matthew Guah},
url = {http://dx.doi.org/10.4018/978-1-60960-183-6.ch019},
year = {2011},
date = {2011-01-01},
booktitle = {Healthcare Delivery Reform and New Technologies: Organizational Initiatives},
pages = {331–344},
publisher = {IGI Global},
address = {Hershey, PA, USA},
abstract = {Pervasive healthcare is the field of application emerging from the combination of healthcare with pervasive computing, which is the computing paradigm that aims to provide users with access to services in a transparent way, wherever they are and whichever their interacting device is. In this paper, a software infrastructure for pervasive healthcare is presented. Such an infrastructure aims at supporting medical practitioners with advanced pervasive access to medical data, which is also context-aware in the sense that the modality to fruit data depends on the device used by the operator and on his physical position within the environment. The paper also describes a service for high quality 3D rendering of medical volume data, which takes advantage of the software infrastructure to distribute the computational load upon the devices available in the environment in a completely transparent way to users.},
keywords = {Healthcare, Visualization},
pubstate = {published},
tppubtype = {incollection}
}
2010
Gallo, Luigi
A Glove-Based Interface for 3D Medical Image Visualization Best Paper Proceedings Article
In: Howlett, Robert J.; Jain, Lakhmi C.; Tsihrintzis, George A.; Damiani, Ernesto; Virvou, Maria; Howlett, Robert J.; Jain, Lakhmi C. (Ed.): Intelligent Interactive Multimedia Systems and Services, pp. 221–230, Springer-Verlag Berlin Heidelberg, Baltimore, USA, 2010, ISBN: 978-3-642-14619-0.
Abstract | Links | BibTeX | Tags: 3D interaction, Data glove, Healthcare, Visualization
@inproceedings{galloGloveBasedInterface3D2010,
title = {A Glove-Based Interface for 3D Medical Image Visualization},
author = { Luigi Gallo},
editor = { Robert J. Howlett and Lakhmi C. Jain and George A. Tsihrintzis and Ernesto Damiani and Maria Virvou and Robert J. Howlett and Lakhmi C. Jain},
doi = {10.1007/978-3-642-14619-0_21},
isbn = {978-3-642-14619-0},
year = {2010},
date = {2010-07-01},
urldate = {2010-07-01},
booktitle = {Intelligent Interactive Multimedia Systems and Services},
volume = {6},
pages = {221--230},
publisher = {Springer-Verlag Berlin Heidelberg},
address = {Baltimore, USA},
series = {Smart Innovation, Systems and Technologies},
abstract = {In this paper, a low cost and portable 3D user interface for exploring medical data is presented. By means of a data glove, equipped with five bend sensors and an accelerometer, and a Wiimote, which tracks additional InfraRed Light Emitting Diodes placed on the glove, 3D imaging data can be visualized and manipulated in a semi-immersive virtual environment. The paper also details the interaction techniques we specifically designed for a medical imaging scenario and provides implementation details of the integration of the interface in an open-source medical image viewer.},
keywords = {3D interaction, Data glove, Healthcare, Visualization},
pubstate = {published},
tppubtype = {inproceedings}
}
Gallo, Luigi
A Glove-Based Interface for 3D Medical Image Visualization Proceedings Article
In: Howlett, Robert J.; Jain, Lakhmi C.; Tsihrintzis, George A.; Damiani, Ernesto; Virvou, Maria; Howlett, Robert J.; Jain, Lakhmi C. (Ed.): Intelligent Interactive Multimedia Systems and Services, pp. 221–230, Springer-Verlag Berlin Heidelberg, Baltimore, USA, 2010, ISBN: 978-3-642-14619-0.
Abstract | Links | BibTeX | Tags: 3D interaction, Data glove, Healthcare, Visualization
@inproceedings{gallo_glove-based_2010,
title = {A Glove-Based Interface for 3D Medical Image Visualization},
author = {Luigi Gallo},
editor = {Robert J. Howlett and Lakhmi C. Jain and George A. Tsihrintzis and Ernesto Damiani and Maria Virvou and Robert J. Howlett and Lakhmi C. Jain},
doi = {10.1007/978-3-642-14619-0_21},
isbn = {978-3-642-14619-0},
year = {2010},
date = {2010-07-01},
booktitle = {Intelligent Interactive Multimedia Systems and Services},
volume = {6},
pages = {221–230},
publisher = {Springer-Verlag Berlin Heidelberg},
address = {Baltimore, USA},
series = {Smart Innovation, Systems and Technologies},
abstract = {In this paper, a low cost and portable 3D user interface for exploring medical data is presented. By means of a data glove, equipped with five bend sensors and an accelerometer, and a Wiimote, which tracks additional InfraRed Light Emitting Diodes placed on the glove, 3D imaging data can be visualized and manipulated in a semi-immersive virtual environment. The paper also details the interaction techniques we specifically designed for a medical imaging scenario and provides implementation details of the integration of the interface in an open-source medical image viewer.},
keywords = {3D interaction, Data glove, Healthcare, Visualization},
pubstate = {published},
tppubtype = {inproceedings}
}
2009
Gallo, Luigi; Ciampi, Mario
Wii Remote-enhanced Hand-Computer Interaction for 3D Medical Image Analysis Proceedings Article
In: CTIT '09: Proceedings of International Conference on the Current Trends in Information Technology, pp. 85–90, IEEE Computer Society, Dubai, United Arab Emirates, 2009, ISBN: 978-1-4244-5755-7.
Abstract | Links | BibTeX | Tags: 3D interaction, Data glove, Healthcare, Visualization, Wiimote
@inproceedings{galloWiiRemoteenhancedHandComputer2009,
title = {Wii Remote-enhanced Hand-Computer Interaction for 3D Medical Image Analysis},
author = { Luigi Gallo and Mario Ciampi},
doi = {10.1109/CTIT.2009.5423137},
isbn = {978-1-4244-5755-7},
year = {2009},
date = {2009-12-01},
booktitle = {CTIT '09: Proceedings of International Conference on the Current Trends in Information Technology},
pages = {85--90},
publisher = {IEEE Computer Society},
address = {Dubai, United Arab Emirates},
abstract = {In this paper, we introduce a low cost, not cumbersome, 3D user interface for exploring medical data. 3D imaging data can be manipulated in a semi-immersive virtual environment by means of an off-the-shelf wireless data glove equipped with an additional infrared LED. The LED is tracked by a Wii Remote, which is able to provide accurate positional information with no need for further six degrees of freedom position trackers. We also describe the interaction techniques we have designed specifically for a medical imaging scenario and provide some implementation details of the integration of the interface in an open-source image processing software for medical research.},
keywords = {3D interaction, Data glove, Healthcare, Visualization, Wiimote},
pubstate = {published},
tppubtype = {inproceedings}
}
Gallo, Luigi; Ciampi, Mario
Wii Remote-enhanced Hand-Computer Interaction for 3D Medical Image Analysis Proceedings Article
In: CTIT '09: Proceedings of International conference on the Current Trends in Information Technology, pp. 85–90, IEEE Computer Society, Dubai, United Arab Emirates, 2009, ISBN: 978-1-4244-5755-7.
Abstract | Links | BibTeX | Tags: 3D interaction, Data glove, Healthcare, Visualization, Wiimote
@inproceedings{gallo_wii_2009,
title = {Wii Remote-enhanced Hand-Computer Interaction for 3D Medical Image Analysis},
author = {Luigi Gallo and Mario Ciampi},
doi = {10.1109/CTIT.2009.5423137},
isbn = {978-1-4244-5755-7},
year = {2009},
date = {2009-12-01},
booktitle = {CTIT '09: Proceedings of International conference on the Current Trends in Information Technology},
pages = {85–90},
publisher = {IEEE Computer Society},
address = {Dubai, United Arab Emirates},
abstract = {In this paper, we introduce a low cost, not cumbersome, 3D user interface for exploring medical data. 3D imaging data can be manipulated in a semi-immersive virtual environment by means of an off-the-shelf wireless data glove equipped with an additional infrared LED. The LED is tracked by a Wii Remote, which is able to provide accurate positional information with no need for further six degrees of freedom position trackers. We also describe the interaction techniques we have designed specifically for a medical imaging scenario and provide some implementation details of the integration of the interface in an open-source image processing software for medical research.},
keywords = {3D interaction, Data glove, Healthcare, Visualization, Wiimote},
pubstate = {published},
tppubtype = {inproceedings}
}