AHCI RESEARCH GROUP
Publications
Papers published in international journals,
proceedings of conferences, workshops and books.
OUR RESEARCH
Scientific Publications
How to
You can use the tag cloud to select only the papers dealing with specific research topics.
You can expand the Abstract, Links and BibTex record for each paper.
2025
Guo, H.; Liu, Z.; Tang, C.; Zhang, X.
An Interactive Framework for Personalized Navigation Based on Metacosmic Cultural Tourism and Large Model Fine-Tuning Journal Article
In: IEEE Access, vol. 13, pp. 81450–81461, 2025, ISSN: 21693536 (ISSN).
Abstract | Links | BibTeX | Tags: Cultural informations, Digital Cultural Heritage, Digital cultural heritages, Digital guide, Fine tuning, fine-tuning, Historical monuments, Language Model, Large language model, Leisure, Metacosmic cultural tourism, Multimodal Interaction, Tourism, Virtual tour
@article{guo_interactive_2025,
title = {An Interactive Framework for Personalized Navigation Based on Metacosmic Cultural Tourism and Large Model Fine-Tuning},
author = {H. Guo and Z. Liu and C. Tang and X. Zhang},
url = {https://www.scopus.com/inward/record.uri?eid=2-s2.0-105004059236&doi=10.1109%2fACCESS.2025.3565359&partnerID=40&md5=45d328831c5795fa31e7e033299912b5},
doi = {10.1109/ACCESS.2025.3565359},
issn = {21693536 (ISSN)},
year = {2025},
date = {2025-01-01},
journal = {IEEE Access},
volume = {13},
pages = {81450–81461},
abstract = {With the wide application of large language models (LLMs) and the rapid growth of metaverse tourism demand, the digital tour and personalized interaction of historical sites have become the key to improving users’ digital travel experience. Creating an environment where users can access rich cultural information and enjoy personalized, immersive experiences is a crucial issue in the field of digital cultural travel. To this end, we propose a tourism information multimodal generation personalized question-answering interactive framework TIGMI (Tourism Information Generation and Multimodal Interaction) based on LLM fine-tuning, which aims to provide a richer and more in-depth experience for virtual tours of historical monuments. Taking Qutan Temple as an example, the framework combines LLM, retrieval augmented generation (RAG), and auto-prompting engineering techniques to retrieve accurate information related to the historical monument from external knowledge bases and seamlessly integrates it into the generated content. This integration mechanism ensures the accuracy and relevance of the generated answers. Through TIGMI’s LLM-driven command interaction mechanism in the 3D digital scene of Qutan Temple, users are able to interact with the building and scene environment in a personalized and real-time manner, successfully integrating historical and cultural information with modern digital technology. This integration significantly enhances the naturalness of interaction and personalizes the user experience, thereby improving user immersion and information acquisition efficiency. Evaluation results show that TIGMI excels in question-answering and multimodal interactions, significantly enhancing the depth and breadth of services provided by the personalized virtual tour. We conclude by addressing the limitations of TIGMI and briefly discuss how future research will focus on further improving the accuracy and user satisfaction of the generated content to adapt to the dynamically changing tourism environment. © 2013 IEEE.},
keywords = {Cultural informations, Digital Cultural Heritage, Digital cultural heritages, Digital guide, Fine tuning, fine-tuning, Historical monuments, Language Model, Large language model, Leisure, Metacosmic cultural tourism, Multimodal Interaction, Tourism, Virtual tour},
pubstate = {published},
tppubtype = {article}
}
Hu, H.; Wan, Y.; Tang, K. Y.; Li, Q.; Wang, X.
Affective-Computing-Driven Personalized Display of Cultural Information for Commercial Heritage Architecture Journal Article
In: Applied Sciences (Switzerland), vol. 15, no. 7, 2025, ISSN: 20763417 (ISSN).
Abstract | Links | BibTeX | Tags: Affective Computing, Cultural informations, Cultural value, Data fusion, Information display, Information fusion, Information presentation, Language Model, Large language model, Multimodal information fusion, User-generated, User-generated content, Virtual environments
@article{hu_affective-computing-driven_2025,
title = {Affective-Computing-Driven Personalized Display of Cultural Information for Commercial Heritage Architecture},
author = {H. Hu and Y. Wan and K. Y. Tang and Q. Li and X. Wang},
url = {https://www.scopus.com/inward/record.uri?eid=2-s2.0-105002467183&doi=10.3390%2fapp15073459&partnerID=40&md5=1dc611258248d58a2bf5f44b6a0e890b},
doi = {10.3390/app15073459},
issn = {20763417 (ISSN)},
year = {2025},
date = {2025-01-01},
journal = {Applied Sciences (Switzerland)},
volume = {15},
number = {7},
abstract = {The display methods for traditional cultural heritage lack personalization and emotional interaction, making it difficult to stimulate the public’s deep cultural awareness. This is especially true in commercialized historical districts, where cultural value is easily overlooked. Balancing cultural value and commercial value in information display has become one of the challenges that needs to be addressed. To solve the above problems, this article focuses on the identification of deep cultural values and the optimization of the information display in Beijing’s Qianmen Street, proposing a framework for cultural information mining and display based on affective computing and large language models. The pre-trained models QwenLM and RoBERTa were employed to analyze text and image data from user-generated content on social media, identifying users’ emotional tendencies toward various cultural value dimensions and quantifying their multilayered understanding of architectural heritage. This study further constructed a multimodal information presentation model driven by emotional feedback, mapping it into virtual reality environments to enable personalized, multilayered cultural information visualization. The framework’s effectiveness was validated through an eye-tracking experiment that assessed how different presentation styles impacted users’ emotional engagement and cognitive outcomes. The results show that the affective computing and multimodal data fusion approach to cultural heritage presentation accurately captures users’ emotions, enhancing their interest and emotional involvement. Personalized presentations of information significantly improve users’ engagement, historical understanding, and cultural experience, thereby fostering a deeper comprehension of historical contexts and architectural details. © 2025 by the authors.},
keywords = {Affective Computing, Cultural informations, Cultural value, Data fusion, Information display, Information fusion, Information presentation, Language Model, Large language model, Multimodal information fusion, User-generated, User-generated content, Virtual environments},
pubstate = {published},
tppubtype = {article}
}