Mostrar el registro sencillo del ítem
Large Language Models for in Situ Knowledge Documentation and Access With Augmented Reality
dc.contributor.author | Izquierdo-Domenech, Juan | |
dc.contributor.author | Linares-Pellicer, Jordi | |
dc.contributor.author | Ferri-Molla, Isabel | |
dc.date | 2023-09 | |
dc.date.accessioned | 2023-10-02T15:57:32Z | |
dc.date.available | 2023-10-02T15:57:32Z | |
dc.identifier.citation | J. Izquierdo-Domenech, J. Linares-Pellicer, I. Ferri-Molla. Large Language Models for in Situ Knowledge Documentation and Access With Augmented Reality, International Journal of Interactive Multimedia and Artificial Intelligence, (2023), http://dx.doi.org/10.9781/ijimai.2023.09.002 | es_ES |
dc.identifier.issn | 1989-1660 | |
dc.identifier.uri | https://reunir.unir.net/handle/123456789/15340 | |
dc.description.abstract | Augmented reality (AR) has become a powerful tool for assisting operators in complex environments, such as shop floors, laboratories, and industrial settings. By displaying synthetic visual elements anchored in real environments and providing information for specific tasks, AR helps to improve efficiency and accuracy. However, a common bottleneck in these environments is introducing all necessary information, which often requires predefined structured formats and needs more ability for multimodal and Natural Language (NL) interaction. This work proposes a new method for dynamically documenting complex environments using AR in a multimodal, non-structured, and interactive manner. Our method employs Large Language Models (LLMs) to allow experts to describe elements from the real environment in NL and select corresponding AR elements in a dynamic and iterative process. This enables a more natural and flexible way of introducing information, allowing experts to describe the environment in their own words rather than being constrained by a predetermined structure. Any operator can then ask about any aspect of the environment in NL to receive a response and visual guidance from the AR system, thus allowing for a more natural and flexible way of introducing and retrieving information. These capabilities ultimately improve the effectiveness and efficiency of tasks in complex environments. | es_ES |
dc.language.iso | eng | es_ES |
dc.publisher | International Journal of Interactive Multimedia and Artificial Intelligence | es_ES |
dc.relation.ispartofseries | ;In Press | |
dc.relation.uri | https://www.ijimai.org/journal/bibcite/reference/3378 | es_ES |
dc.rights | openAccess | es_ES |
dc.subject | augmented reality | es_ES |
dc.subject | deep learning | es_ES |
dc.subject | multimodal | es_ES |
dc.subject | large language models | es_ES |
dc.subject | transformer | es_ES |
dc.subject | IJIMAI | es_ES |
dc.title | Large Language Models for in Situ Knowledge Documentation and Access With Augmented Reality | es_ES |
dc.type | article | es_ES |
reunir.tag | ~IJIMAI | es_ES |
dc.identifier.doi | https://doi.org/10.9781/ijimai.2023.09.002 |