Multimodal Multi-User Mixed Reality Human–Robot Interface for Remote Operations in Hazardous Environments
Ver/ Abrir
Impacto
Scholar |
Otros documentos de la autoría: Szczurek, Krzysztof Adam; Marin, Raul; Matheson, Eloise; Rodriguez-Nogueira, Jose; Di Castro, Mario
Metadatos
Mostrar el registro completo del ítemcomunitat-uji-handle:10234/9
comunitat-uji-handle2:10234/7036
comunitat-uji-handle3:10234/8620
comunitat-uji-handle4:
INVESTIGACIONMetadatos
Título
Multimodal Multi-User Mixed Reality Human–Robot Interface for Remote Operations in Hazardous EnvironmentsAutoría
Fecha de publicación
2023-02-15Editor
Institute of Electrical and Electronics Engineers Inc.ISSN
2169-3536Cita bibliográfica
K. A. Szczurek, R. M. Prades, E. Matheson, J. Rodriguez-Nogueira and M. D. Castro, "Multimodal Multi-User Mixed Reality Human–Robot Interface for Remote Operations in Hazardous Environments," in IEEE Access, vol. 11, pp. 17305-17333, 2023, doi: 10.1109/ACCESS.2023.3245833.Tipo de documento
info:eu-repo/semantics/articleVersión de la editorial
https://ieeexplore.ieee.org/abstract/document/10045681Versión
info:eu-repo/semantics/publishedVersionPalabras clave / Materias
Resumen
In hazardous environments, where conditions present risks for humans, the maintenance and interventions are often done with teleoperated remote systems or mobile robotic manipulators to avoid human exposure to dangers. ... [+]
In hazardous environments, where conditions present risks for humans, the maintenance and interventions are often done with teleoperated remote systems or mobile robotic manipulators to avoid human exposure to dangers. The increasing need for safe and efficient teleoperation requires advanced environmental awareness and collision avoidance. The up-to-date screen-based 2D or 3D interfaces do not fully allow the operator to immerse in the controlled scenario. This problem can be addressed with the emerging Mixed Reality (MR) technologies with Head-Mounted Devices (HMDs) that offer stereoscopic immersion and interaction with virtual objects. Such human-robot interfaces have not yet been demonstrated in telerobotic interventions in particle physics accelerators. Moreover, the operations often require a few experts to collaborate, which increases the system complexity and requires sharing an Augmented Reality (AR) workspace. The multi-user mobile telerobotics in hazardous environments with shared control in the AR has not yet been approached in the state-of-the-art. In this work, the developed MR human-robot interface using the AR HMD is presented. The interface adapts to the constrained wireless networks in particle accelerator facilities and provides reliable high-precision interaction and specialized visualization. The multimodal operation uses hands, eyes and user motion tracking, and voice recognition for control, as well as offers video, 3D point cloud and audio feedback from the robot. Multiple experts can collaborate in the AR workspace locally or remotely, and share or monitor the robot's control. Ten operators tested the interface in intervention scenarios in the European Organization for Nuclear Research (CERN) with complete network characterization and measurements to conclude if operational requirements were met and if the network architecture could support single and multi-user communication load. The interface system has proved to be operationally ready at the Technical Readiness Level (TRL) 8 and was validated through successful demonstration in single and multi-user missions. Some system limitations and further work areas were identified, such as optimizing the network architecture for multi-user scenarios or high-level interface actions applying automatic interaction strategies depending on network conditions. [-]
Publicado en
IEEE Access, vol. 11 (2023)Derechos de acceso
© 2013 IEEE.
info:eu-repo/semantics/openAccess
info:eu-repo/semantics/openAccess
Aparece en las colecciones
- ICC_Articles [423]