Vision for Robust Robot Manipulation
Ver/ Abrir
Metadatos
Mostrar el registro completo del ítemcomunitat-uji-handle:10234/9
comunitat-uji-handle2:10234/7036
comunitat-uji-handle3:10234/8620
comunitat-uji-handle4:
INVESTIGACIONMetadatos
Título
Vision for Robust Robot ManipulationFecha de publicación
2019Editor
MDPIISSN
1424-8220; 1424-8220Cita bibliográfica
Martinez-Martin, Ester; del Pobil, Angel P. "Vision for Robust Robot Manipulation." Sensors, 2019, vol. 19, núm. 7, p. 1648Tipo de documento
info:eu-repo/semantics/articleVersión de la editorial
https://www.mdpi.com/1424-8220/19/7/1648Versión
info:eu-repo/semantics/publishedVersionPalabras clave / Materias
Resumen
Advances in Robotics are leading to a new generation of assistant robots working in
ordinary, domestic settings. This evolution raises new challenges in the tasks to be accomplished by
the robots. This is the case ... [+]
Advances in Robotics are leading to a new generation of assistant robots working in
ordinary, domestic settings. This evolution raises new challenges in the tasks to be accomplished by
the robots. This is the case for object manipulation where the detect-approach-grasp loop requires a
robust recovery stage, especially when the held object slides. Several proprioceptive sensors have
been developed in the last decades, such as tactile sensors or contact switches, that can be used for
that purpose; nevertheless, their implementation may considerably restrict the gripper’s flexibility
and functionality, increasing their cost and complexity. Alternatively, vision can be used since it
is an undoubtedly rich source of information, and in particular, depth vision sensors. We present
an approach based on depth cameras to robustly evaluate the manipulation success, continuously
reporting about any object loss and, consequently, allowing it to robustly recover from this situation.
For that, a Lab-colour segmentation allows the robot to identify potential robot manipulators in the
image. Then, the depth information is used to detect any edge resulting from two-object contact.
The combination of those techniques allows the robot to accurately detect the presence or absence
of contact points between the robot manipulator and a held object. An experimental evaluation in
realistic indoor environments supports our approach. [-]
Publicado en
Sensors, 2019, vol. 19, núm. 7, p. 1648Proyecto de investigación
This research was partially funded by Ministerio de Economía y Competitividad grant number DPI2015-69041-R. This paper describes research done at UJI Robotic Intelligence Laboratory. Support for this laboratory is provided in part by Ministerio de Economía y Competitividad and by Universitat Jaume I (UJI-B2018-74).Derechos de acceso
info:eu-repo/semantics/openAccess
Aparece en las colecciones
- ICC_Articles [424]
El ítem tiene asociados los siguientes ficheros de licencia:
Excepto si se señala otra cosa, la licencia del ítem se describe como: © 2019 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access
article distributed under the terms and conditions of the Creative Commons Attribution
(CC BY) license (http://creativecommons.org/licenses/by/4.0/).