Self-supervised learning of depth-based navigation affordances from haptic cues

This paper presents a ground vehicle capable of exploiting haptic cues to learn navigation affordances from depth cues. A simple pan-tilt telescopic antenna and a Kinect sensor, both fitted to the robot’s body frame, provide the required haptic and depth sensory feedback, respectively. With the ante...

Full description

Bibliographic Details
Main Author: Baleia, J. (author)
Other Authors: Santana, P. (author), Barata, J. (author)
Format: conferenceObject
Language:eng
Published: 2022
Subjects:
Online Access:http://hdl.handle.net/10071/25845
Country:Portugal
Oai:oai:repositorio.iscte-iul.pt:10071/25845
Description
Summary:This paper presents a ground vehicle capable of exploiting haptic cues to learn navigation affordances from depth cues. A simple pan-tilt telescopic antenna and a Kinect sensor, both fitted to the robot’s body frame, provide the required haptic and depth sensory feedback, respectively. With the antenna, the robot determines whether an object is traversable by the robot. Then, the interaction outcome is associated to the object’s depth-based descriptor. Later on, the robot to predict if a newly observed object is traversable just by inspecting its depth-based appearance uses this acquired knowledge. A set of field trials show the ability of the to robot progressively learn which elements of the environment are traversable.