To operate autonomously in a various unfamiliar settings and successfully complete missions, mobile robots should be able to adapt to changes in their surroundings. Visual teach and repeat (VT&R) systems are a promising class of approaches for training robots to adaptively navigate environments.
As their name suggests, VT&R systems are based on two key stages: the teach and the repeat steps. During the teach step, the systems learn from demonstrations of paths taken by human operators. Subsequently, during the repeat stage, the robots try to replicate what the humans did in the demonstration, walking down the same path autonomously and as consistently as possible.
Researchers at the Oxford Robotics Institute have recently developed a new controller that could help to enhance VT&R systems. Their approach, presented in a paper published in IEEE Robotics and Automation Letters, could help to develop robots that are better at navigating unfamiliar environments.
“The recent paper is part of our work on VT&R navigation,” Matias Mattamala, one of the authors, told TechXplore. “This is useful to quickly deploy robots to inspect new places and collect data without having to build a precise map of the environment. In our previous work, we demonstrated robustness to visual occlusions by switching between different cameras on the robot, such as when someone is walking by.”
In their previous studies, Mattamala and his colleagues were able to train models to access different cameras on a robot at different times, using data they collected during human demonstrations. Despite this remarkable achievement, their models did not allow robots to actively avoid potentially obstacles in their surroundings while replicating the trajectory demonstrated by human agents.
“We started to work on this ‘safety layer’ some time ago and our recent paper presents it fully functional,” Mattamala explained. “Our controller is based on a recent approach developed by Nvidia called Riemannian Motion Policies (RMP).”
The controller developed by the researchers partly resembles potential field controllers, tools that allow robots to compute a combination of different forces, such as a attraction forces (i.e., those driving them toward completing a goal) and repulsion forces (i.e., those helping them to stay away from obstacles), to ultimately determine what direction to move in. Nvidia’s RMP approach, however, takes their controller one step further, as it introduces dynamic weights (called metrics) that leverage these forces in different ways, depending on the state of the robot.
“For example, you don’t need to always avoid obstacles, but only when you are close to them or pointing in their direction,” Mattamala explained. “In this way, you can prevent some situations in which attraction and reaction forces cancel each other.”
The interacting forces processed by the team’s controller are computed from a local map that is generated on the fly and adapts as a robot moves in its surrounding environment. By analyzing this local map, the system can generate fields that are easy to interpret and can be used as data to enhance a robot’s navigation skills. This includes a signed distance field (SDF), which characterizes obstacles, and a geodisc distance field (GDF), which conveys the closest distance to a goal or target location. When processing these fields, the controller accounts for the fact that there is a certain amount of space in the surrounding environment that the robot cannot move in or traverse.
“In our study, we were able to explore novel control techniques such as RMP, which have so far only been applied to robot manipulators or small wheeled robots,” Mattamala said. “In addition, we deployed our controller on the ANYbotics’ ANYmal quadruped and performed closed-loop experiments in a decommissioned mine, which was quite exciting to test.”
In contrast with other previously proposed approaches, the controller created by Mattamala and his colleagues is intrinsically reactive, as it does not require robots and developers to plan ahead and predict the obstacles a robot will encounter in a specific environment. Interestingly, in their evaluations, the team found that by using better environment representations to generate attraction and reaction forces, they could achieve similar results to those attained by models that plan for missions in advance.
“For example, we placed some obstacles blocking the reference path and the robot was able to go around without planning,” Mattamala explained. “We also extended our VT&R system to work with fisheye cameras, such as the Sevensense Alphasense rig that we used in our experiments. We achieved comparable results to previous experiments with Realsense cameras, which demonstrated the flexibility of our system.”
So far, the researchers have tested their controller in a series of indoor cluttered spaces and in an underground mine. In these initial experiments, their system achieved very promising results, suggesting that it could soon help to enhance the navigation capabilities of both existing and newly developed mobile robots. Notably, the controller could be applied to a variety of systems, as it only requires a local map generated using data collected by depth cameras or LiDAR technology.
In their next studies, Mattamala and his colleagues plan to apply and test their controller on other robots developed in their lab. In addition, they would like to evaluate its performance in a broader range of dynamic, real-world environments.
“Our future work considers extending our VT&R system to achieve the long-term visual navigation of legged robots in industrial and natural environments,” Mattamala explained. “This requires (1) better visual localization systems, since drastic appearance changes due to lighting or weather conditions will challenge our current system, and (2) better walking controllers to achieve reliable navigation in rough terrain, which should interact with the high-level navigation. Imagine teaching the robot to traverse forest trails or to hike along a mountain trail, and then repeating the trajectory autonomously, no matter the terrain or the weather—that’s what we aim to achieve.”
Matias Mattamala et al, An Efficient Locally Reactive Controller for Safe Navigation in Visual Teach and Repeat Missions, IEEE Robotics and Automation Letters (2022). DOI: 10.1109/LRA.2022.3143196
© 2022 Science X Network
A locally reactive controller to enhance visual teach and repeat systems (2022, February 8)
retrieved 8 February 2022
This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no
part may be reproduced without the written permission. The content is provided for information purposes only.