Learn to efficiently exploit cost maps by combining RRT∗ with Reinforcement Learning

Riccardo Franceschini, Matteo Fumagalli, Julian Cayero Becerra

Research output: Chapter in Book/Report/Conference proceedingArticle in proceedingsResearchpeer-review

Abstract

Safe autonomous navigation of robots in complex and cluttered environments is a crucial task and is still an open challenge even in 2D environments. Being able to efficiently minimize multiple constraints such as safety or battery drain requires the ability to understand and leverage information from different cost maps. Rapid-exploring random trees (RRT) methods are often used in current path planning methods, thanks to their efficiency in finding a quick path to the goal. However, these approaches suffer from a slow convergence towards an optimal solution, especially when the planner's goal must consider other aspects like safety or battery consumption besides simply achieving the goal. Therefore, it is proposed a sample-efficient and cost-aware sampling RRT∗ method that can overcome previous methods by exploiting the information gathered from map analysis. In particular, the use of a Reinforcement Learning agent is leveraged to guide the RRT∗ sampling toward an almost optimal solution. The performance of the proposed method is demonstrated against different RRT∗ implementations in multiple synthetic environments.

Original languageEnglish
Title of host publicationProceedings of the IEEE International Symposium on Safety, Security, and Rescue Robotics (SSRR)
PublisherIEEE
Publication date2022
Pages251-256
ISBN (Print)978-1-6654-5681-4
ISBN (Electronic)978-1-6654-5680-7
DOIs
Publication statusPublished - 2022
Event2022 IEEE International Symposium on Safety, Security, and Rescue Robotics - Sevilla, Spain
Duration: 8 Nov 202210 Nov 2022

Conference

Conference2022 IEEE International Symposium on Safety, Security, and Rescue Robotics
Country/TerritorySpain
CitySevilla
Period08/11/202210/11/2022

Fingerprint

Dive into the research topics of 'Learn to efficiently exploit cost maps by combining RRT∗ with Reinforcement Learning'. Together they form a unique fingerprint.

Cite this