TY - GEN
T1 - Mobile robot navigation based on deep reinforcement learning with 2D-LiDAR sensor using stochastic approach
AU - Beomsoo, Han
AU - Ravankar, Ankit A.
AU - Emaru, Takanori
N1 - Publisher Copyright:
© 2021 IEEE.
PY - 2021/3/4
Y1 - 2021/3/4
N2 - In recent years, there has been a significant progress in mobile robotics and their applications in different fields. Currently, mobile robots are employed for applications such as service robots for delivery, exploration, mapping, search and rescue, and warehouses. Recent advances in computing efficiency and machine learning algorithms have increased the variations of intelligent robots that can navigate autonomously using sensor data. Particularly, reinforcement learning has recently enjoyed a wide variety of success in controlling the robot motion in an unknown environment. However, most of the reinforcement learning-based navigation gets the path plan with a deterministic method, which results in some errors. Therefore, we present a navigation policy for a mobile robot equipped with a 2D range sensor based on the Proximal Policy Optimization of a stochastic approach. The tested algorithm also includes a stochastic operation, which simplifies the policy network model. We trained a differential drive robot in multiple training environments, and based on such stochastic learning, the training data accumulates faster than before. We tested our algorithm in a virtual environment and present the results of successful planning and navigation for mobile robots.
AB - In recent years, there has been a significant progress in mobile robotics and their applications in different fields. Currently, mobile robots are employed for applications such as service robots for delivery, exploration, mapping, search and rescue, and warehouses. Recent advances in computing efficiency and machine learning algorithms have increased the variations of intelligent robots that can navigate autonomously using sensor data. Particularly, reinforcement learning has recently enjoyed a wide variety of success in controlling the robot motion in an unknown environment. However, most of the reinforcement learning-based navigation gets the path plan with a deterministic method, which results in some errors. Therefore, we present a navigation policy for a mobile robot equipped with a 2D range sensor based on the Proximal Policy Optimization of a stochastic approach. The tested algorithm also includes a stochastic operation, which simplifies the policy network model. We trained a differential drive robot in multiple training environments, and based on such stochastic learning, the training data accumulates faster than before. We tested our algorithm in a virtual environment and present the results of successful planning and navigation for mobile robots.
UR - http://www.scopus.com/inward/record.url?scp=85106508778&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85106508778&partnerID=8YFLogxK
U2 - 10.1109/ISR50024.2021.9419565
DO - 10.1109/ISR50024.2021.9419565
M3 - Conference contribution
AN - SCOPUS:85106508778
T3 - ISR 2021 - 2021 IEEE International Conference on Intelligence and Safety for Robotics
SP - 417
EP - 422
BT - ISR 2021 - 2021 IEEE International Conference on Intelligence and Safety for Robotics
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 2nd IEEE International Conference on Intelligence and Safety for Robotics, ISR 2021
Y2 - 4 March 2021 through 6 March 2021
ER -