TY - JOUR
T1 - Synergy Emergence in Deep Reinforcement Learning for Full-Dimensional Arm Manipulation
AU - Han, Jihui
AU - Chai, Jiazheng
AU - Hayashibe, Mitsuhiro
N1 - Funding Information:
This work was supported by the JSPS Grant-in-Aid for Scientific Research on Innovative Areas Hyper-Adaptability Project under Grant 20H05458.
Publisher Copyright:
© 2018 IEEE.
PY - 2021/5
Y1 - 2021/5
N2 - Full-dimensional natural arm manipulation is a challenging task in the field of model-based control due to its high degree of freedom and unknown dynamics of the given system. Deep reinforcement learning (DRL) offers a promising model-free approach for handling high-dimensional robotics problems. Although impressive results for the arm manipulation task have been reported, it still remains an open problem on how we can create human-like synergetic reaching motion using learning algorithms. In this study, we apply DRL for managing full-dimensional arm manipulation in a simulation study, and verify the relations among motion error, energy, and synergy emergence, to reveal the mechanism of employing motor synergy. Although synergy information has never been encoded into the reward function, the synergy naturally emerges along with feedforward control, leading to a similar situation as human motion learning. To the best of our knowledge, this is a pioneer study demonstrating the error and energy optimization issue exists behind the motor synergy employment in DRL for reaching tasks. In addition, our proposed feedback-augmented DRL controller shows better capability over DRL in terms of synergy development and the coupled criteria of error-energy index. This implies that feedback control can support the learning process under redundancy by voiding unnecessary random exploration.
AB - Full-dimensional natural arm manipulation is a challenging task in the field of model-based control due to its high degree of freedom and unknown dynamics of the given system. Deep reinforcement learning (DRL) offers a promising model-free approach for handling high-dimensional robotics problems. Although impressive results for the arm manipulation task have been reported, it still remains an open problem on how we can create human-like synergetic reaching motion using learning algorithms. In this study, we apply DRL for managing full-dimensional arm manipulation in a simulation study, and verify the relations among motion error, energy, and synergy emergence, to reveal the mechanism of employing motor synergy. Although synergy information has never been encoded into the reward function, the synergy naturally emerges along with feedforward control, leading to a similar situation as human motion learning. To the best of our knowledge, this is a pioneer study demonstrating the error and energy optimization issue exists behind the motor synergy employment in DRL for reaching tasks. In addition, our proposed feedback-augmented DRL controller shows better capability over DRL in terms of synergy development and the coupled criteria of error-energy index. This implies that feedback control can support the learning process under redundancy by voiding unnecessary random exploration.
KW - Deep learning in robotics
KW - human motor learning
KW - motor synergy
KW - reaching task
KW - redundancy
UR - http://www.scopus.com/inward/record.url?scp=85115872799&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85115872799&partnerID=8YFLogxK
U2 - 10.1109/TMRB.2021.3056924
DO - 10.1109/TMRB.2021.3056924
M3 - Article
AN - SCOPUS:85115872799
SN - 2576-3202
VL - 3
SP - 498
EP - 509
JO - IEEE Transactions on Medical Robotics and Bionics
JF - IEEE Transactions on Medical Robotics and Bionics
IS - 2
M1 - 9345796
ER -