Abstract
This paper presents a Deep Reinforcement Learning (DRL) method based on a mechanical (work) Energy reward function applied to a reconfigurable RSTAR robot to overcome obstacles. The RSTAR is a crawling robot that can reconfigure its shape and shift the location of its center of mass via a sprawl and a four-bar extension mechanism. The DRL was applied in a simulated environment with a physical engine (UNITY TM). The robot was trained on a step obstacle and a two-stage narrow passage obstacle composed of a horizontal and a vertical channel. To evaluate the benefits of the proposed Energy reward function, it was compared to time-based and movement-based reward functions. The results showed that the Energy-based reward produced superior results in terms of obstacle height, energy requirements, and time to overcome the obstacle. The Energy-based reward method also converged faster to the solution compared to the other reward methods. The DRL's results for all the methods (energy, time and movement- based rewards) were superior to the best results produced by the human experts (see attached video).
Original language | English |
---|---|
Article number | 10121761 |
Pages (from-to) | 47681-47689 |
Number of pages | 9 |
Journal | IEEE Access |
Volume | 11 |
Issue number | 99 |
DOIs | |
State | Published - 9 May 2023 |
Keywords
- Obstacle negotiation
- reconfigurable robot
- reinforcement learning (RL)
- reward shaping
ASJC Scopus subject areas
- Computer Science (all)
- Materials Science (all)
- Engineering (all)