loading
Papers Papers/2022 Papers Papers/2022

Research.Publish.Connect.

Paper

Authors: Chenxing Li 1 ; 2 ; Yinlong Liu 3 ; Zhenshan Bing 3 ; Fabian Schreier 1 ; 2 ; Jan Seyler 2 and Shahram Eivazi 1 ; 2

Affiliations: 1 University of Tübingen, Tübingen, Germany ; 2 Festo, Esslingen, Germany ; 3 Technical University of Munich, Munich, Germany

Keyword(s): Q-function Targets Via Optimization, Data Efficiency, Hindsight Goals Techniques, Offline Data Collection, Dynamic Buffer.

Abstract: In this paper, we examine three extensions to the Q-function Targets via Optimization (QT-Opt) algorithm and empirically studies their effects on training time over complex robotic tasks. The vanilla QT-Opt algorithm requires lots of offline data (several months with multiple robots) for training which is hard to collect in practice. To bridge the gap between basic reinforcement learning research and real world robotic applications, first we propose to use hindsight goals techniques (Hindsight Experience Replay, Hindsight Goal Generation) and Energy-Based Prioritization (EBP) to increase data efficiency in reinforcement learning. Then, an efficient offline data collection method using PD control method and dynamic buffer are proposed. Our experiments show that both data collection and training the agent for a robotic grasping task takes about one day only, besides, the learning performance maintains high level (80% successful rate). This work serves as a step towards accelerating the training of reinforcement learning for complex real world robotics tasks. (More)

CC BY-NC-ND 4.0

Sign In Guest: Register as new SciTePress user now for free.

Sign In SciTePress user: please login.

PDF ImageMy Papers

You are not signed in, therefore limits apply to your IP address 3.12.161.77

In the current month:
Recent papers: 100 available of 100 total
2+ years older papers: 200 available of 200 total

Paper citation in several formats:
Li, C.; Liu, Y.; Bing, Z.; Schreier, F.; Seyler, J. and Eivazi, S. (2023). Accelerate Training of Reinforcement Learning Agent by Utilization of Current and Previous Experience. In Proceedings of the 15th International Conference on Agents and Artificial Intelligence - Volume 3: ICAART; ISBN 978-989-758-623-1; ISSN 2184-433X, SciTePress, pages 698-705. DOI: 10.5220/0011745600003393

@conference{icaart23,
author={Chenxing Li. and Yinlong Liu. and Zhenshan Bing. and Fabian Schreier. and Jan Seyler. and Shahram Eivazi.},
title={Accelerate Training of Reinforcement Learning Agent by Utilization of Current and Previous Experience},
booktitle={Proceedings of the 15th International Conference on Agents and Artificial Intelligence - Volume 3: ICAART},
year={2023},
pages={698-705},
publisher={SciTePress},
organization={INSTICC},
doi={10.5220/0011745600003393},
isbn={978-989-758-623-1},
issn={2184-433X},
}

TY - CONF

JO - Proceedings of the 15th International Conference on Agents and Artificial Intelligence - Volume 3: ICAART
TI - Accelerate Training of Reinforcement Learning Agent by Utilization of Current and Previous Experience
SN - 978-989-758-623-1
IS - 2184-433X
AU - Li, C.
AU - Liu, Y.
AU - Bing, Z.
AU - Schreier, F.
AU - Seyler, J.
AU - Eivazi, S.
PY - 2023
SP - 698
EP - 705
DO - 10.5220/0011745600003393
PB - SciTePress