Kulkarni, Padmaja und Kober, Jens und Babuška, Robert und Della Santina, Cosimo (2021) Learning Assembly Tasks in a Few Minutes by Combining Impedance Control and Residual Recurrent Reinforcement Learning. Advanced Intelligent Systems, 4 (1). Wiley. doi: 10.1002/aisy.202100095. ISSN 2640-4567.
PDF
- Verlagsversion (veröffentlichte Fassung)
2MB |
Offizielle URL: https://dx.doi.org/10.1002/aisy.202100095
Kurzfassung
Adapting to uncertainties is essential yet challenging for robots while conducting assembly tasks in real-world scenarios. Reinforcement learning (RL) methods provide a promising solution for these cases. However, training robots with RL can be a data-extensive, time-consuming, and potentially unsafe process. In contrast, classical control strategies can have near-optimal performance without training and be certifiably safe. However, this is achieved at the cost of assuming that the environment is known up to small uncertainties. Herein, an architecture aiming at getting the best out of the two worlds, by combining RL and classical strategies so that each one deals with the right portion of the assembly problem, is proposed. A time-varying weighted sum combines a recurrent RL method with a nominal strategy. The output serves as the reference for a task space impedance controller. The proposed approach can learn to insert an object in a frame within a few minutes of real-world training. A success rate of 94% in the presence of considerable uncertainties is observed. Furthermore, the approach is robust to changes in the experimental setup and task, even when no retrain is performed. For example, the same policy achieves a success rate of 85% when the object properties change.
elib-URL des Eintrags: | https://elib.dlr.de/193634/ | ||||||||||||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
Dokumentart: | Zeitschriftenbeitrag | ||||||||||||||||||||
Titel: | Learning Assembly Tasks in a Few Minutes by Combining Impedance Control and Residual Recurrent Reinforcement Learning | ||||||||||||||||||||
Autoren: |
| ||||||||||||||||||||
Datum: | 2 September 2021 | ||||||||||||||||||||
Erschienen in: | Advanced Intelligent Systems | ||||||||||||||||||||
Referierte Publikation: | Ja | ||||||||||||||||||||
Open Access: | Ja | ||||||||||||||||||||
Gold Open Access: | Ja | ||||||||||||||||||||
In SCOPUS: | Nein | ||||||||||||||||||||
In ISI Web of Science: | Ja | ||||||||||||||||||||
Band: | 4 | ||||||||||||||||||||
DOI: | 10.1002/aisy.202100095 | ||||||||||||||||||||
Verlag: | Wiley | ||||||||||||||||||||
ISSN: | 2640-4567 | ||||||||||||||||||||
Status: | veröffentlicht | ||||||||||||||||||||
Stichwörter: | impedance control; reinforcement learning | ||||||||||||||||||||
HGF - Forschungsbereich: | Luftfahrt, Raumfahrt und Verkehr | ||||||||||||||||||||
HGF - Programm: | Raumfahrt | ||||||||||||||||||||
HGF - Programmthema: | Robotik | ||||||||||||||||||||
DLR - Schwerpunkt: | Raumfahrt | ||||||||||||||||||||
DLR - Forschungsgebiet: | R RO - Robotik | ||||||||||||||||||||
DLR - Teilgebiet (Projekt, Vorhaben): | R - Roboterdynamik & Simulation [RO] | ||||||||||||||||||||
Standort: | Oberpfaffenhofen | ||||||||||||||||||||
Institute & Einrichtungen: | Institut für Robotik und Mechatronik (ab 2013) > Analyse und Regelung komplexer Robotersysteme Institut für Robotik und Mechatronik (ab 2013) | ||||||||||||||||||||
Hinterlegt von: | Strobl, Dr. Klaus H. | ||||||||||||||||||||
Hinterlegt am: | 28 Jan 2023 12:04 | ||||||||||||||||||||
Letzte Änderung: | 30 Jan 2023 07:56 |
Nur für Mitarbeiter des Archivs: Kontrollseite des Eintrags