Özer, Baran (2026) Guided Reinforcement Learning with Vision Feedback. Masterarbeit, Technical University of Munich (TUM).
|
PDF
8MB |
Kurzfassung
This thesis investigates how to improve reinforcement learning (RL) for contact-rich robotic assembly by combining probabilistic trajectory guidance with visual representation learning. Building on Kernelized Guided Reinforcement Learning (KGRL), where Kernelized Movement Primitives (KMP) encode demonstrations as probabilistic trajectory priors and guide policy learning through uncertainty-aware null space actions, we introduce Vision-KGRL, which augments this framework with compact visual features learned via task-specific Variational Autoencoders (VAEs). The approach is evaluated on peg insertion, gear meshing, and nut threading tasks in the NVIDIA Isaac Lab Factory and Forge environments, showing that visual augmentation preserves the faster convergence of KGRL over standard RL, with stronger gains in higher-dimensional action spaces. Systematic ablations across observation modalities (proprioception, force, vision) and action spaces (4D vs 6D) highlight the complementary roles of trajectory guidance and learned visual features, while results further show that KMP-guided policies significantly reduce interaction forces, leading to smoother and more stable behavior. Overall, Vision-KGRL provides a data-efficient and robust solution for contact-rich manipulation, combining learned visual representations with trajectory priors to improve both learning performance and execution safety.
| elib-URL des Eintrags: | https://elib.dlr.de/224243/ | ||||||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| Dokumentart: | Hochschulschrift (Masterarbeit) | ||||||||||||
| Titel: | Guided Reinforcement Learning with Vision Feedback | ||||||||||||
| Autoren: |
| ||||||||||||
| DLR-Supervisor: |
| ||||||||||||
| Datum: | 2026 | ||||||||||||
| Erschienen in: | Guided Reinforcement Learning with Vision Feedback | ||||||||||||
| Open Access: | Ja | ||||||||||||
| Seitenanzahl: | 81 | ||||||||||||
| Status: | veröffentlicht | ||||||||||||
| Stichwörter: | Reinforcement Learning; Imitation Learning; Robot Learning | ||||||||||||
| Institution: | Technical University of Munich (TUM) | ||||||||||||
| HGF - Forschungsbereich: | Luftfahrt, Raumfahrt und Verkehr | ||||||||||||
| HGF - Programm: | Verkehr | ||||||||||||
| HGF - Programmthema: | Straßenverkehr | ||||||||||||
| DLR - Schwerpunkt: | Verkehr | ||||||||||||
| DLR - Forschungsgebiet: | V ST Straßenverkehr | ||||||||||||
| DLR - Teilgebiet (Projekt, Vorhaben): | V - ASPIRO - Aerospace production using intelligent robotic systems | ||||||||||||
| Standort: | Oberpfaffenhofen | ||||||||||||
| Institute & Einrichtungen: | Institut für Robotik und Mechatronik (ab 2013) > Kognitive Robotik | ||||||||||||
| Hinterlegt von: | Silverio, Joao | ||||||||||||
| Hinterlegt am: | 05 Mai 2026 10:25 | ||||||||||||
| Letzte Änderung: | 05 Mai 2026 10:25 |
Nur für Mitarbeiter des Archivs: Kontrollseite des Eintrags