elib
DLR-Header
DLR-Logo -> http://www.dlr.de
DLR Portal Home | Imprint | Privacy Policy | Contact | Deutsch
Fontsize: [-] Text [+]

Fusion of multimodal imaging techniques towards autonomous navigation

Sharif, Helia (2021) Fusion of multimodal imaging techniques towards autonomous navigation. Dissertation, Universität Bremen. doi: 10.26092/elib/1077.

[img] PDF - Only accessible within DLR
115MB

Abstract

"Earth is the cradle of humanity, but one cannot live in a cradle forever." -Konstantin E. Tsiolkovsky, an early pioneer of rocketry and astronautics. Space robotics enable humans to explore beyond our home planet. Traditional techniques for tele-operated robotic guidance make it possible for a driver to direct a rover that is up to 245.55Mkm away. However, relying on manual terrestrial operators for guidance is a key limitation for exploration missions today, as real-time communication between rovers and operators is delayed by long distances and limited uplink opportunities. Moreover, autonomous guidance techniques in use today are generally limited in scope and capacity; for example, some autonomous techniques presently in use require the application of special markers on targets in order to enable detection, while other techniques provide autonomous vision-based flight navigation but only at limited altitudes in ideal visibility conditions. Improving autonomy is thus essential to expanding the scope of viable space missions. In this thesis, a fusion of monocular visible and infrared imaging cameras is employed to estimate the relative pose of a nearby target while compensating for each spectrum's shortcomings. The robustness of the algorithm was tested in a number of different scenarios by simulating harsh space environments while imaging a subject of similar characteristics to a spacecraft in orbit. It is shown that the fusion of visual odometries from two spectrums performs well where knowledge of the target's physical characteristics is limited. The result of this thesis research is an autonomous, robust vision-based tracking system designed for space applications. This appealing solution can be used onboard most spacecraft and adapted for the specific application of any given mission.

Item URL in elib:https://elib.dlr.de/145543/
Document Type:Thesis (Dissertation)
Title:Fusion of multimodal imaging techniques towards autonomous navigation
Authors:
AuthorsInstitution or Email of AuthorsAuthor's ORCID iDORCID Put Code
Sharif, HeliaUNSPECIFIEDhttps://orcid.org/0000-0002-1256-1329UNSPECIFIED
Date:17 September 2021
Refereed publication:Yes
Open Access:No
DOI:10.26092/elib/1077
Status:Published
Keywords:fusion of multimodal sensors; visual odometry; monocular imaging; Extended Kalman Filter; autonomous vision-based navigation for space applications;
Institution:Universität Bremen
Department:Fachbereich 03: Mathematik/Informatik (FB 03)
HGF - Research field:Aeronautics, Space and Transport
HGF - Program:Space
HGF - Program Themes:Space System Technology
DLR - Research area:Raumfahrt
DLR - Program:R SY - Space System Technology
DLR - Research theme (Project):R - Optical navigation on hybrid avionics architecture
Location: Bremen
Institutes and Institutions:Institute of Space Systems > Navigation and Control Systems
Deposited By: Theil, Dr.-Ing. Stephan
Deposited On:12 Nov 2021 09:06
Last Modified:12 Nov 2021 09:06

Repository Staff Only: item control page

Browse
Search
Help & Contact
Information
OpenAIRE Validator logo electronic library is running on EPrints 3.3.12
Website and database design: Copyright © German Aerospace Center (DLR). All rights reserved.