Logo des Repositoriums
Zur Startseite
  • English
  • Deutsch
Anmelden
  1. Startseite
  2. SuUB
  3. Dissertationen
  4. Fusion of multimodal imaging techniques towards autonomous navigation
 
Zitierlink DOI
10.26092/elib/1077

Fusion of multimodal imaging techniques towards autonomous navigation

Veröffentlichungsdatum
2021-09-17
Autoren
Sharif, Helia  
Betreuer
Suppa, Michael  
Gutachter
Frese, Udo  
Zusammenfassung
“Earth is the cradle of humanity, but one cannot live in a cradle forever.”
-Konstantin E. Tsiolkovsky, an early pioneer of rocketry and astronautics.

Space robotics enable humans to explore beyond our home planet. Traditional techniques for tele-operated robotic guidance make it possible for a driver to direct a rover that is up to 245.55Mkm away. However, relying on manual terrestrial operators for guidance is a key limitation for exploration missions today, as real-time communication between rovers and operators is delayed by long distances and limited uplink opportunities. Moreover, autonomous guidance techniques in use today are generally limited in scope and capacity; for example, some autonomous techniques presently in use require the application of special markers on targets in order to enable detection, while other techniques provide autonomous vision-based flight navigation but only at limited altitudes in ideal visibility conditions. Improving autonomy is thus essential to expanding the scope of viable space missions.

In this thesis, a fusion of monocular visible and infrared imaging cameras is employed to estimate the relative pose of a nearby target while compensating for each spectrum's shortcomings. The robustness of the algorithm was tested in a number of different scenarios by simulating harsh space environments while imaging a subject of similar characteristics to a spacecraft in orbit. It is shown that the fusion of visual odometries from two spectrums performs well where knowledge of the target's physical characteristics is limited.

The result of this thesis research is an autonomous, robust vision-based tracking system designed for space applications. This appealing solution can be used onboard most spacecraft and adapted for the specific application of any given mission.
Schlagwörter
fusion of multimodal sensors

; 

visual odometry

; 

monocular imaging

; 

Extended Kalman Filter

; 

autonomous vision-based navigation for space applications
Institution
Universität Bremen  
Fachbereich
Fachbereich 03: Mathematik/Informatik (FB 03)  
Dokumenttyp
Dissertation
Zweitveröffentlichung
Nein
Sprache
Englisch
Dateien
Lade...
Vorschaubild
Name

Helia_Sharif_Fusion_of_Multimodal_Imaging_Techniques_Towards_Autonomous_Navigation.pdf

Size

110.03 MB

Format

Adobe PDF

Checksum

(MD5):d4a61eee5b533939dd47bc87b54dd283

Built with DSpace-CRIS software - Extension maintained and optimized by 4Science

  • Datenschutzbestimmungen
  • Endnutzervereinbarung
  • Feedback schicken