Declarative reasoning about space and motion in visual imagery - theoretical foundations and applications
Veröffentlichungsdatum
2022-06-22
Autoren
Betreuer
Gutachter
Zusammenfassung
Perceptual sensemaking of dynamic visual imagery, e.g., involving semantic grounding, explanation, and learning, is central to a range of tasks where artificial intelligent systems have to make decisions and interact with humans. Towards this, commonsense characterisations of space and motion encompassing spatio-temporal relations, motion patterns, and events provide an abstraction layer to perform semantic reasoning about (embodied) spatio-temporal interactions observed from visuospatial imagery.
This thesis develops: (1). a general theory about space and motion for representing and reasoning about interactions founded in declaratively grounded models pertaining to space, time, space-time, motion, and events, and (2). a computational cognitive vision framework for perceptual sensemaking with visuospatial imagery, systematically developed to be compliant with declarative programming methods such as Constraint Logic Programming (CLP), Answer-Set Programming (ASP), and Inductive Logic Programming (ILP).
The thesis provides general tools and methods for declarative reasoning with visuospatial imagery, encompassing question-answering, abduction, and integration of reasoning and learning; contributed publications in this thesis focus on:
1. Grounded Semantic Interpretation and Question-Answering rooted to expressive declarative models of (embodied) visuospatial semantics to characterise (human) interactions with respect to their relational spatio-temporal structure;
2. Visuospatial Abduction, for hypothesising object interactions explaining perceived visuospatial dynamics, tightly integrating low-level (neural) visual processing and high-level (relational) abductive reasoning; and
3. Declarative Explainability and Inductive Generalisation based on declarative formalisations of visuospatial image characteristics grounded in (symbolic and subsymbolic) image elements and (neural) image features thereby providing a relational abstraction layer suitable for relational (inductive) learning.
These developed representation and reasoning capabilities are demonstrated and evaluated in the context of real-world applications (with requirements such as real-time processing, robustness against noise, etc.), where the processing and semantic interpretation of (potentially large volumes of) highly dynamic visuospatial imagery is central. Example applications included in this thesis encompass cognitive robotics, autonomous vehicles, and assistive technologies for human behaviour research.
This thesis develops: (1). a general theory about space and motion for representing and reasoning about interactions founded in declaratively grounded models pertaining to space, time, space-time, motion, and events, and (2). a computational cognitive vision framework for perceptual sensemaking with visuospatial imagery, systematically developed to be compliant with declarative programming methods such as Constraint Logic Programming (CLP), Answer-Set Programming (ASP), and Inductive Logic Programming (ILP).
The thesis provides general tools and methods for declarative reasoning with visuospatial imagery, encompassing question-answering, abduction, and integration of reasoning and learning; contributed publications in this thesis focus on:
1. Grounded Semantic Interpretation and Question-Answering rooted to expressive declarative models of (embodied) visuospatial semantics to characterise (human) interactions with respect to their relational spatio-temporal structure;
2. Visuospatial Abduction, for hypothesising object interactions explaining perceived visuospatial dynamics, tightly integrating low-level (neural) visual processing and high-level (relational) abductive reasoning; and
3. Declarative Explainability and Inductive Generalisation based on declarative formalisations of visuospatial image characteristics grounded in (symbolic and subsymbolic) image elements and (neural) image features thereby providing a relational abstraction layer suitable for relational (inductive) learning.
These developed representation and reasoning capabilities are demonstrated and evaluated in the context of real-world applications (with requirements such as real-time processing, robustness against noise, etc.), where the processing and semantic interpretation of (potentially large volumes of) highly dynamic visuospatial imagery is central. Example applications included in this thesis encompass cognitive robotics, autonomous vehicles, and assistive technologies for human behaviour research.
Schlagwörter
Declarative space and motion
;
Cognitive vision
;
Visuospatial sensemaking
;
Vision and semantics
;
Commonsense reasoning
;
Knowledge representation and reasoning
;
Human-centred AI
Institution
Fachbereich
FB3 - Mathematik/Informatik
Dokumenttyp
Dissertation
Zweitveröffentlichung
Nein
Sprache
Englisch
Dateien![Vorschaubild]()
Lade...
Name
doctoral_thesis-jakob_suchan-declarative_reasoning_about_space_and_motion_in_visual_imagery_PDFA.pdf
Size
35.44 MB
Format
Adobe PDF
Checksum
(MD5):a022152becf765dc226768326024b1e6