Mobile robotic systems need to perceive their surroundings in order to act independently. In this work a perception framework is developed which interprets the data of a binocular camera in order to transform it into a compact, expressive model of the environment. This model enables a mobile system to move in a targeted way and interact with its surroundings. It is shown how the developed methods also provide a solid basis for technical assistive aids for visually impaired people.
Umfang: VII, 129 S.
Preis: €41.00 | £38.00 | $72.00
These are words or phrases in the text that have been automatically identified by the Named Entity Recognition and Disambiguation service, which provides Wikipedia () and Wikidata () links for these entities.
Schwarze, T. 2018. Compact Environment Modelling from Unconstrained Camera Platforms. Karlsruhe: KIT Scientific Publishing. DOI: https://doi.org/10.5445/KSP/1000083235
Dieses Buch ist lizenziert unter Creative Commons Attribution + ShareAlike 4.0 Dedication
Dieses Buch ist Peer reviewed. Informationen dazu finden Sie hier
Veröffentlicht am 25. September 2018