Multi range real-time depth inference from a monocular stabilized footage using a fully convolutional neural network

Clement Pinard, Laure Chevalley, Antoine Manzanera, David Filliat

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

We propose a neural network architecture for depth map inference from monocular stabilized videos with application to UAV videos in rigid scenes. Training is based on a novel synthetic dataset for navigation that mimics aerial footage from gimbal stabilized monocular camera in rigid scenes. Based on this network, we propose a multi-range architecture for unconstrained UAV flight, leveraging flight data from sensors to make accurate depth maps for uncluttered outdoor environment. We try our algorithm on both synthetic scenes and real UAV flight data. Quantitative results are given for synthetic scenes with a slightly noisy orientation, and show that our multi-range architecture improves depth inference. Along with this article is a video that present our results more thoroughly.

Original languageEnglish
Title of host publication2017 European Conference on Mobile Robots, ECMR 2017
PublisherInstitute of Electrical and Electronics Engineers Inc.
ISBN (Electronic)9781538610961
DOIs
Publication statusPublished - 6 Nov 2017
Externally publishedYes
Event2017 European Conference on Mobile Robots, ECMR 2017 - Paris, France
Duration: 6 Sept 20178 Sept 2017

Publication series

Name2017 European Conference on Mobile Robots, ECMR 2017

Conference

Conference2017 European Conference on Mobile Robots, ECMR 2017
Country/TerritoryFrance
CityParis
Period6/09/178/09/17

Fingerprint

Dive into the research topics of 'Multi range real-time depth inference from a monocular stabilized footage using a fully convolutional neural network'. Together they form a unique fingerprint.

Cite this