The Montefiore Institute Dataset of Aerial Images and Records
Mid-Air, The Montefiore Institute Dataset of Aerial Images and Records, is a multi-purpose synthetic dataset for low altitude drone flights. It provides a large amount of synchronized data corresponding to flight records for multi-modal vision sensors and navigation sensors mounted on board of a flying quadcopter. Our multi-modal vision sensors capture RGB pictures, relative surface normal orientation, depth, object semantics and stereo disparity.Additionally, each flight trajectory was recorded several times in the same place but in different climate conditions in order to change the visuals of the scene. This offers the opportunity to train algorithms for robustness to visual changes. A test set for benchmarking this particular criteria is proposed alongside the training data.Large training setOur dataset contains 79 minutes of drone flight records extracted out of more than 5 hours of flight records. Records were captured by manually flying the drone in a virtual environment thanks to an RC controller connected to the computer. The 79 minutes of flight are divided into 54 individual trajectories of equal length.Since each trajectory is rendered several times for different climate scenarios, Mid-Air offers more than 420,000 individual training frames.Multi-modal sensorsOne of the important features of Mid-Air is the types of data which are proposed. Our drone is equipped with 3 RGB cameras and records several ground-truth visual maps such as relative surface normal orientation, depth, object semantics, and stereo disparity.On top of that, our dataset provides drone positioning information. Additionally to ground truths, flight records also contain data logs for several simulated positioning sensors, i.e. accelerometer, gyroscope and GPS.Train for robustnessWith its 4 weather setups, 3 different seasons, 3 environment maps, and high quality visuals, our dataset should give a good insight on potential performances which can be expected from algorithms in real-world scenarios.Additionally, since each trajectory is recorded several times in different climate conditions, Mid-Air can be used to test the robustness of vision algorithms to visual changes. By the way, we propose a benchmark designed to assess the latter.