- Project ID PercEvite
- Project duration 2017-09-01 > 2020-08-31
- Total EUR 899 007,50
- EU Contr. EUR 899 007,50
- Status Closed
We will develop a sensor, communication, and processing suite for small drones for autonomously detecting and avoiding “ground-based” obstacles and flying objects.
To avoid ground-based obstacles, we aim for a lightweight, energy-efficient sensor and processing package that maximizes payload capacity. Self-supervised learning will allow for a breakthrough in perception range. This will enable effective fusion of stereo vision, motion, appearance, ranging and audio information. Our learning process will allow obstacle detection as far as the camera ‘sees’, rather than the current ± 30 m. For close distances, our solution does without energy expensive active sensors such as lasers or sonar.
For collaborative avoidance between drones and other air vehicles, we achieve an interoperable solution by combining multiple communication hardware types (ADSB, 4/5G, WiFi) to exchange information on position, speed, and future waypoints. This will enable drones to successfully avoid other flying vehicles even in a very densely used air space. The probability for a collision in a collaborative scenario will be in the order of 10-9.
For non-collaborative avoidance, we rely on sensors and even the communication hardware mentioned above. If a non-collaborative aircraft emits communication signals, for instance to a ground station, this hardware allows to retrieve angular measurements. These measurements can be fused with detection and angle estimations performed with multiple tiny microphones and cameras on board of the detecting drone. We estimate the collision probability in a non-collaborative scenario as 10-6.
These performances will be assessed by simulations and extensive real-world tests. The consortium will benefit from the partners’ academic and industrial background with expertise in autonomous flight of very light-weight drones, robust wireless communication, drone design, production, and operation to realize a commercially viable platform.
This project has received funding from the SESAR Joint Undertaking under the European Union's Horizon 2020 research and innovation programme under grant agreement No 763702