Summary
We will develop a sensor, communication, and processing suite for small drones for autonomously detecting and avoiding “ground-based” obstacles and flying objects.
To avoid ground-based obstacles, we aim for a lightweight, energy-efficient sensor and processing package that maximizes payload capacity. Self-supervised learning will allow for a breakthrough in perception range. This will enable effective fusion of stereo vision, motion, appearance, ranging and audio information. Our learning process will allow obstacle detection as far as the camera ‘sees’, rather than the current ± 30 m. For close distances, our solution does without energy expensive active sensors such as lasers or sonar.
For collaborative avoidance between drones and other air vehicles, we achieve an interoperable solution by combining multiple communication hardware types (ADSB, 4/5G, WiFi) to exchange information on position, speed, and future waypoints. This will enable drones to successfully avoid other flying vehicles even in a very densely used air space. The probability for a collision in a collaborative scenario will be in the order of 10-9.
For non-collaborative avoidance, we rely on sensors and even the communication hardware mentioned above. If a non-collaborative aircraft emits communication signals, for instance to a ground station, this hardware allows to retrieve angular measurements. These measurements can be fused with detection and angle estimations performed with multiple tiny microphones and cameras on board of the detecting drone. We estimate the collision probability in a non-collaborative scenario as 10-6.
These performances will be assessed by simulations and extensive real-world tests. The consortium will benefit from the partners’ academic and industrial background with expertise in autonomous flight of very light-weight drones, robust wireless communication, drone design, production, and operation to realize a commercially viable platform.
To avoid ground-based obstacles, we aim for a lightweight, energy-efficient sensor and processing package that maximizes payload capacity. Self-supervised learning will allow for a breakthrough in perception range. This will enable effective fusion of stereo vision, motion, appearance, ranging and audio information. Our learning process will allow obstacle detection as far as the camera ‘sees’, rather than the current ± 30 m. For close distances, our solution does without energy expensive active sensors such as lasers or sonar.
For collaborative avoidance between drones and other air vehicles, we achieve an interoperable solution by combining multiple communication hardware types (ADSB, 4/5G, WiFi) to exchange information on position, speed, and future waypoints. This will enable drones to successfully avoid other flying vehicles even in a very densely used air space. The probability for a collision in a collaborative scenario will be in the order of 10-9.
For non-collaborative avoidance, we rely on sensors and even the communication hardware mentioned above. If a non-collaborative aircraft emits communication signals, for instance to a ground station, this hardware allows to retrieve angular measurements. These measurements can be fused with detection and angle estimations performed with multiple tiny microphones and cameras on board of the detecting drone. We estimate the collision probability in a non-collaborative scenario as 10-6.
These performances will be assessed by simulations and extensive real-world tests. The consortium will benefit from the partners’ academic and industrial background with expertise in autonomous flight of very light-weight drones, robust wireless communication, drone design, production, and operation to realize a commercially viable platform.
Unfold all
/
Fold all
More information & hyperlinks
Web resources: | https://cordis.europa.eu/project/id/763702 |
Start date: | 01-09-2017 |
End date: | 31-08-2020 |
Total budget - Public funding: | 899 007,50 Euro - 899 007,00 Euro |
Cordis data
Original description
We will develop a sensor, communication, and processing suite for small drones for autonomously detecting and avoiding “ground-based” obstacles and flying objects.To avoid ground-based obstacles, we aim for a lightweight, energy-efficient sensor and processing package that maximizes payload capacity. Self-supervised learning will allow for a breakthrough in perception range. This will enable effective fusion of stereo vision, motion, appearance, ranging and audio information. Our learning process will allow obstacle detection as far as the camera ‘sees’, rather than the current ± 30 m. For close distances, our solution does without energy expensive active sensors such as lasers or sonar.
For collaborative avoidance between drones and other air vehicles, we achieve an interoperable solution by combining multiple communication hardware types (ADSB, 4/5G, WiFi) to exchange information on position, speed, and future waypoints. This will enable drones to successfully avoid other flying vehicles even in a very densely used air space. The probability for a collision in a collaborative scenario will be in the order of 10-9.
For non-collaborative avoidance, we rely on sensors and even the communication hardware mentioned above. If a non-collaborative aircraft emits communication signals, for instance to a ground station, this hardware allows to retrieve angular measurements. These measurements can be fused with detection and angle estimations performed with multiple tiny microphones and cameras on board of the detecting drone. We estimate the collision probability in a non-collaborative scenario as 10-6.
These performances will be assessed by simulations and extensive real-world tests. The consortium will benefit from the partners’ academic and industrial background with expertise in autonomous flight of very light-weight drones, robust wireless communication, drone design, production, and operation to realize a commercially viable platform.
Status
CLOSEDCall topic
RPAS-07Update Date
27-10-2022
Images
No images available.
Geographical location(s)
Structured mapping