Photonics Spectra BioPhotonics Vision Spectra Photonics Showcase Photonics Buyers' Guide Photonics Handbook Photonics Dictionary Newsletters Bookstore
Latest News Latest Products Features All Things Photonics Podcast
Marketplace Supplier Search Product Search Career Center
Webinars Photonics Media Virtual Events Industry Events Calendar
White Papers Videos Contribute an Article Suggest a Webinar Submit a Press Release Subscribe Advertise Become a Member


Deep Neural Network Guides Drone to Smooth Landing

Researchers at the California Institute of Technology (Caltech) have collaborated on a system that uses a deep neural network (DNN) to help autonomous drones “learn” how to land more safely and quickly, while consuming less power. The system, dubbed the Neural Lander, is a learning-based controller that tracks the position and speed of the drone and modifies its landing trajectory and rotor speed accordingly to achieve a steady landing. To make sure that the drone flies smoothly, the team employed spectral normalization to even out the neural network’s outputs so that it does not make wildly varying predictions as inputs or conditions shift. 


The Neural Lander system is tested in the Aerodrome, a three-story drone arena at Caltech’s Center for Autonomous Systems and Technologies. Courtesy of Caltech.

The researchers measured improvements in landing by examining the drone’s deviation from an idealized trajectory in 3D space. When tested, the Neural Lander was found to decrease vertical error by 100%, allowing for controlled landings, and to reduce lateral drift by up to 90%. It achieved actual landing rather than getting stuck about 10 to 15 cm above the ground, as unmodified conventional flight controllers often do. In addition, the Neural Lander produced a much smoother transition as the drone went from skimming across the table to flying in free space.

“This interdisciplinary effort brings together experts from machine learning and control systems. We have barely started to explore the rich connections between the two areas,” said professor Anima Anandkumar.

The researchers have filed a patent on the new system, which could be used for projects currently under development at Caltech’s Center for Autonomous Systems and Technologies, including an autonomous medical transport vehicle that could land in hard-to-reach locations such as in gridlocked traffic.

The research was presented at the Institute of Electrical and Electronics Engineers (IEEE) International Conference on Robotics and Automation May 22, 2019, in Montreal (https://arxiv.org/abs/1811.08027).


Engineers and computer scientists at Caltech's Center for Autonomous Systems and Technologies (CAST) use a deep neural network to help autonomous drones compensate for complex turbulence to skim and land more efficiently. Courtesy of Caltech.



Explore related content from Photonics Media




LATEST NEWS

Terms & Conditions Privacy Policy About Us Contact Us

©2024 Photonics Media