In a new paper published on the preprint server Arxiv.org, researchers at Intel, the University of Zurich, and ETH Zurich describe an AI system that enables autonomous drones to perform acrobatics like barrel rolls, loops, and flips with only onboard sensing and computation. By training entirely in simulation and leveraging demonstrations from a controller module, the system can deploy directly onto a real-world robot without fine-tuning, according to the coauthors.
Acrobatic flight with drones is extremely challenging. Human pilots often train for years to master moves like power loops and rolls, and existing autonomous systems that perform agile maneuvers require external sensing and computation. That said, the acrobatics are worth pursuing because they represent a challenge for all of a drone’s components. Vision-based systems usually fail as a result of factors like motion blur, and the harsh requirements of high-speed fast and precise control make it difficult to tune controllers — even the tiniest mistake can result in catastrophic outcomes.
The researchers’ technique entails training the abovementioned controller to predict actions from a series of drone sensor measurements and user-defined reference trajectories. A front-facing camera image, the trajectories, and an inertial measurement serve as inputs to the system, while the output is an action in the form of thrust and angular velocity values.
The controller trains via privileged learning, where a policy learns from demonstrations provided by a so-called privileged expert. This expert has access to privileged information that isn’t available to the controller, and it’s built on a planning and control pipeline that tracks a reference trajectory from the state (i.e., the drone’s position and orientation). To facilitate the transfer from simulation to reality, the controller doesn’t access raw sensor data like color images; instead, it acts on an abstraction of the input in the form of feature points (which depend on scene structure and motion) extracted via computer vision. A series of checks ensures it doesn’t exceed the drone’s physical limits.
The coauthors chose the open source Gazebo simulator to train their policies, simulating the AscTec Hummingbird multirotor as opposed to the custom quadrotor they used in real-world experiments. They then tested the polices’ robustness by having the custom quadrotor perform a loop, roll, and flip at high accelerations and fast angular velocities.
The results over 10 training runs show that the controller managed to complete each maneuver successfully 100% of the time, without intervention or breaks. “Our approach is first to enable an autonomous flying machine to perform a wide range of acrobatics maneuvers that are highly challenging even for expert human pilots,” the researchers wrote. “We have shown that designing appropriate abstraction of the input facilities direct transfer of the policies from simulation to physical reality. The presented methodology is not limited to autonomous flight and can enable progress in other areas of robotics.”
VentureBeatVentureBeat's mission is to be a digital town square for technical decision-makers to gain knowledge about transformative technology and transact. Our site delivers essential information on data technologies and strategies to guide you as you lead your organizations. We invite you to become a member of our community, to access:
- up-to-date information on the subjects of interest to you
- our newsletters
- gated thought-leader content and discounted access to our prized events, such as Transform
- networking features, and more