June 11, 2015
How the hawkmoth sees, hovers and tracks flowers in the dark
It’s difficult enough to see things in the dark, but what if you also had to hover in mid-air while tracking a flower moving in the wind? That’s the challenge the hummingbird-sized hawkmoth (Manduca sexta) must overcome while feeding on the nectar of its favorite flowers.
Using high-speed infrared cameras and 3-D-printed robotic flowers, scientists have now learned how this insect juggles these complex sensing and control challenges – all while adjusting to changing light conditions. The work shows that the creatures can slow their brains to improve vision under low-light conditions – while continuing to perform demanding tasks.
What the researchers have discovered could help the next generation of small flying robots operate efficiently under a broad range of lighting conditions. The research is published in the June 12 edition of Science.
“There has been a lot of interest in understanding how animals deal with challenging sensing environments, especially when they are also doing difficult tasks like hovering in mid-air,” said lead author Simon Sponberg, a former University of Washington postdoctoral researcher who is now an assistant professor in the School of Physics and School of Applied Physiology at the Georgia Institute of Technology. “This is also a very significant challenge for micro air vehicles.”
The hawkmoth has been studied extensively to investigate the fundamental principles governing the development and function of its neural system, said co-author Tom Daniel, a UW biology professor and director of the new Air Force Center of Excellence on Nature-Inspired Flight Technologies and Ideas at the UW.
Daniel’s research group has experimentally characterized the response of flying hawkmoths using a sensory input comprised of the linear sum of sine waves. This new paper extends application of the “sum of sines” approach, he said.
“Simon’s work took the formal methods of control theory to dissect out how neural circuits adapt to vast ranges of luminance levels,” added Daniel. “By looking at the time delays in the movement dynamics of a freely-flying moth – interacting with the input of a robotically moved flower – Simon was able to extract the luminance dependent processing of the moth’s central nervous system.”
Scientists already knew that the moths, which feed on flower nectar during the evening and at dusk and dawn, use specialized eye structures to maximize the amount of light they can capture. But they also surmised that the insects might be slowing their nervous systems to make the best use of this limited light. But if they were slowing their brains to see better, wouldn’t that hurt their ability to hover and track the motion of flowers?
Sponberg and colleagues at the UW studied this question using high-speed infrared cameras and nectar-dispensing robotic flowers that could be moved from side-to-side at different rates. While varying both the light conditions and the frequency at which the flowers moved, the researchers studied how well free-flying moths kept their tongues – known as proboscises – in the flowers.
They also measured real flowers blowing in the wind to determine the range of motion the insects had to contend with in the wild.
“We expected to see a tradeoff with the moths doing significantly worse at tracking flowers in low light conditions,” said Sponberg. “What we saw was that while the moths did slow down, that only made a difference if the flower was moving rapidly – faster than they actually move in nature.”
In the experiments, the moths tracked robotic flowers that were oscillating at rates of up to 20 hertz – twenty oscillations per second. That was considerably faster than the two-hertz maximum rate observed in real flowers. Because the moth’s wings beat at a rate of about 25 strokes per second, they had to adjust their direction of movement with nearly every wingstroke – a major sensing, computational and control accomplishment.
“This is really an extreme behavior, though the moth makes it look simple and elegant,” said Sponberg. “To maneuver like this is really quite challenging.”
In the natural world, light intensity varies 10 billion-fold from noon on a sunny day to midnight a cloudy evening. Operating in that range of luminosity is a challenge for both moths and the sensors on human-engineered systems. Understanding how natural systems adjust to this range of conditions could therefore have broader benefits.
“If we want to have robots or machine vision systems that are working under this broad range of conditions, understanding how these moths function under these varying light conditions would be very useful,” Sponberg said.
To gather the data reported in this paper, the researchers used a robotic flower able to move in one dimension. Recently, they’ve used the actuator devices from a 3-D printer to build a robotic flower that moves in two or three dimensions, providing an additional challenge for the moths.
In future research, Sponberg and his colleagues hope to incorporate their robotic flower into a low-speed wind tunnel to study the aerodynamic challenges the moths overcome – including the role of wing vortices and the flow-effect interaction of the insect’s wings with the flowers.
Other co-authors are Jonathan Dyhr at Northwest University and Robert Hall at the UW.
The research was funded by the National Science Foundation and the Air Force Office of Scientific Research.
For more information, contact Daniel at email@example.com or 206-543-1659 and Sponberg at firstname.lastname@example.org.
This story was adapted from a Georgia Institute of Technology news release.