It's got praying mantis eyes


By Jennifer McManamay, University of Virginia School of Engineering and Applied Science
Thursday, 01 August, 2024

It's got praying mantis eyes

Self-driving cars occasionally crash because their visual systems can’t always process static or slow-moving objects in 3D space. In that regard, they’re like the monocular vision of many insects, whose compound eyes provide great motion-tracking and a wide field of view but poor depth perception.

Except for the praying mantis.

A praying mantis’s field of view also overlaps between its left and right eyes, creating binocular vision with depth perception in 3D space.

Combining this insight with some nifty optoelectrical engineering and innovative “edge” computing — processing data in or near the sensors that capture it — researchers at the University of Virginia School of Engineering and Applied Science have developed artificial compound eyes that overcome vexing limitations in the way machines currently collect and process real-world visual data.

These limitations include accuracy issues, data processing lag times and the need for substantial computational power.

“After studying how praying mantis eyes work, we realised a biomimetic system that replicates their biological capabilities required developing new technologies,” said Byungjoon Bae, a PhD candidate in the Charles L. Brown Department of Electrical and Computer Engineering.

About those biomimetic peepers

The team’s meticulously designed “eyes” mimic nature by integrating microlenses and multiple photodiodes, which produce an electrical current when exposed to light. The team used flexible semiconductor materials to emulate the convex shapes and faceted positions within mantis eyes.

A photograph of the artificial compound eye prototype. Image credit: UVA Engineering.

“Making the sensor in hemispherical geometry while maintaining its functionality is a state-of-the-art achievement, providing a wide field of view and superior depth perception,” Bae said.

“The system delivers precise spatial awareness in real time, which is essential for applications that interact with dynamic surroundings.”

Such uses include low-power vehicles and drones, self-driving vehicles, robotic assembly, surveillance and security systems, and smart home devices.

Bae, whose adviser is Kyusang Lee, an associate professor in the department with a secondary appointment in materials science and engineering, is first author of the team’s recent paper in Science Robotics.

Among the team’s important findings on the lab’s prototype system was a potential reduction in power consumption by more than 400 times compared to traditional visual systems.

Benefits of computing on the edge

Rather than using cloud computing, Lee’s system can process visual information in real time, nearly eliminating the time and resource costs of data transfer and external computation, while minimising energy usage.

“The technological breakthrough of this work lies in the integration of flexible semiconductor materials, conformal devices that preserve the exact angles within the device, an in-sensor memory component, and unique post-processing algorithms,” Bae said.

The key is that the sensor array continuously monitors changes in the scene, identifying which pixels have changed and encoding this information into smaller data sets for processing.

The approach mirrors how insects perceive the world through visual cues, differentiating pixels between scenes to understand motion and spatial data. For example, like other insects — and humans, too — the praying mantis can process visual data rapidly by using the phenomenon of motion parallax, in which nearer objects appear to move faster than distant objects. Only one eye is needed to achieve the effect, but motion parallax alone isn’t sufficient for accurate depth perception.

Praying mantis eyes are special because, like us, they use stereopsis — seeing with both eyes to perceive depth — in addition to their hemispherical compound eye geometries and motion parallax to understand their surroundings.

“The seamless fusion of these advanced materials and algorithms enables real-time, efficient and accurate 3D spatiotemporal perception,” said Lee, a prolific early-career researcher in thin-film semiconductors and smart sensors.

“Our team’s work represents a significant scientific insight that could inspire other engineers and scientists by demonstrating a clever, biomimetic solution to complex visual processing challenges,” he said.

This is a modified version of a news item published by the University of Virginia School of Engineering and Applied Science. The original version of the news item can be accessed here.

Top image credit: iStock.com/Valentin Baciu

Related News

Using quantum dots for solid-state optical cooling

Researchers have investigated a cooling phenomenon in a halide perovskite-based...

Zero-gravity tech could enable circuit repairs in space

Researchers have developed a special ink and 3D printer that can function in zero gravity.

Electron microscopy reveals colours of outermost electron layer

Researchers have used atomic-resolution secondary electron (SE) imaging to capture the atomic...


  • All content Copyright © 2024 Westwick-Farrow Pty Ltd