Light Processing Improves Robotic Sensing, Army Research Finds

Editor’s Note: U.S. Army researchers have uncovered an avenue to improving sensors in robots. A story from Space Daily details that this discovery will enable autonomous robots to become even better teammates with their human counterparts. Let’s face it, if robots can measure up to the lofty standards of our military—where autonomy is a top priority—these advancements should someday find their way to utilization in stores.

A team of Army researchers uncovered how the human brain processes bright and contrasting light, which they say is a key to improving robotic sensing and enabling autonomous agents to team with humans.

To enable developments in autonomy, a top Army priority, machine sensing must be resilient across changing environments, researchers said.

"When we develop machine vision algorithms, real-world images are usually compressed to a narrower range, as a cellphone camera does, in a process called tone mapping," said Andre Harrison, a researcher at the U.S. Army Combat Capabilities Development Command's Army Research Laboratory. "This can contribute to the brittleness of machine vision algorithms because they are based on artificial images that don't quite match the patterns we see in the real world."

By developing a new system with 100,000-to-1 display capability, the team discovered the brain's computations, under more real-world conditions, so they could build biological resilience into sensors, Harrison said.

Current vision algorithms are based on human and animal studies with computer monitors, which have a limited range in luminance of about 100-to-1, the ratio between the brightest and darkest pixels. In the real world, that variation could be a ratio of 100,000-to-1, a condition called high dynamic range, or HDR.

"Changes and significant variations in light can challenge Army systems — drones flying under a forest canopy could be confused by reflectance changes when wind blows through the leaves, or autonomous vehicles driving on rough terrain might not recognize potholes or other obstacles because the lighting conditions are slightly different from those on which their vision algorithms were trained," said Army researcher Dr. Chou Po Hung.

The research team sought to understand how the brain automatically takes the 100,000-to-1 input from the real world and compresses it to a narrower range, which enables humans to interpret shape. The team studied early visual processing under HDR, examining how simple features like HDR luminance and edges interact, as a way to uncover the underlying brain mechanisms.

"The brain has more than 30 visual areas, and we still have only a rudimentary understanding of how these areas process the eye's image into an understanding of 3D shape," Hung said.

"Our results with HDR luminance studies, based on human behavior and scalp recordings, show just how little we truly know about how to bridge the gap from laboratory to real-world environments. But, these findings break us out of that box, showing that our previous assumptions from standard computer monitors have limited ability to generalize to the real world, and they reveal principles that can guide our modeling toward the correct mechanisms."

The Journal of Vision published the team's research findings, Abrupt darkening under high dynamic range (HDR) luminance invokes facilitation for high contrast targets and grouping by luminance similarity.

Researchers said the discovery of how light and contrast edges interact in the brain's visual representation will help improve the effectiveness of algorithms for reconstructing the true 3D world under real-world luminance, by correcting for ambiguities that are unavoidable when estimating 3D shape from 2D information.

"Through millions of years of evolution, our brains have evolved effective shortcuts for reconstructing 3D from 2D information," Hung said. "It's a decades-old problem that continues to challenge machine vision scientists, even with the recent advances in AI."

In addition to vision for autonomy, this discovery will also be helpful to develop other AI-enabled devices such as radar and remote speech understanding that depend on sensing across wide dynamic ranges.

With their results, the researchers are working with partners in academia to develop computational models, specifically with spiking neurons that may have advantages for both HDR computation and for more power-efficient vision processing - both important considerations for low-powered drones.

"The issue of dynamic range is not just a sensing problem," Hung said.

"It may also be a more general problem in brain computation because individual neurons have tens of thousands of inputs. How do you build algorithms and architectures that can listen to the right inputs across different contexts? We hope that, by working on this problem at a sensory level, we can confirm that we are on the right track, so that we can have the right tools when we build more complex AIs."

 

This article was from SpaceDaily.com and was legally licensed through the Industry Dive publisher network. Please direct all licensing questions to legal@industrydive.com.