Purr-fect Computer Vision

This cat-inspired artificial vision system enhances object detection accuracy by adjusting for poor lighting without extra computation.

nickbild
over 1 year ago Sensors
Cats' eyes naturally adjust to low-light levels (📷: M. Kim et al.)

Modern object detection and image classification algorithms have revolutionized the field of computer vision. They have given autonomous navigation systems, such as those found in self-driving vehicles, humanoid robots, and drones, the ability to understand their surroundings in great detail. But as anyone who has worked with these technologies knows, despite their fantastic capabilities these systems can also be quite fragile. When real-world data distributions differ from what was seen in the training dataset — due to factors like lighting or glare — the algorithms often get confused and perform very poorly. And that can lead to disastrous failures at the worst possible time.

One of the primary methods used to overcome this problem involves collecting a larger, more diverse training dataset that covers these less common conditions. This does provide algorithms with the knowledge needed to recognize these situations, which makes them more resilient. However, while this approach works well for constrained problems, the full range of possible conditions in the real world cannot be completely captured in a dataset. So for applications like self-driving cars, where the unexpected is certain to happen from time to time, alternative solutions to this problem are needed.

An overview of the artificial vision system (📷: M. Kim et al.)

A team led by researchers at Seoul National University in Korea approached this problem from a different angle that eliminates the need for ever-larger training datasets. Rather than focusing on the algorithm, or the knowledge encoded in it, they set their sights on the imaging system that feeds into it. They were inspired by the eyes of cats, which have natural mechanisms that enable them to both filter out glare and enhance perception under low-light conditions. This inspiration led them to develop an artificial vision system that blurs out irrelevant details while focusing on important objects, all without any computational processing.

The artificial visual system consists of two main components: a custom optical lens with adjustable apertures and a hemispherical silicon photodetector array combined with patterned silver reflectors. The adjustable aperture, similar to a cat’s pupil, controls light intensity and enables depth of field asymmetry for targeted imaging. The hemispherical shape of the photodetector array minimizes optical distortions, reducing the complexity of lens requirements.

The patterned silver reflectors, inspired by the feline eye's tapetum lucidum, enhance light absorption efficiency by 52 percent, compensating for any limitations in the silicon image sensors. This design allows the system to adapt to different lighting conditions and break camouflage, providing high-contrast imaging.

The composition of individual photodiodes (📷: M. Kim et al.)

Tests comparing the system's performance to a conventional circular-pupil (CP) setup show that the system equipped with a variable pupil (VP) consistently outperforms the CP. In object tracking tasks, the VP system showed superior accuracy, surpassing the CP by more than 1.5 times in five out of seven metrics. For object recognition, a convolutional neural network was used on datasets like MNIST and Fashion-MNIST to assess accuracy under noisy conditions. The VP system demonstrated high recognition accuracy, particularly with noisy backgrounds, where it achieved 94.44 percent accuracy compared to 88.8 percent with the CP system.

In scenarios with grayscale images and light saturation, as seen in the Fashion-MNIST dataset, the VP system maintained a 10 percent higher accuracy rate than the CP under noisy conditions. Over 50 training epochs, the VP-equipped vision system consistently provided better predictive performance, highlighting its effectiveness in recognizing objects in both binary and grayscale environments, even when background interference is present.

One limitation of this technology is that the changes in pupil size limit the camera’s field of view. The team suggests that this may be overcome in the future by mimicking the eye and head movements seen in animals. But in any case, with some refinement, this system could help us to overcome some of the roadblocks that the field of computer vision is starting to rub up against.

nickbild

R&D, creativity, and building the next big thing you never knew you wanted are my specialties.

Latest Articles