Search
Menu
Spectrogon US - Optical Filters 2024 LB

Nanophotonic Processor with Optical Camera Could Improve AI Efficiency

Facebook X LinkedIn Email
SEATTLE, Feb. 14, 2025 — Increasing demand for high-performance AI has engendered interest in using photonic processing instead of conventional electronic processing for AI computations. Optical computing has the potential to boost AI’s computational throughput, processing speed, and energy efficiency by orders of magnitude.

But first, optical neural networks must achieve recognition accuracy that is on par with electronic neural networks. A nanophotonic neural network, developed by researchers at the University of Washington and Princeton University, aims to overcome this limitation.

The researchers embedded parallelized optical computation into flat camera optics 4 mm in length. The camera performs neural network computations during image capture, before recording on the sensor. The team developed a spatially varying convolutional network learned through a low-dimensional reparameterization and incorporated the network inside the camera lens with a nanophotonic array with angle-dependent responses.

The nanophotonic neural network achieves an image classification accuracy of 72.76% on the CIFAR-10 database and 48.64% on ImageNet (1000-class), shrinking the gap between photonic and electronic AI while ensuring generalization to diverse vision tasks without the need to fabricate new optics.

“This is a completely new way of thinking about optics, which is very different from traditional optics,” professor Arka Majumdar said. “It’s an end-to-end design, where the optics are designed in conjunction with the computational block. Here, we replaced the camera lens with engineered optics, which allows us to put a lot of the computation into the optics.”

University of Washington professor Arka Majumdar and his students collaborated with Princeton University to build a new type of compact camera engineered for computer vision. Their prototype, shown here, uses optics for computing, significantly reducing power consumption and enabling the camera to identify objects at the speed of light. Courtesy of Ilya Chugunov/Princeton University.
The compact camera prototype, shown here, uses optics for computing, significantly reducing power consumption and enabling the camera to identify objects at the speed of light. Courtesy of Ilya Chugunov/Princeton University.
The approach to computer vision demonstrated by the researchers’ prototype could be used, for example, in autonomous vehicles, robotics, medical devices, and smartphone applications. “Nowadays, every iPhone has AI or vision technology in it,” professor Felix Heide said.

With a compact footprint and CMOS sensor compatibility, the optical system is both a photonic accelerator and an ultracompact computational camera.

Instead of a traditional lens, the camera uses an array of 50 metalenses — flat, lightweight optical components that manipulate light — to pick up different features of the object. The metalenses also function as an optical neural network.

“Our idea was to use some of the work that Arka pioneered on metasurfaces to bring some of those computations that are traditionally done electronically into the optics at the speed of light,” Heide said. “By doing so, we produced a new computer vision system that performs a lot of the computation optically.”

Lambda Research Optics, Inc. - Custom Optics

The accuracy of the system is comparable to conventionally supported neural networks. Because it performs many computations at lightspeed, it can identify and classify images more than 200 times faster than neural networks that use conventional computer hardware. The optics in the camera are powered with light instead of electricity, reducing power consumption significantly.

Instead of using a traditional camera lens made of glass or plastic, the optics in the camera rely on layers of 50 metalenses — flat, lightweight optical components that use microscopic nanostructures to manipulate light. These metalenses fit into a compact, optical computing chip, shown here. Courtesy of Ilya Chugunov/Princeton University.
Instead of using a traditional camera lens made of glass or plastic, the optics in the camera rely on layers of 50 metalenses. These metalenses fit into a compact, optical computing chip, shown here. Courtesy of Ilya Chugunov/Princeton University.
Heide and his students at Princeton provided the design for the optical chip-based camera prototype. Majumdar helped engineer the camera, and he and his students fabricated the chip in the Washington Nanofabrication Laboratory.

Majumdar and Heide said that they intend to continue their collaboration and are planning further iterations of the prototype to make it more relevant for autonomous navigation in self-driving vehicles. They also plan to work with more complex data sets and problems that require greater computing power to solve, such as object detection (i.e., locating specific objects within an image).

“Right now, this optical computing system is a research prototype, and it works for one particular application,” Majumdar said. “However, we see it eventually becoming broadly applicable to many technologies. That, of course, remains to be seen, but here, we demonstrated the first step. And it is a big step forward compared to all other existing optical implementations of neural networks.”

The nanophotonic processor and compact optical camera could strengthen recognition technology in optical neural networks, bolstering their capacity for deep learning. “There are really broad applications for this research, from self-driving cars, self-driving trucks, and other robotics to medical devices and smartphones,” Heide said. “This work is still at a very early stage, but all of these applications could someday benefit from what we are developing.”

The research was published in Science Advances (www.science.org/doi/10.1126/sciadv.adp0391).

Published: February 2025
Glossary
artificial intelligence
The ability of a machine to perform certain complex functions normally associated with human intelligence, such as judgment, pattern recognition, understanding, learning, planning, and problem solving.
deep learning
Deep learning is a subset of machine learning that involves the use of artificial neural networks to model and solve complex problems. The term "deep" in deep learning refers to the use of deep neural networks, which are neural networks with multiple layers (deep architectures). These networks, often called deep neural networks or deep neural architectures, have the ability to automatically learn hierarchical representations of data. Key concepts and components of deep learning include: ...
embedded vision
Embedded vision refers to the integration of computer vision technologies into various embedded systems, devices, or machines. Computer vision involves teaching machines to interpret and understand visual information from the world, much like human vision. Embedded vision takes this concept and applies it to systems where the processing occurs locally within the device, as opposed to relying on external servers or cloud-based services. Key components of embedded vision systems include: ...
nano
An SI prefix meaning one billionth (10-9). Nano can also be used to indicate the study of atoms, molecules and other structures and particles on the nanometer scale. Nano-optics (also referred to as nanophotonics), for example, is the study of how light and light-matter interactions behave on the nanometer scale. See nanophotonics.
nanophotonics
Nanophotonics is a branch of science and technology that explores the behavior of light on the nanometer scale, typically at dimensions smaller than the wavelength of light. It involves the study and manipulation of light using nanoscale structures and materials, often at dimensions comparable to or smaller than the wavelength of the light being manipulated. Aspects and applications of nanophotonics include: Nanoscale optical components: Nanophotonics involves the design and fabrication of...
machine vision
Machine vision, also known as computer vision or computer sight, refers to the technology that enables machines, typically computers, to interpret and understand visual information from the world, much like the human visual system. It involves the development and application of algorithms and systems that allow machines to acquire, process, analyze, and make decisions based on visual data. Key aspects of machine vision include: Image acquisition: Machine vision systems use various...
Research & TechnologyeducationAmericasUniversity of WashingtonPrinceton UniversityImagingLight SourcesOpticsSensors & DetectorslensescamerasCMOSartificial intelligencedeep learningEmbedded Visionneural networksautomotiveConsumerBiophotonicsmedicalnanonanophotonicsmachine vision

We use cookies to improve user experience and analyze our website traffic as stated in our Privacy Policy. By using this website, you agree to the use of cookies unless you have disabled them.