New! Sign up for our free email newsletter.
Science News
from research organizations

Engineers bring efficient optical neural networks into focus

Date:
August 12, 2024
Source:
Ecole Polytechnique Fédérale de Lausanne
Summary:
Researchers have published a programmable framework that overcomes a key computational bottleneck of optics-based artificial intelligence systems. In a series of image classification experiments, they used scattered light from a low-power laser to perform accurate, scalable computations using a fraction of the energy of electronics.
Share:
FULL STORY

EPFL researchers have published a programmable framework that overcomes a key computational bottleneck of optics-based artificial intelligence systems. In a series of image classification experiments, they used scattered light from a low-power laser to perform accurate, scalable computations using a fraction of the energy of electronics.

As digital artificial intelligence systems grow in size and impact, so does the energy required to train and deploy them -- not to mention the associated carbon emissions. Recent research suggests that if current AI server production continues at its current pace, their annual energy consumption could outstrip that of a small country by 2027. Deep neural networks, inspired by the architecture of the human brain, are especially power-hungry due to the millions or even billions of connections between multiple layers of neuron-like processors.

To counteract this mushrooming energy demand, researchers have doubled down on efforts to implement optical computing systems, which have existed experimentally since the 1980s. These systems rely on photons to process data, and although light can theoretically be used to perform computations much faster and more efficiently than electrons, a key challenge has hindered optical systems' ability to surpass the electronic state-of-the art.

"In order to classify data in a neural network, each node, or 'neuron', must make a 'decision' to fire or not based on weighted input data. This decision leads to what is known as a nonlinear transformation of the data, meaning the output is not directly proportional to the input," says Christophe Moser, head of the Laboratory of Applied Photonics Devices in EPFL's School of Engineering.

Moser explains that while digital neural networks can easily perform nonlinear transformations with transistors, in optical systems, this step requires very powerful lasers. Moser worked with students Mustafa Yildirim, Niyazi Ulas Dinc, and Ilker Oguz, as well as Optics Laboratory head Demetri Psaltis, to develop an energy-efficient method for performing these nonlinear computations optically. Their new approach involves encoding data, like the pixels of an image, in the spatial modulation of a low-power laser beam. The beam reflects back on itself several times, leading to a nonlinear multiplication of the pixels.

"Our image classification experiments on three different datasets showed that our method is scalable, and up to 1,000 times more power-efficient than state-of-the-art deep digital networks, making it a promising platform for realizing optical neural networks," says Psaltis.

The research, supported by a Sinergia grant from the Swiss National Science Foundation, has recently been published in Nature Photonics.

A simple structural solution

In nature, photons do not directly interact with each other the way charged electrons do. To achieve nonlinear transformations in optical systems, scientists have therefore had to 'force' photons to interact indirectly, for example by using a light intense enough to modify the optical properties of the glass or other material it passes through.

The scientists worked around this need for a high-power laser with an elegantly simple solution: they encoded the pixels of an image spatially on the surface of a low-power laser beam. By performing this encoding twice, via adjustment of the trajectory of the beam in the encoder, the pixels are multiplied by themselves, i.e., squared. Since squaring is a non-linear transformation, this structural modification achieves the non-linearity essential to neural network calculations, at a fraction of the energy cost. This encoding can be carried out two, three or even ten times, increasing the non-linearity of the transformation and the precision of the calculation.

"We estimate that using our system, the energy required to optically compute a multiplication is eight orders of magnitude less than that required for an electronic system," Psaltis says.

Moser and Psaltis emphasize that the scalability of their low-energy approach is a major advantage, as the ultimate goal would be to use hybrid electronic-optical systems to mitigate the energy consumption of digital neural networks. However, further engineering research is needed to achieve such scale-up. For example, because optical systems use different hardware than electronic systems, a next step that the researchers are already working on is developing a compiler to translate digital data into code that optical systems can use.


Story Source:

Materials provided by Ecole Polytechnique Fédérale de Lausanne. Original written by Celia Luterbacher. Note: Content may be edited for style and length.


Journal Reference:

  1. Mustafa Yildirim, Niyazi Ulas Dinc, Ilker Oguz, Demetri Psaltis, Christophe Moser. Nonlinear processing with linear optics. Nature Photonics, 2024; DOI: 10.1038/s41566-024-01494-z

Cite This Page:

Ecole Polytechnique Fédérale de Lausanne. "Engineers bring efficient optical neural networks into focus." ScienceDaily. ScienceDaily, 12 August 2024. <www.sciencedaily.com/releases/2024/08/240812123155.htm>.
Ecole Polytechnique Fédérale de Lausanne. (2024, August 12). Engineers bring efficient optical neural networks into focus. ScienceDaily. Retrieved September 10, 2024 from www.sciencedaily.com/releases/2024/08/240812123155.htm
Ecole Polytechnique Fédérale de Lausanne. "Engineers bring efficient optical neural networks into focus." ScienceDaily. www.sciencedaily.com/releases/2024/08/240812123155.htm (accessed September 10, 2024).

Explore More

from ScienceDaily

RELATED STORIES