Researchers at MIT have developed a photonic processor that can perform deep neural network computations optically on a chip, promising faster and more energy-efficient machine learning for demanding applications like lidar, astronomy, and high-speed telecommunications.
The device, fabricated using commercial foundry processes, overcomes previous limitations by integrating nonlinear optical function units (NOFUs) to implement nonlinear operations on the chip. Led by Saumil Bandyopadhyay, a visiting scientist at MIT’s Research Laboratory of Electronics, the team demonstrated a fully integrated photonic processor that can complete key computations in under half a nanosecond with over 92 percent accuracy.
This breakthrough could enable real-time learning and high-speed processing for applications where speed and energy efficiency are critical. The research, published in Nature Photonics, was funded by the US National Science Foundation, the US Air Force Office of Scientific Research, and NTT Research, and involved collaboration with companies like Nokia and Periplous.
“This work demonstrates that computing — at its essence, the mapping of inputs to outputs — can be compiled onto new architectures of linear and nonlinear physics that enable a fundamentally different scaling law of computation versus effort needed,” .
Dirk Englund, a professor in the Department of Electrical Engineering and Computer Science, principal investigator of the Quantum Photonics and Artificial Intelligence Group and of RLE
Photonic Processors: A New Frontier for Ultrafast AI Computations
The development of photonic processors has opened up new possibilities for ultrafast artificial intelligence (AI) computations with extreme energy efficiency. Researchers from MIT and elsewhere have demonstrated a fully integrated photonic processor that can perform all key computations of a deep neural network optically on the chip, paving the way for faster and more energy-efficient deep learning.
Overcoming Limitations of Traditional Computing Hardware
Deep neural network models have grown increasingly complex, pushing the limits of traditional electronic computing hardware. Photonic hardware, which uses light to process information, offers a promising solution to this problem. However, one major challenge in developing photonic processors is implementing nonlinear operations on the chip. Nonlinear operations, such as activation functions, are essential for deep neural networks to learn complex patterns.
Breakthrough: Nonlinear Optical Function Units (NOFUs)
To overcome this challenge, researchers designed devices called nonlinear optical function units (NOFUs), which combine electronics and optics to implement nonlinear operations on the chip. NOFUs siphon off a small amount of light to photodiodes that convert optical signals to electric current, eliminating the need for an external amplifier and consuming very little energy.
A Fully-Integrated Optical Deep Neural Network
The researchers built an optical deep neural network on a photonic chip using three layers of devices that perform linear and nonlinear operations. The system encodes the parameters of a deep neural network into light, which is then processed through an array of programmable beamsplitters and NOFUs. This enables the system to stay in the optical domain until the end, when the answer is read out, achieving ultra-low latency.
Achieving Ultra-Low Latency and High Accuracy
The photonic system achieved more than 96 percent accuracy during training tests and more than 92 percent accuracy during inference, comparable to traditional hardware. Moreover, the chip performs key computations in less than half a nanosecond, enabling efficient training of deep neural networks on the chip.
Scalability and Future Directions
The entire circuit was fabricated using the same infrastructure and foundry processes that produce CMOS computer chips, making it possible to manufacture at scale with minimal error. Future work will focus on scaling up the device and integrating it with real-world electronics like cameras or telecommunications systems. Researchers also aim to explore algorithms that can leverage the advantages of optics to train systems faster and with better energy efficiency.
Implications for AI and Beyond
This breakthrough has far-reaching implications for AI research, enabling faster and more efficient processing of complex data sets. Moreover, it demonstrates that computing can be compiled onto new architectures of linear and nonlinear physics, enabling a fundamentally different scaling law of computation versus effort needed. As researchers continue to push the boundaries of photonic processors, we may see significant advancements in fields beyond AI, such as telecommunications, navigation, and more.
External Link: Click Here For More
