image: Researchers from Xidian University have developed a groundbreaking photonic transformer chip (PTC) that uses optical interference—the interaction of light waves—as its fundamental operation, replacing the multiply-accumulate units found in electronic chips. This "all-interference" approach, termed Kramers-Kronig attention (KKA), is intrinsically more efficient for executing the attention mechanism that powers modern AI.
Credit: Tian et al., PhotonIX (2025)
A research team from Xidian University has unveiled the world's first "Photonic Transformer Chip" (PTC) in the journal PhotoniX, showcasing a paradigm shift in AI hardware. This groundbreaking chip replaces traditional electronic multiplication-and-accumulation (MAC) operations with optical interference, implementing a novel all-optical attention mechanism termed "Kramers-Kronig Attention" at the hardware level.
The computational and energy demands for Transformer models, which power large language models like ChatGPT, are growing exponentially. Conventional electronic chips, constrained by power and memory bottlenecks, struggle to keep pace. While photonic computing offers high bandwidth and low power, existing architectures are primarily designed for static neural networks and are inefficient for the dynamic computations required by Transformer attention mechanisms.
This research introduces a new paradigm where "interference is all you need." It utilizes interference between optical fields to perform the core matrix operations of the attention mechanism directly. Crucially, the team leverages the inherent amplitude-phase coupling in optics—often considered a nuisance—to introduce essential nonlinearity, effectively eliminating the need for the standard SoftMax activation function.
The researchers not only provided a theoretical foundation for this mechanism using Random Fourier Features theory but also experimentally validated it on a custom-designed silicon photonic chip featuring a 10x1 optical interference unit array.
The prototype PTC demonstrated an energy efficiency of 7 TOPS/W at a 5 GHz clock frequency. Theoretical scaling analysis indicates that by expanding the core to a 512x512 interferometer array, the architecture could deliver a staggering computational throughput exceeding 200 POPS, a computational density over 1 POPS/mm² and an energy efficiency surpassing 500 TOPS/W, outperforming state-of-the-art electronic counterparts by more than two orders of magnitude.
"We are not just using light to accelerate computation; we are forging a unique path for photonic AI, establishing its own computing paradigm," said corresponding author Professor Shuiying Xiang. "This work demonstrates that for photonic neural networks, interference could indeed be all you need."
The chip achieved 94% inference accuracy on the MNIST handwritten digit recognition task, confirming its practical learning capability. This technology holds immediate promise for application in large language models, autonomous driving, and generative AI, offering a light-speed path to overcome the current AI computational bottleneck.
Journal
PhotoniX
Method of Research
Computational simulation/modeling
Subject of Research
Not applicable
Article Title
Photonic transformer chip: interference is all you need
Article Publication Date
31-Oct-2025
COI Statement
The authors declare that there is no conflict of interest.