ITHACA, N.Y. – A new innovation from Cornell University researchers lowers the energy use needed to power artificial intelligence – a step toward shrinking the carbon footprints of data centers and AI infrastructure.
As AI systems become increasingly powerful, they also become more power-hungry – raising questions about sustainability. The research team is tackling that challenge by rethinking the hardware that powers AI, aiming to make it faster, more efficient and less energy-intensive.
The researchers received a Best Paper Award for their findings, presented at the 2025 International Conference on Field-Programmable Logic and Applications, held from Sept. 1 to 5 in Leiden, Netherlands.
Their work focuses on a type of computer chip called a Field-Programmable Gate Array (FPGA). Unlike traditional chips, they can be reprogrammed for different tasks after manufacturing. This makes them especially useful in rapidly evolving fields such as AI, cloud computing and wireless communication.
“FPGAs are everywhere – from network cards and communication base stations to ultrasound machines, CAT scans, and even washing machines,” said co-author Mohamed Abdelfattah, assistant professor at Cornell Tech. “AI is coming to all of these devices, and this architecture helps make that transition more efficient.”
Inside each FPGA chip are computing units called logic blocks. These blocks contain components that can handle different types of computing. Lookup Tables (LUTs) are components that can conduct a wide range of logical operations depending on what the chip needs to do. Adder chains are components that perform fast arithmetic operations such as adding numbers – essential for tasks like image recognition and natural language processing.
In conventional FPGA designs, these components are tightly linked, meaning the adder chains can only be accessed through the LUTs. This limits the chip’s efficiency, especially for AI workloads that rely heavily on arithmetic operations.
The research team developed “Double Duty,” a new chip architecture, to address this problem. The design allows LUTs and adder chains to work independently and simultaneously within the same logic block. In other words, the chip can now do more with the exact same processing resources.
This innovation is particularly impactful for deep neural networks, AI models that mimic the human brain’s processing of information. These models are often “unrolled” onto FPGAs – laid out as fixed circuits for faster, more efficient processing.
“We focused on a mode where FPGAs are actually really good at AI acceleration,” said Abdelfattah. “By making a small architectural change, we make these unrolled neural networks much more efficient, playing to the strengths of FPGAs instead of treating them like generic processors.”
In testing, the Double Duty design reduced the space needed for specific AI tasks by more than 20% and improved overall performance on a large suite of circuits by nearly 10%. That means fewer chips could be used to perform the same work, resulting in lower energy use.
For additional information, read this Cornell Chronicle story.
-30-