News Release

Less is more: Efficient pruning for reducing AI memory and computational cost

The emergence of advanced AI systems that solve complex problems requires large memory and great computation costs. This reality begs the question: Is there a smart and effective way of reducing that cost?

Peer-Reviewed Publication

Bar-Ilan University

How does deep learning work?

video: 

Researchers from Bar-Ilan University have developed a groundbreaking method to drastically reduce the size and energy consumption of deep learning systems—without compromising performance. Published in Physical Review E, their study reveals that by better understanding how deep networks learn, it's possible to prune up to 90% of parameters in certain layers while maintaining accuracy. This advancement, led by Prof. Ido Kanter and PhD student Yarden Tzach, could make AI more efficient, sustainable, and scalable for real-world applications.  

 

view more 

Credit: Prof. Ido Kanter, Bar-Ilan University

Deep learning and AI systems have made great headway in recent years, especially in their capabilities of automating complex computational tasks such as image recognition, computer vision and natural language processing. Yet, these systems consist of billions of parameters and require great memory usage as well as expensive computational cost.

This reality raises the question: Can we optimize, or more correctly, prune, the parameters in those systems without compromising their capabilities? In a study just published in Physical Review E by researchers from Bar-Ilan University, the answer is a resounding yes.

In the article the researchers show how better understanding of the mechanism underlying successful deep learning leads to an efficient pruning of unnecessary parameters in a deep architecture without affecting its performance.

“It all hinges on an initial understanding of what happens in deep networks, how they learn and what parameters are essential to its learning,” said Prof. Ido Kanter, of Bar-Ilan's Department of Physics and Gonda (Goldschmied) Multidisciplinary Brain Research Center, who led the research. “It’s the ever-present reality of scientific research, the more we know the better we understand, and in turn, the better and more efficient the technology we can create.”

“There are many methods that attempt to improve memory and data usage,” said PhD student Yarden Tzach, a key contributor to this research. “They were able to improve memory usage and computational complexity, but our method was able to prune up to 90% of the parameters of certain layers, without hindering the system’s accuracy at all!” he added.

These results can lead to better usage of AI systems, both in memory as well as energy consumption. As AI becomes more and more prevalent in our day to day lives, reducing its energy cost will be of utmost importance.

A video on advanced pruning in deep learning: https://www.youtube.com/watch?reload=9&v=IzfpNPKSFCc


Disclaimer: AAAS and EurekAlert! are not responsible for the accuracy of news releases posted to EurekAlert! by contributing institutions or for the use of any information through the EurekAlert system.