Multi-objective evolutionary optimization for hardware-aware neural network pruning
en-GBde-DEes-ESfr-FR

Multi-objective evolutionary optimization for hardware-aware neural network pruning

25/09/2024 TranSpread

Neural network pruning is a key technique for deploying artificial intelligence (AI) models based on deep neural networks (DNNs) on resource-constrained platforms, such as mobile devices. However, hardware conditions and resource availability vary greatly across different platforms, making it essential to design pruned models optimally suited to specific hardware configurations. Hardware-aware neural network pruning offers an effective way to automate this process, but it requires balancing multiple conflicting objectives, such as network accuracy, inference latency, and memory usage, that traditional mathematical methods struggle to solve.

In a study (doi: https://doi.org/10.1016/j.fmre.2022.07.013) published in the journal Fundamental Research, a group of researchers from Shenzhen, China, present a novel hardware-aware neural network pruning approach based on multi-objective evolutionary optimization.

“We propose to employ Multi-Objective Evolutionary Algorithms (MOEAs) to solve the hardware neural network pruning problem,” shares Ke Tang, senior and corresponding author of the study.

Compared with conventional optimization algorithms, MOEAs have two advantages in tackling this problem. One is that MOEAs do not require particular assumptions like differentiability or continuity and possess strong capacity for black-box optimization. The other is their ability to find multiple Pareto-optimal solutions in a single simulation run, which is very useful in practice because it offers flexibility to meet different user requirements.

Specifically, once such a set of solutions has been found. End users can easily choose their preferred configurations of DNN compression, such as latency first or memory consumption first, with just one click on the corresponding solutions,” adds Tang.

The team's findings further revealed that, while multi-objective evolutionary algorithms hold significant potential, they still struggle with low search efficiency. To that end, the researchers developed a new MOEA, namely Hardware-Aware Multi-objective evolutionary network Pruning (HAMP), to address this issue.

“It is a memetic MOEA that combines an efficient portfolio-based selection and a surrogate-assist local search operator. HAMP is currently the only network pruning approach that can effectively handle multiple hardware direct feedback and accuracy simultaneously.” explains first author Wenjing Hong. “Experimental studies on the mobile NVIDIA Jetson Nano demonstrate the effectiveness of HAMP over the state-of-the-art and the potential of MOEAs for hardware-aware network pruning.”

The team’s results show that HAMP not only manages to achieve solutions that are better on all objectives, but also delivers simultaneously a set of alternative solutions.

“These solutions present different trade-offs between latency, memory consumption, and accuracy, and hence can facilitate a rapid deployment of DNNs in practice,” concludes Hong.

###

References

DOI

10.1016/j.fmre.2022.07.013

Original Source URL

https://doi.org/10.1016/j.fmre.2022.07.013

Funding information

This work was supported by grants from the National Natural Science Foundation of China (62106098), the Stable Support Plan Program of Shenzhen Natural Science Fund (20200925154942002), and the MOE University Scientific-Technological Innovation Plan Program.

About Fundamental Research

Fundamental Research is an open access, peer-reviewed, multidisciplinary journal, which is supervised by the National Natural Science Foundation of China (NSFC). Published bimonthly, it features high-calibre research covering all areas of the natural sciences and high-tech fields.

Paper title: Multi‐objective evolutionary optimization for hardware‐aware neural network pruning
Attached files
  • MULTI-OBJECTIVE EVOLUTIONARY OPTIMIZATION FOR HARDWARE-AWARE NEURAL NETWORK PRUNING.
25/09/2024 TranSpread
Regions: North America, United States, Asia, China
Keywords: Applied science, Computing

Disclaimer: AlphaGalileo is not responsible for the accuracy of news releases posted to AlphaGalileo by contributing institutions or for the use of any information through the AlphaGalileo system.

Testimonials

For well over a decade, in my capacity as a researcher, broadcaster, and producer, I have relied heavily on Alphagalileo.
All of my work trips have been planned around stories that I've found on this site.
The under embargo section allows us to plan ahead and the news releases enable us to find key experts.
Going through the tailored daily updates is the best way to start the day. It's such a critical service for me and many of my colleagues.
Koula Bouloukos, Senior manager, Editorial & Production Underknown
We have used AlphaGalileo since its foundation but frankly we need it more than ever now to ensure our research news is heard across Europe, Asia and North America. As one of the UK’s leading research universities we want to continue to work with other outstanding researchers in Europe. AlphaGalileo helps us to continue to bring our research story to them and the rest of the world.
Peter Dunn, Director of Press and Media Relations at the University of Warwick
AlphaGalileo has helped us more than double our reach at SciDev.Net. The service has enabled our journalists around the world to reach the mainstream media with articles about the impact of science on people in low- and middle-income countries, leading to big increases in the number of SciDev.Net articles that have been republished.
Ben Deighton, SciDevNet

We Work Closely With...


  • BBC
  • The Times
  • National Geographic
  • The University of Edinburgh
  • University of Cambridge
  • iesResearch
Copyright 2024 by AlphaGalileo Terms Of Use Privacy Statement