Network pruning can skew deep learning models

North Carolina State University researchers learn that network pruning can adversely affect the performance of the model at identifying certain groups.

By Matt Shipman November 2, 2022
Image courtesy: Brett Sayles

Deep learning insights

  • Deep learning is a type of artificial intelligence (AI) that can classify things such as images, text or sound and is growing in priority in many different fields.
  • While deep learning has a lot of potential, it requires a great deal of computing power to operate.
  • North Carolina State researchers have found that network pruning can have an adverse effect on deep learning models and harm their ability to operate.

Computer science researchers have demonstrated that a widely used technique called neural network pruning can adversely affect the performance of deep learning models, detailed what causes these performance problems, and demonstrated a technique for addressing the challenge.

Deep learning is a type of artificial intelligence (AI) that can be used to classify things, such as images, text or sound. For example, it can be used to identify individuals based on facial images. However, deep learning models often require a lot of computing resources to operate. This poses challenges when a deep learning model is put into practice for some applications.

To address these challenges, some systems engage in “neural network pruning.” This effectively makes the deep learning model more compact and, therefore, able to operate while using fewer computing resources.

“However, our research shows that this network pruning can impair the ability of deep learning models to identify some groups,” said Jung-Eun Kim, co-author of a paper on the work and an assistant professor of computer science at North Carolina State University.

“For example, if a security system uses deep learning to scan people’s faces in order to determine whether they have access to a building, the deep learning model would have to be made compact so that it can operate efficiently. This may work fine most of the time, but the network pruning could also affect the deep learning model’s ability to identify some faces.”

How network pruning impairs deep learning models

In their new paper, the researchers lay out why network pruning can adversely affect the performance of the model at identifying certain groups – which the literature calls “minority groups” – and demonstrate a new technique for addressing these challenges.

Two factors explain how network pruning can impair the performance of deep learning models.

In technical terms, these two factors are: disparity in gradient norms across groups; and disparity in Hessian norms associated with inaccuracies of a group’s data. In practical terms, this means that deep learning models can become less accurate in recognizing specific categories of images, sounds or text. Specifically, the network pruning can amplify accuracy deficiencies that already existed in the model.

For example, if a deep learning model is trained to recognize faces using a data set that includes the faces of 100 white people and 60 Asian people, it might be more accurate at recognizing white faces, but could still achieve adequate performance for recognizing Asian faces. After network pruning, the model is more likely to be unable to recognize some Asian faces.

“The deficiency may not have been noticeable in the original model, but because it’s amplified by the network pruning, the deficiency may become noticeable,” Kim said. “To mitigate this problem, we’ve demonstrated an approach that uses mathematical techniques to equalize the groups that the deep learning model is using to categorize data samples. In other words, we are using algorithms to address the gap in accuracy across groups.”

In testing, the researchers demonstrated that using their mitigation technique improved the fairness of a deep learning model that had undergone network pruning, essentially returning it to pre-pruning levels of accuracy.

“I think the most important aspect of this work is that we now have a more thorough understanding of exactly how network pruning can influence the performance of deep learning models to identify minority groups, both theoretically and empirically,” Kim said.


Author Bio: Matt Shipman, research communications lead, North Carolina State University