Repository logo
 
Loading...
Thumbnail Image
Publication

DyPrune: dynamic pruning rates for neural networks

Use this identifier to reference this record.

Advisor(s)

Abstract(s)

Neural networks have achieved remarkable success in various applications such as image classification, speech recognition, and natural language processing. However, the growing size of neural networks poses significant challenges in terms of memory usage, computational cost, and deployment on resource-constrained devices. Pruning is a popular technique to reduce the complexity of neural networks by removing unnecessary connections, neurons, or filters. In this paper, we present novel pruning algorithms that can reduce the number of parameters in neural networks by up to 98% without sacrificing accuracy. This is done by scaling the pruning rate of the models to the size of the model and scheduling the pruning to execute throughout the training of the model. Code related to this work is openly available.

Description

Keywords

Machine learning Neural networks Pruning

Pedagogical Context

Citation

Jonker, Richard A.A.; Poudel, Roshan; Fajarda, Olga; Oliveira, José Luís; Lopes, Rui Pedro; Matos, Sérgio (2023). DyPrune: dynamic pruning rates for neural networks. In Progress in Artificial Intelligence (EPIA). Cham: Springer. 14115. p. 146-157. ISBN 978-3-031-49007-1

Organizational Units

Journal Issue