TY - GEN
T1 - ONE-CYCLE PRUNING
T2 - 29th IEEE International Conference on Image Processing, ICIP 2022
AU - Hubens, Nathan
AU - Mancas, Matei
AU - Gosselin, Bernard
AU - Preda, Marius
AU - Zaharia, Titus
N1 - Publisher Copyright:
© 2022 IEEE.
PY - 2022/1/1
Y1 - 2022/1/1
N2 - Introducing sparsity in a convnet has been an efficient way to reduce its complexity while keeping its performance almost intact. Most of the time, sparsity is introduced using a three-stage pipeline: 1) training the model to convergence, 2) pruning the model, 3) fine-tuning the pruned model to recover performance. The last two steps are often performed iteratively, leading to reasonable results but also to a time-consuming process. In our work, we propose to remove the first step of the pipeline and to combine the two others in a single training-pruning cycle, allowing the model to jointly learn the optimal weights while being pruned. We do this by introducing a novel pruning schedule, named One-Cycle Pruning (OCP), which starts pruning from the beginning of the training, and until its very end. Experiments conducted on a variety of combinations between architectures (VGG-16, ResNet-18), datasets (CIFAR-10, CIFAR-100, Caltech-101), and sparsity values (80%, 90%, 95%) show that not only OCP consistently outperforms common pruning schedules such as One-Shot, Iterative and Automated Gradual Pruning, but also that it drastically reduces the required training budget. Moreover, experiments following the Lottery Ticket Hypothesis show that OCP allows to find higher quality and more stable pruned networks.
AB - Introducing sparsity in a convnet has been an efficient way to reduce its complexity while keeping its performance almost intact. Most of the time, sparsity is introduced using a three-stage pipeline: 1) training the model to convergence, 2) pruning the model, 3) fine-tuning the pruned model to recover performance. The last two steps are often performed iteratively, leading to reasonable results but also to a time-consuming process. In our work, we propose to remove the first step of the pipeline and to combine the two others in a single training-pruning cycle, allowing the model to jointly learn the optimal weights while being pruned. We do this by introducing a novel pruning schedule, named One-Cycle Pruning (OCP), which starts pruning from the beginning of the training, and until its very end. Experiments conducted on a variety of combinations between architectures (VGG-16, ResNet-18), datasets (CIFAR-10, CIFAR-100, Caltech-101), and sparsity values (80%, 90%, 95%) show that not only OCP consistently outperforms common pruning schedules such as One-Shot, Iterative and Automated Gradual Pruning, but also that it drastically reduces the required training budget. Moreover, experiments following the Lottery Ticket Hypothesis show that OCP allows to find higher quality and more stable pruned networks.
KW - Neural Network Compression
KW - Neural Network Pruning
KW - Pruning Schedule
UR - https://www.scopus.com/pages/publications/85142447798
U2 - 10.1109/ICIP46576.2022.9897980
DO - 10.1109/ICIP46576.2022.9897980
M3 - Conference contribution
AN - SCOPUS:85142447798
T3 - Proceedings - International Conference on Image Processing, ICIP
SP - 4128
EP - 4132
BT - 2022 IEEE International Conference on Image Processing, ICIP 2022 - Proceedings
PB - IEEE Computer Society
Y2 - 16 October 2022 through 19 October 2022
ER -