Hu, W., Gibson, P. and Cano, J. (2023) ICE-Pick: Iterative Cost-Efficient Pruning for DNNs. ICML 2023 Workshop on Neural Compression, Honolulu, Hawaii, USA, 28-29 July 2023.
![]() |
Text
303684.pdf - Published Version 1MB |
Abstract
Pruning is one of the main compression methods for Deep Neural Networks (DNNs), where less relevant parameters are removed from a DNN model to reduce its memory footprint. To get better final accuracy, pruning is often performed iteratively with increasing amounts of parameters being removed in each step, and fine-tuning (i.e., additional training epochs) being applied to the remaining parameters. However, this process can be very time-consuming, since the finetuning process is applied after every pruning step and calculates gradients for the whole model. Motivated by these overheads, in this paper we propose ICE-Pick, a novel threshold-guided finetuning method which freezes less sensitive layers and leverages a custom pruning-aware learning rate scheduler. We evaluate ICE-Pick using ResNet-110, ResNet-152, and MobileNetV2 (all defined for CIFAR-10), and show that it can save up to 87.6% of the pruning time while maintaining accuracy.
Item Type: | Conference or Workshop Item |
---|---|
Status: | Published |
Refereed: | Yes |
Glasgow Author(s) Enlighten ID: | Cano Reyes, Dr Jose and Gibson, Mr Perry |
Authors: | Hu, W., Gibson, P., and Cano, J. |
College/School: | College of Science and Engineering > School of Computing Science |
Copyright Holders: | Copyright 2023 by the author(s). |
Publisher Policy: | Reproduced with the permission of the authors |
Related URLs: |
University Staff: Request a correction | Enlighten Editors: Update this record