An extension version of our CVPR 2020, oral: HRank: Filter Pruning using High-Rank Feature Map (Link).
Prior code version can be found here.
Any problem, please contact the authors via emails: [email protected] or [email protected]. Do not post issues with github as much as possible, just in case that I could not receive the emails from github thus ignore the posted issues.
If you find HRank useful in your research, please consider citing:
@inproceedings{lin2020hrank,
title={HRank: Filter Pruning using High-Rank Feature Map},
author={Lin, Mingbao and Ji, Rongrong and Wang, Yan and Zhang, Yichen and Zhang, Baochang and Tian, Yonghong and Shao, Ling},
booktitle={Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)},
pages={1529--1538},
year={2020}
}
In this code, you can run our models on CIFAR-10 and ImageNet dataset. (Task on object detection is preparing)
For rank generation, the code has been tested by Python 3.6, Pytorch 1.0 and CUDA 9.0 on Ubuntu 16.04. And for evaluating(fine-tuning), we recommend you to run with PyTorch 1.6 for the usage of Thop.
python rank_generation.py \
--dataset [dataset name] \
--data_dir [dataset dir] \
--pretrain_dir [pre-trained model dir] \
--arch [model arch name] \
--limit [batch numbers] \
--gpu [gpu_id]
For the ease of reproducibility, we provide the extracted ranks here:
For model training, our code will automatically save the previous-epoch checkpoint file in checkpoint.pth.tar
and best checkpoint in model_best.pth.tar
after every epoch. And with the resume
option, you can resume training from the previous checkpoint under the same job directory.
python evaluate_cifar.py \
--data_dir [CIFAR-10 dataset dir] \
--job_dir ./result/[model name]/[folder name] \
--arch [model name](vgg_16_bn, resnet_56, resnet_110, googlenet, densenet_40) \
--rank_conv_prefix [rank folder dir] \
--use_pretrain \
--pretrain_dir [pre-trained model dir] \
--compress_rate [compress rate] \
--gpu [gpu_id]
python evaluate.py \
--data_dir [ImageNet dataset dir] \
--use_dali \
--job_dir ./result/[model name]/[folder name] \
--arch [model name](resnet_50, mobilenet_v1, mobilenet_v2) \
--rank_conv_prefix [rank folder dir] \
--lr_type (cos or step) \
--use_pretrain \
--pretrain_dir [pre-trained model dir] \
--compress_rate [compress rate] \
--gpu [gpu_id]
optional arguments:
--data_dir Dataset directory.
--job_dir The directory where the summaries will be stored.
--rank_conv_prefix Prefix directory of rank folder.
--compress_rate Compress rate of each convolutional layer.
--arch Architecture of model. default: vgg_16_bn
Optional: resnet_50, mobilenet_v2, mobilenet_v1, vgg_16_bn, resnet_56, resnet_110, densenet_40, googlenet
--learning_rate Initial learning rate. default: 0.1
--lr_type Learning rate decay schedule. default: step. optional: step, cos
--momentum Momentum for optimizer. default: 0.9
--epochs The number of epochs to train. default: 90
--batch_size Batch size for both training and validation. default: 64
--weight_decay The weight decay of loss function. default: 1e-4
--label_smooth Label smooth parameter. default: 0.1
--resume If this parameter exists, model training by resuming from previous ckpt in the same directory.
--use_dali If this parameter exists, use dali module to load ImageNet data.
--use_pretrain If this parameter exists, use pretrained model for finetuning.
--pretrain_dir Pretrained model directory.
--test_only If this parameter exists, only validate the model performance without training.
--test_model_dir Test model directory
--gpu Select gpu to use. default: 0
The following are the examples of compression rate setting for several models: (Please note that the following compression rates are only used to demonstrate the parameter format, which may not be used in our experiment. For the specific pruning rate, please refer to the configuration files of pruned model in the next section)
Model | Compression Rate |
---|---|
VGG-16-BN | [0.45]*7+[0.78]*5 |
ResNet-56 | [0.]+[0.18]*29 |
ResNet-110 | [0.]+[0.2]*2+[0.3]*18+[0.35]*36 |
GoogLeNet | [0.3]+[0.6]*2+[0.7]*5+[0.8]*2 |
DenseNet-40 | [0.]+[0.4]*12+[0.2]+[0.5]*12+[0.3]+[0.6]*12 |
ResNet-50 | [0.1]+[0.2]*3+[0.5]*16 |
MobileNet-V1 | [0.]+[0.3]*12 |
MobileNet-V2 | [0.]+[0.3]*7 |
With the compression rate, our training module (evaluate_cifar.py and evaluate.py) can automatically calculate the params and FLOPs of that model and record them in the training logger.
After training, a total of four files can be found in the job_dir
, including best model, final model, config file and logger file.
We provide our pre-trained models, and pruned models. For your ease of reproducibility, the training loggers and configuration files are attached with the url link as well.
(The percentages in parentheses indicate the pruning rate)
Architecture | Params | Flops | Accuracy | Model |
---|---|---|---|---|
VGG-16-BN(Baseline) | 14.98M(0.0%) | 313.73M(0.0%) | 93.96% | pre-trained |
VGG-16-BN | 2.76M(81.6%) | 131.17M(58.1%) | 93.73%(-0.23%) | pruned |
VGG-16-BN | 2.50M(83.3%) | 104.78M(66.6%) | 93.56%(-0.40%) | pruned |
VGG-16-BN | 1.90M(87.3%) | 66.95M(78.6%) | 93.10%(-0.86%) | pruned |
ResNet-56(Baseline) | 0.85M(0.0%) | 125.49M(0.0%) | 93.26% | pre-trained |
ResNet-56 | 0.66M(22.3%) | 90.35M(28.0%) | 93.85%(+0.59%) | pruned |
ResNet-56 | 0.48M(42.8%) | 65.94M(47.4%) | 93.57%(+0.31%) | pruned |
ResNet-56 | 0.24M(70.0%) | 34.78M(74.1%) | 92.32%(-0.94%) | pruned |
ResNet-110(Baseline) | 1.72M(0.0%) | 252.89M(0.0%) | 93.50% | pre-trained |
ResNet-110 | 1.04M(39.1%) | 140.54M(44.4%) | 94.20%(+0.70%) | pruned |
ResNet-110 | 0.72M(58.1%) | 101.97M(59.6%) | 93.81%(+0.31%) | pruned |
ResNet-110 | 0.54M(68.3%) | 71.69M(71.6%) | 93.23%(-0.27%) | pruned |
GoogLeNet(Baseline) | 6.15M(0.0%) | 1520M(0.0%) | 95.05% | pre-trained |
GoogLeNet | 2.85M(53.5%) | 649.19M(57.2%) | 95.04%(-0.01%) | pruned |
GoogLeNet | 2.09M(65.8%) | 395.42M(73.9%) | 94.82%(-0.23%) | pruned |
DenseNet-40(Baseline) | 1.04M(0.0%) | 282.00M(0.0%) | 94.81% | pre-trained |
DenseNet-40 | 0.62M(40.1%) | 173.39M(38.5%) | 94.51%(-0.30%) | pruned |
DenseNet-40 | 0.45M(56.5%) | 133.17M(52.7%) | 93.91%(-0.90%) | pruned |
DenseNet-40 | 0.39M(61.9%) | 113.08M(59.9%) | 93.66%(-1.21%) | pruned |
Architecture | Params | Flops | Use Dali | Lr Type | Top-1 Accuracy | Top-5 Accuracy | Model |
---|---|---|---|---|---|---|---|
ResNet-50(baseline) | 25.55M(0.0%) | 4.11B(0.0%) | 76.15% | 92.87% | pre-trained | ||
ResNet-50 | 15.09M(40.8%) | 2.26B(44.8%) | No | step | 75.56% | 92.63% | pruned |
ResNet-50 | 11.05M(56.7%) | 1.52B(62.8%) | No | step | 74.19% | 91.94% | pruned |
ResNet-50 | 8.02M(68.6%) | 0.95B(76.7%) | No | step | 72.30% | 90.74% | pruned |
ResNet-50 | 11.05M(56.7%) | 1.52B(62.8%) | No | cos | 75.83% | 92.85% | pruned |
ResNet-50 | 11.81M(53.7%) | 1.38B(66.2%) | No | cos | 75.60% | 92.61% | pruned |
ResNet-50 | 6.56M(74.3%) | 1.02B(75.0%) | No | cos | 74.24% | 91.8% | pruned |
To verify our model performance, please use the command below (make sure you are using the corresponding compression rate in the configuration file of that model):
python evaluate_cifar.py \
--data_dir [CIFAR-10 dataset dir] \
--job_dir ./result/[model name]/[folder name] \
--arch [model name](vgg_16_bn, resnet_56, resnet_110, googlenet, densenet_40) \
--test_only \
--test_model_dir [test model dir] \
--compress_rate [compress rate] \
--gpu [gpu_id]
python evaluate.py \
--data_dir [ImageNet dataset dir] \
--job_dir ./result/[model name]/[folder name] \
--arch [model name](resnet_50, mobilenet_v1, mobilenet_v2) \
--test_only \
--test_model_dir [test model dir] \
--compress_rate [compress rate] \
--gpu [gpu_id]
Additionally, we provide the pre-trained models used in our experiments.
Vgg-16 | ResNet56 | ResNet110 | DenseNet-40 | GoogLeNet