Skip to content

Knowledge Distillation: CVPR2020 Oral, Revisiting Knowledge Distillation via Label Smoothing Regularization

License

Notifications You must be signed in to change notification settings

leoozy/Teacher-free-Knowledge-Distillation

 
 

Repository files navigation

Teacher-free-Knowledge-Distillation

Implementation for our paper: Revisiting Knowledge Distillation via Label Smoothing Regularization, arxiv

Our work suggests that: when a neural network is too powerful to find stronger teacher models, or computation resource is limited to train teacher models, "self-training" or "manually-designed regularization" can be applied.

For example, ResNeXt101-32x8d is a powerful model with 88.79M parameters and 16.51G FLOPs on ImageNet, and it is hard or computation expensive to train a stronger teacher model for this student. Our strategy can further improve this powerful student model by 0.48% without extra computation on ImageNet. Similarly, when taking a powerful single model ResNeXt29-8x64d with 34.53M parameters as a student model, our self-training implementation achieves more than 1.0% improvement on CIFAR100 (from 81.03% to 82.08%).

1. Preparations

Clone this repository:

git clone https://github.com/yuanli2333/Teacher-free-Knowledge-Distillation.git

1.1 Environment

Build a new environment and install:

pip install -r requirements.txt

Better use: NVIDIA GPU + CUDA9.0 + Pytorch 1.2.0

Please do not use other versions of pytorch, otherwise, some experiment results may not be reproduced because some slight difference would make the hyper-parameters different.

1.2 Dataset

CIFAR100, CIFAR10 and Tiny_ImageNet; For CIFAR100 and CIFAR10, our codes will download the datasets automatically. For Tiny-ImageNet, you should download and put in the dir: "data/". The follow instruction and commands are for CIFAR100.

2. Train baseline models

You can skip this step by using our pre-trained models in here. Download and unzip to: experiments/pretrained_teacher_models/

Use ''--model_dir'' to specify the directory of "parameters", model saving and log saving.

For example, normally train ResNet18 to obtain the pre-trained teacher:

CUDA_VISIBLE_DEVICES=0 python main.py --model_dir experiments/base_experiments/base_resnet18/

We ignore the command ''CUDA_VISIBLE_DEVICES=gpu_id'' in the following commands

Normally train MobileNetV2 to obtain the baseline model and baseline accuracy:

python main.py --model_dir experiments/base_experiments/base_mobilenetv2/

Normally train ResNeXt29 to obtain the baseline model and baseline accuracy:

python main.py --model_dir experiments/base_experiments/base_resnext29/

The baseline accuracy (in %) on CIFAR100 is:

Model Baseline Acc
MobileNetV2 68.38
ShuffleNetV2 70.34
ResNet18 75.87
ResNet50 78.16
GoogLeNet 78.72
Desenet121 79.04
ResNeXt29 81.03

3. Exploratory experiments (Section 2 in our paper)

3.1 Reversed KD (Re-KD)

Normal KD: ResNet18 teach MobileNetV2

python main.py --model_dir experiments/kd_experiments/mobilenet_v2_distill/resnet18_teacher/

Re-KD: MobileNetV2 teach ResNet18

python main.py --model_dir experiments/kd_experiments/resnet18_distill/mobilenet_v2_teacher/

Re-KD: ShuffleNetV2 teach ResNet18

python main.py --model_dir experiments/kd_experiments/resnet18_distill/shufflenet_v2_teacher/

Re-KD experiment results on CIFAR100:

3.2 Defective KD (De-KD)

Use the arguments "--pt_teacher" to switch to Defective KD experiment.

For expample, use a pooly-trained Teacher ResNet18 with 15.48% accuracy (just trained one epoch) to teach MobileNetV2:

python main.py --model_dir experiments/kd_experiments/mobilenet_v2_distill/resnet18_teacher/ --pt_teacher

Use one-epoch trained teacher ResNet18 to teach ShuffleNetV2:

python main.py --model_dir experiments/kd_experiments/shufflenet_v2_distill/resnet18_teacher/ --pt_teacher

Use 50-epoch-trained teacher ResNet50 (acc:45.82%) to teach ShuffleNetV2:

python main.py --model_dir experiments/kd_experiments/shufflenet_v2_distill/resnet50_teacher/ --pt_teacher

Use 50-epoch-trained teacher ResNet50 (acc:45.82%) to teach ResNet18:

python main.py --model_dir experiments/kd_experiments/resnet18_distill/resnet50_teacher/ --pt_teacher

Use 50-epoch-trained teacher ResNeXt29 (acc:51.94%) to teach MobileNetV2:

python main.py --model_dir experiments/kd_experiments/mobilenet_v2_distill/resnext29_teacher/ --pt_teacher

De-KD experiment results on CIFAR100:

4. Teacher-free KD (Tf-KD) (Section 5 in our paper)

We have two implementations to achieve Tf-KD, the first one is self-training, the second one is manually-designed teacher(regularization)

4.1 Tf-KD self-training

Use the arguments ''--self_training'' to control this training. The --model_dir should be experiment dirctory, should be "experiments/kd_experiments/student/student_self_teacher"

MobileNetV2 self training:

python main.py --model_dir experiments/kd_experiments/mobilenet_v2_distill/mobilenet_self_teacher/ --self_training

ShuffleNetV2 self training:

python main.py --model_dir experiments/kd_experiments/shufflenet_v2_distill/shufflenet_self_teacher/ --self_training

ResNet18 self training:

python main.py --model_dir experiments/kd_experiments/resnet18_distill/resnet18_self_teacher/ --self_training

Our method achieve more than 1.0% improvement for a big single model ResNeXt29, run self-training for ResNeXt29:

python main.py --model_dir experiments/kd_experiments/resnext29_distill/resnext29_self_teacher/ --self_training

Tf-KD self-training experiment results on CIFAR100:

4.2 Tf-KD manually-designed teacher(regularization)

MobileNetV2 taught by mannually-designed regularization:

python main.py --model_dir experiments/base_experiments/base_mobilenetv2/  --regularization

ShuffleNetV2 taught by mannually-designed regularization:

python main.py --model_dir experiments/base_experiments/base_shufflenetv2/ --regularization

ResNet18 taught by mannually-designed regularization:

python main.py --model_dir experiments/base_experiments/base_resnet18/  --regularization

GoogLeNet taught by mannually-designed regularization:

python main.py --model_dir experiments/base_experiments/base_googlenet/ --regularization

4.3 Lable Smoothing Regularization

MobileNetV2 Lable Smoothing:

python main.py --model_dir experiments/base_experiments/base_mobilenetv2/  --label_smoothing

ShuffleNetV2 Lable Smoothing:

python main.py --model_dir experiments/base_experiments/base_shufflenetv2/ --label_smoothing

Tf-KD regularization and LSR experiment results on CIFAR100:

Reference

If you find this repo useful, please consider citing:

@inproceedings{yuan2020revisiting,
  title={Revisiting Knowledge Distillation via Label Smoothing Regularization},
  author={Yuan, Li and Tay, Francis EH and Li, Guilin and Wang, Tao and Feng, Jiashi},
  booktitle={Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition},
  pages={3903--3911},
  year={2020}
}

About

Knowledge Distillation: CVPR2020 Oral, Revisiting Knowledge Distillation via Label Smoothing Regularization

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 100.0%