Skip to content
/ DiffKD Public

Official implementation for paper "Knowledge Diffusion for Distillation", NeurIPS 2023

License

Notifications You must be signed in to change notification settings

hunto/DiffKD

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

7 Commits
 
 
 
 
 
 
 
 

Repository files navigation

Knowledge Diffusion for Distillation (DiffKD)

Official implementation for paper "Knowledge Diffusion for Distillation" (DiffKD), NeurIPS 2023


Reproducing our results

git clone https://github.com/hunto/DiffKD.git --recurse-submodules
cd DiffKD

The implementation of DiffKD is in classification/lib/models/losses/diffkd.

  • classification: prepare your environment and datasets following the README.md in classification.

ImageNet

cd classification
sh tools/dist_train.sh 8 ${CONFIG} ${MODEL} --teacher-model ${T_MODEL} --experiment ${EXP_NAME}

Example script for reproducing DiffKD on ResNet-34 teacher and ResNet-18 student with B1 baseline setting:

sh tools/dist_train.sh 8 configs/strategies/distill/diffkd/diffkd_b1.yaml tv_resnet18 --teacher-model tv_resnet34 --experiment diffkd_res34_res18
  • Baseline settings (R34-R18 and R50-MBV1):
    CONFIG=configs/strategies/distill/TODO
    
    Student Teacher DiffKD MODEL T_MODEL Log Ckpt
    ResNet-18 (69.76) ResNet-34 (73.31) 72.20 tv_resnet18 tv_resnet34 log ckpt
    MobileNet V1 (70.13) ResNet-50 (76.16) 73.24 mobilenet_v1 tv_resnet50 to be reproduced

License

This project is released under the Apache 2.0 license.

Citation

@article{huang2023knowledge,
  title={Knowledge Diffusion for Distillation},
  author={Huang, Tao and Zhang, Yuan and Zheng, Mingkai and You, Shan and Wang, Fei and Qian, Chen and Xu, Chang},
  journal={arXiv preprint arXiv:2305.15712},
  year={2023}
}

About

Official implementation for paper "Knowledge Diffusion for Distillation", NeurIPS 2023

Resources

License

Stars

Watchers

Forks

Packages

No packages published