Skip to content

code of the paper "instance temporary knowledge distillation"

License

Notifications You must be signed in to change notification settings

Zhengbo-Zhang/ITKD

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

13 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

ITKD

code for the paper "instance temporary knowledge distillation"

Project page | arXiv

installation

# create env
conda create -n ITKD python=3.9
conda activate ITKD
# pytorch version
conda install pytorch==1.11.0 torchvision==0.12.0 torchaudio==0.11.0 cudatoolkit=11.3 -c pytorch
# other dependencies
pip install -r requirements.txt

Weights

weights are putted into pretrained folder.

Training on cifar100

All training scripts reside within the file train.sh. Adapt this file to align with your specific environment, then proceed to select a configuration from the available options. Note that the dataset will be automatically downloaded during the execution phase.

# modify the content inside and just run it
bash train.sh

Inference on cifar100

The testing scripts mirror the training scripts in their structure and usage. Simply modify the content within the scripts to harmonize with your environment.

# modify the content inside and just run it
bash test.sh

About

code of the paper "instance temporary knowledge distillation"

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published