Study of three major model compression techniques: Knowledge distillation, Quantization, and Pruning.
This was a part of class project for the Deep Learning course of the graduate AI program in Oregon State University. Please refer to the report PDF in the repository for more details on the experiments.