技术栈

akd

__如果
1 年前
论文阅读·人工智能·深度学习·知识蒸馏·nas·akd
论文阅读--Search to DistillStandard Knowledge Distillation (KD) approaches distill the knowledge of a cumbersome teacher model into the parameters of a student model with a pre-defined architecture. However, the knowledge of a neural network, which is represented by the network’s o