Can knowledge distillation be used for 3 classes?

后端 未结 0 648
轮回少年
轮回少年 2021-02-02 10:55

I have a classification problem (3 classes only) and I trained a teacher-student model.

The trained student model (with knowledge distillation) performs less than the stu

相关标签:
回答
  • 消灭零回复
提交回复
热议问题