Knowledge Distillation
How teacher-student training compresses large models into smaller, faster ones while preserving most of the original accuracy.
How teacher-student training compresses large models into smaller, faster ones while preserving most of the original accuracy.