All Stories

  1. Hierarchical Multi-Attention Transfer for Knowledge Distillation