"Knowledge Distillation Using Output Errors for Self-attention End-to-end ..."

Ho-Gyeong Kim et al. (2019)
a service of Schloss Dagstuhl - Leibniz Center for Informatics