Inproceedings,

Align-to-Distill: Trainable Attention Alignment for Knowledge Distillation in Neural Machine Translation.

, , , , , and .
LREC/COLING, page 722-732. ELRA and ICCL, (2024)

Meta data

Tags

Users

  • @dblp

Comments and Reviews