Inproceedings,

Layer-wise Pruning of Transformer Attention Heads for Efficient Language Modeling.

, , , and .
ISOCC, page 357-358. IEEE, (2021)

Meta data

Tags

Users

  • @dblp

Comments and Reviews