MiniLMv2: Multi-Head Self-Attention Relation Distillation for Compressing Pretrained Transformers

Wenhui Wang | Hangbo Bao | Shaohan Huang | Li Dong | Furu Wei |

Paper Details:

Month: August
Year: 2021
Location: Online
Venue: Findings |

Citations

URL