MixKD: Towards Efficient Distillation of Large-scale Language Models

Kevin J Liang, Weituo Hao, Dinghan Shen, Yufan Zhou, Weizhu Chen, Changyou Chen, Lawrence Carin

Research output: Contribution to journalArticlepeer-review

11 Downloads (Pure)

Fingerprint

Dive into the research topics of 'MixKD: Towards Efficient Distillation of Large-scale Language Models'. Together they form a unique fingerprint.

Engineering

Computer Science

Keyphrases