Deep Compression of Pre-trained Transformer Models.
NEURIPS 2022(2022)
关键词
Quantization,Sparsity,Pruning,Pre-trained,Transformer,Foundation Model,Inference,NLP,vision,speech,BERT,Wav2vec,ViT
AI 理解论文
溯源树
样例

生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要
NEURIPS 2022(2022)