【20230414组会汇报】MiniLM: Deep self-attention distillation for task-agnostic compression of pre-trained transformers

论文原文链接:https://proceedings.neurips.cc/paper/2020/file/3f5ee243547dee91fbd053c1c4a845aa-Paper.pdf

借用这篇博文测试一下可不可以放PPT,今天突然发现已经引用830了,好牛

#此处放PPT,加载略慢,请耐心等待#