2021-04-06发表2021-04-30更新paper reading11 分钟读完 (大约1668个字)(NewsBERT)Distilling Pre-trained Language Model for Intelligent News Application 论文笔记这篇论文介绍了知识蒸馏技术以及BERT在新闻领域进行知识蒸馏的应用阅读更多
2021-04-30(DICE) Disentangling User Interest and Conformity for Recommendation with Causal Embedding 论文阅读paper reading
2021-04-30(ESAM)Discriminative Domain Adaptation with Non-Displayed Items to Improve Long-Tail Performance 论文笔记paper reading