Here we list further (or related, preliminary) readings for the course. Due to our limited time to prepare for the course, we must have missed many excellent resources. Please help us improve and expand the list by creating pull requests! If we find it useful, we will actively merge your pull request (and of course, thus you will be listed as one of the contributors to the course).
- 【视频】CS11-747 CMU自然语言处理课程
- 【视频】CS224n 斯坦福自然语言处理课程
- 【视频】大模型概况简介
- 【网页】十个自然语言处理大模型
- 【论文】Distributed Representations of Words and Phrases and their Compositionality
- 【论文】基础模型综述(李飞飞等)
- 【论文】BMInf: An Efficient Toolkit for Big Model Inference and Tuning
- 【网页】BMTrain:为大模型训练计算成本节省9成
- 【网页】不止于ZeRO:BMTrain技术原理浅析
- 【论文】ZeRO: Memory Optimizations Toward Training Trillion Parameter Models
- 【论文】TinyBERT: Distilling BERT for Natural Language Understanding
- 【论文】Structured Pruning Learns Compact and Accurate Models
- 【论文】MoEfication: Transformer Feed-forward Layers are Mixtures of Experts
-
信息检索
-
阅读理解/问答
-
文本生成
- 【论文】AlphaFold2: Highly accurate protein structure prediction with AlphaFold
- 【论文】Enformer: Effective gene expression prediction from sequence by integrating long-range interactions
- 【论文】DNABERT: pre-trained Bidirectional Encoder Representations from Transformers model for DNA-language in genome
- 【论文】biomedical NLP综述: Domain-Specific Language Model Pretraining for Biomedical Natural Language Processing.
- 【工具包】TorchDrug
- 【论文】How does NLP benefit legal system: A summary of legal artificial intelligence
- 【论文】Lawformer: A pre-trained language model for chinese legal long documents
- 【论文】LeCaRD: a legal case retrieval dataset for Chinese law system
- 【论文】LEVEN: A Large-Scale Chinese Legal Event Detection Dataset
- 【论文】LEGAL-BERT: The muppets straight out of law school
- 【论文】MoEfication: Transformer Feed-forward Layers are Mixtures of Experts
- 【论文】Finding Experts in Transformer Models
- 【论文】Interpreting and improving natural-language processing (in machines) with natural language-processing (in the brain)
- 【论文】Exploring Universal Intrinsic Task Subspace via Prompt Tuning
- 【论文】On Transferability of Prompt Tuning for Natural Language Understanding