Lists (13)
Sort Name ascending (A-Z)
Stars
code for piccolo embedding model from SenseTime
Offical Repo for "Programming Every Example: Lifting Pre-training Data Quality Like Experts at Scale"
[ICLR 2025] LongWriter: Unleashing 10,000+ Word Generation from Long Context LLMs
雅意信息抽取大模型:在百万级人工构造的高质量信息抽取数据上进行指令微调,由中科闻歌算法团队研发。 (Repo for YAYI Unified Information Extraction Model)
Freeing data processing from scripting madness by providing a set of platform-agnostic customizable pipeline processing blocks.
CoreNet: A library for training deep neural networks
[ACL 20] TACRED Revisited: A Thorough Evaluation of the TACRED Relation Extraction Task
Dataset and codes for ACL 2019 DocRED: A Large-Scale Document-Level Relation Extraction Dataset.
[LREC 2022] An off-the-shelf pre-trained Tweet NLP Toolkit (NER, tokenization, lemmatization, POS tagging, dependency parsing) + Tweebank-NER dataset
Repository for the paper "MultiNERD: A Multilingual, Multi-Genre and Fine-Grained Dataset for Named Entity Recognition (and Disambiguation)" (NAACL 2022).
A new dataset HarveyNER with fine-grained locations annotated in tweets with strong baseline models using Curriculum Learning.
The Broad Twitter Corpus, an NER dataset in English stratified for time, location, social media genre, socioeconomic factors (COLING 2016)
Guideline following Large Language Model for Information Extraction
Awesome papers about generative Information Extraction (IE) using Large Language Models (LLMs)
[EMNLP 2023 Demo] Video-LLaMA: An Instruction-tuned Audio-Visual Language Model for Video Understanding
large language model training-3-stages+deployment
This is the official repo for "PromptAgent: Strategic Planning with Language Models Enables Expert-level Prompt Optimization". PromptAgent is a novel automatic prompt optimization method that auton…
[ACL 2023] This is the code repo for our ACL'23 paper "Augmentation-Adapted Retriever Improves Generalization of Language Models as Generic Plug-In".
中文Mixtral-8x7B(Chinese-Mixtral-8x7B)
Official inference library for Mistral models
Deita: Data-Efficient Instruction Tuning for Alignment [ICLR2024]
A principled instruction benchmark on formulating effective queries and prompts for large language models (LLMs). Our paper: https://arxiv.org/abs/2312.16171
ChatGLM3 series: Open Bilingual Chat LLMs | 开源双语对话语言模型