Stars
Scalable RL solution for advanced reasoning of language models
Recipes to train reward model for RLHF.
🔥中文 prompt 精选🔥,ChatGPT 使用指南,提升 ChatGPT 可玩性和可用性!🚀
[NeurIPS 2024] SimPO: Simple Preference Optimization with a Reference-Free Reward
Tools for merging pretrained large language models.
Codes of our MICCAI 2023 paper "DARC: Distribution-Aware Re-Coloring Model for Generalizable Nucleus Segmentation"
A repository for ebooks, including C, C plus plus, Linux Kernel, Compiler, OS, Algorithm, Security, Database, Network, ML and DL
[ICLR 2024] COLLIE: Systematic Construction of Constrained Text Generation Tasks
The RedPajama-Data repository contains code for preparing large datasets for training large language models.
SeqGPT: An Out-of-the-box Large Language Model for Open Domain Sequence Understanding
InsTag: A Tool for Data Analysis in LLM Supervised Fine-tuning
[NAACL'24] Self-data filtering of LLM instruction-tuning data using a novel perplexity-based difficulty score, without using any other models
🐋 An unofficial implementation of Self-Alignment with Instruction Backtranslation.
A series of large language models developed by Baichuan Intelligent Technology
The official repo of Qwen (通义千问) chat & pretrained large language model proposed by Alibaba Cloud.
Awesome Pretrained Chinese NLP Models,高质量中文预训练模型&大模型&多模态模型&大语言模型集合
Reference implementation for DPO (Direct Preference Optimization)
ChatGLM2-6B: An Open Bilingual Chat LLM | 开源双语对话语言模型
Hackable and optimized Transformers building blocks, supporting a composable construction.
TigerBot: A multi-language multi-task LLM
A large-scale 7B pretraining language model developed by BaiChuan-Inc.
LLMs build upon Evol Insturct: WizardLM, WizardCoder, WizardMath
This repository contains code to quantitatively evaluate instruction-tuned models such as Alpaca and Flan-T5 on held-out tasks.
OpenLLaMA, a permissively licensed open source reproduction of Meta AI’s LLaMA 7B trained on the RedPajama dataset