Stars
Ongoing research training transformer models at scale
The simplest, fastest repository for training/finetuning medium-sized GPTs.
Video+code lecture on building nanoGPT from scratch
心理健康大模型、LLM、The Big Model of Mental Health、Finetune、InternLM2、InternLM2.5、Qwen、ChatGLM、Baichuan、DeepSeek、Mixtral、LLama3、GLM4、Qwen2、LLama3.1
Get up and running with Llama 3.3, Mistral, Gemma 2, and other large language models.
Unified Efficient Fine-Tuning of 100+ LLMs (ACL 2024)
Mapping the Grokking Coding Interview Patterns to LeetCode
A minimal PyTorch re-implementation of the OpenAI GPT (Generative Pretrained Transformer) training
Reasoning in Large Language Models: Papers and Resources, including Chain-of-Thought and OpenAI o1 🍓
Minimal, clean code for the Byte Pair Encoding (BPE) algorithm commonly used in LLM tokenization.
A curated reading list for large language model (LLM) alignment. Take a look at our new survey "Large Language Model Alignment: A Survey" for more details!
A Bilingual Role Evaluation Benchmark for Large Language Models
The papers are organized according to our survey: Evaluating Large Language Models: A Comprehensive Survey.
Dependency free publish/subscribe for JavaScript
Cross-browser storage for all use cases, used across the web.
An image loading and caching library for Android focused on smooth scrolling
Demonstrate all the questions on LeetCode in the form of animation.(用动画的形式呈现解LeetCode题目的思路)
LeetCode Solutions: A Record of My Problem Solving Journey.( leetcode题解,记录自己的leetcode解题之路。)