Stars
SGLang is a fast serving framework for large language models and vision language models.
A chatbot/GraphRAG framework that creates multi-llm-agents from social platform user comments and let them debate on specific topics.
Implement a ChatGPT-like LLM in PyTorch from scratch, step by step
Advanced Multi-Turn QA System with LLM and Intent Recognition. 基于LLM大语言模型意图识别、参数抽取结合slot词槽技术实现多轮问答、NL2API. 打造Function Call多轮问答最佳实践
State-of-the-Art Deep Learning scripts organized by models - easy to train and deploy with reproducible accuracy and performance on enterprise-grade infrastructure.
A one stop repository for generative AI research updates, interview resources, notebooks and much more!
🧑🏫 60+ Implementations/tutorials of deep learning papers with side-by-side notes 📝; including transformers (original, xl, switch, feedback, vit, ...), optimizers (adam, adabelief, sophia, ...), ga…
Data processing for and with foundation models! 🍎 🍋 🌽 ➡️ ➡️🍸 🍹 🍷
Dense X Retrieval: What Retrieval Granularity Should We Use?
FinQwen: 致力于构建一个开放、稳定、高质量的金融大模型项目,基于大模型搭建金融场景智能问答系统,利用开源开放来促进「AI+金融」。
利用lightgbm做(learning to rank)排序学习,包括数据处理、模型训练、模型决策可视化、模型可解释性以及预测等。Use LightGBM to learn ranking, including data processing, model training, model decision visualization, model interpretability and …
[SIGIR 2022] The official repo for the paper "Curriculum Learning for Dense Retrieval Distillation".
A set of examples around pytorch in Vision, Text, Reinforcement Learning, etc.
Code, Dataset, and Pretrained Models for Audio and Speech Large Language Model "Listen, Think, and Understand".
A high-throughput and memory-efficient inference and serving engine for LLMs
[ACL'23 Findings] This is the code repo for our ACL'23 Findings paper "ReGen: Zero-Shot Text Classification via Training Data Generation with Progressive Dense Retrieval".
🦜🔗 Build context-aware reasoning applications
Unified Efficient Fine-Tuning of 100+ LLMs & VLMs (ACL 2024)
手把手带你实战 Huggingface Transformers 课程视频同步更新在B站与YouTube
Code and documentation to train Stanford's Alpaca models, and generate the data.
Toolkit for fine-tuning, ablating and unit-testing open-source LLMs.
llama2 finetuning with deepspeed and lora