Stars
🚀 Efficient implementations of state-of-the-art linear attention models in Torch and Triton
Muon optimizer: +>30% sample efficiency with <3% wallclock overhead
An AI-powered research assistant that performs iterative, deep research on any topic by combining search engines, web scraping, and large language models. The goal of this repo is to provide the si…
(Supports DeepSeek R1) An AI-powered research assistant that performs iterative, deep research on any topic by combining search engines, web scraping, and large language models.
A lightweight data processing framework built on DuckDB and 3FS.
基于大模型搭建的聊天机器人,同时支持 微信公众号、企业微信应用、飞书、钉钉 等接入,可选择GPT3.5/GPT-4o/GPT-o1/ DeepSeek/Claude/文心一言/讯飞星火/通义千问/ Gemini/GLM-4/Claude/Kimi/LinkAI,能处理文本、语音和图片,访问操作系统和互联网,支持基于自有知识库进行定制企业智能客服。
🤖一个基于 WeChaty 结合 DeepSeek / ChatGPT / Kimi / 讯飞等Ai服务实现的微信机器人 ,可以用来帮助你自动回复微信消息,或者管理微信群/好友,检测僵尸粉等...
Integrate the DeepSeek API into popular softwares
DeepEP: an efficient expert-parallel communication library
Production-tested AI infrastructure tools for efficient AGI development and community-driven innovation
Scalable RL solution for advanced reasoning of language models
Official Repo for Open-Reasoner-Zero
MoBA: Mixture of Block Attention for Long-Context LLMs
Democratizing Reinforcement Learning for LLMs
Clean, minimal, accessible reproduction of DeepSeek R1-Zero
This is a replicate of DeepSeek-R1-Zero and DeepSeek-R1 training on small models with limited data
Codebase for Iterative DPO Using Rule-based Rewards
verl: Volcano Engine Reinforcement Learning for LLMs
TransMLA: Multi-Head Latent Attention Is All You Need
A Flexible Framework for Experiencing Cutting-edge LLM Inference Optimizations
Official implementation of paper: SFT Memorizes, RL Generalizes: A Comparative Study of Foundation Model Post-training
Reasoning in LLMs: Papers and Resources, including Chain-of-Thought, OpenAI o1, and DeepSeek-R1 🍓
Fully open reproduction of DeepSeek-R1
SOTA RL fine-tuning solution for advanced math reasoning of LLM