AI
The simplest, fastest repository for training/finetuning medium-sized GPTs.
A playbook for systematically maximizing the performance of deep learning models.
A minimal PyTorch re-implementation of the OpenAI GPT (Generative Pretrained Transformer) training
⚡️ Python client for the unofficial ChatGPT API with auto token regeneration, conversation tracking, proxy support and more.
Official implementation for "Multimodal Chain-of-Thought Reasoning in Language Models" (stay tuned and more will be updated)
An unnecessarily tiny implementation of GPT-2 in NumPy.
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
为GPT/GLM等LLM大语言模型提供实用化交互接口,特别优化论文阅读/润色/写作体验,模块化设计,支持自定义快捷按钮&函数插件,支持Python和C++等项目剖析&自译解功能,PDF/LaTex论文翻译&总结功能,支持并行问询多种LLM模型,支持chatglm3等本地模型。接入通义千问, deepseekcoder, 讯飞星火, 文心一言, llama2, rwkv, claude2, m…
AIMET is a library that provides advanced quantization and compression techniques for trained neural network models.
ChatGLM-6B: An Open Bilingual Dialogue Language Model | 开源双语对话语言模型
Langchain-Chatchat(原Langchain-ChatGLM)基于 Langchain 与 ChatGLM, Qwen 与 Llama 等语言模型的 RAG 与 Agent 应用 | Langchain-Chatchat (formerly langchain-ChatGLM), local knowledge based LLM (like ChatGLM, Qwen and…
GLM-130B: An Open Bilingual Pre-Trained Model (ICLR 2023)
LlamaIndex is the leading framework for building LLM-powered agents over your data.
Code for loralib, an implementation of "LoRA: Low-Rank Adaptation of Large Language Models"
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
Robust Speech Recognition via Large-Scale Weak Supervision
Running large language models on a single GPU for throughput-oriented scenarios.
Holistic Evaluation of Language Models (HELM), a framework to increase the transparency of language models (https://arxiv.org/abs/2211.09110). This framework is also used to evaluate text-to-image …
An open platform for training, serving, and evaluating large language models. Release repo for Vicuna and Chatbot Arena.
Open deep learning compiler stack for cpu, gpu and specialized accelerators
Chinese-Vicuna: A Chinese Instruction-following LLaMA-based Model —— 一个中文低资源的llama+lora方案,结构参考alpaca
🎉 Repo for LaWGPT, Chinese-Llama tuned with Chinese Legal knowledge. 基于中文法律知识的大语言模型
Fast and memory-efficient exact attention
[NeurIPS 2023 Spotlight] LightZero: A Unified Benchmark for Monte Carlo Tree Search in General Sequential Decision Scenarios (awesome MCTS)
Platform to experiment with the AI Software Engineer. Terminal based. NOTE: Very different from https://gptengineer.app