Stars
Clean, minimal, accessible reproduction of DeepSeek R1-Zero
The most no-nonsense, locally or API-hosted AI code completion plugin for Visual Studio Code - like GitHub Copilot but 100% free.
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
An LLM playground you can run on your laptop
This repo includes ChatGPT prompt curation to use ChatGPT and other LLM tools better.
ChatRWKV is like ChatGPT but powered by RWKV (100% RNN) language model, and open source.
A Unified Library for Parameter-Efficient and Modular Transfer Learning
Deep learning based content moderation from text, audio, video & image input modalities.
Letta (formerly MemGPT) is a framework for creating LLM services with memory.
Building modular LMs with parameter-efficient fine-tuning.
[CVPR 2024] Official Code for the Paper "Compositional Chain-of-Thought Prompting for Large Multimodal Models"
JARVIS-1: Open-world Multi-task Agents with Memory-Augmented Multimodal Language Models
The Cradle framework is a first attempt at General Computer Control (GCC). Cradle supports agents to ace any computer task by enabling strong reasoning abilities, self-improvment, and skill curatio…
[NeurIPS 2024 Spotlight] Buffer of Thoughts: Thought-Augmented Reasoning with Large Language Models
Label Studio is a multi-type data labeling and annotation tool with standardized output format
A curated list of awesome data labeling tools
Get up and running with Llama 3.3, DeepSeek-R1, Phi-4, Gemma 2, and other large language models.
A proof-of-concept project that showcases the potential for using small, locally trainable LLMs to create next-generation documentation tools.
MiniCPM-o 2.6: A GPT-4o Level MLLM for Vision, Speech and Multimodal Live Streaming on Your Phone
Code for the paper Watch-And-Help: A Challenge for Social Perception and Human-AI Collaboration
📋 A list of open LLMs available for commercial use.
Welcome to the Llama Cookbook! This is your go to guide for Building with Llama: Getting started with Inference, Fine-Tuning, RAG. We also show you how to solve end to end problems using Llama mode…
Experiments for "Automatic Calibration and Error Correction for Large Language Models via Pareto Optimal Self-Supervision"
WikiChat is an improved RAG. It stops the hallucination of large language models by retrieving data from a corpus.
Atmospheric adventure chat for AI language models (KoboldAI, NovelAI, Pygmalion, OpenAI chatgpt, gpt-4)
Weave is a toolkit for developing AI-powered applications, built by Weights & Biases.