Starred repositories
Learn how to design large-scale systems. Prep for the system design interview. Includes Anki flashcards.
AutoGPT is the vision of accessible AI for everyone, to use and to build on. Our mission is to provide the tools, so that you can focus on what matters.
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
Tensors and Dynamic neural networks in Python with strong GPU acceleration
Models and examples built with TensorFlow
Robust Speech Recognition via Large-Scale Weak Supervision
The most powerful and modular diffusion model GUI, api and backend with a graph/nodes interface.
scikit-learn: machine learning in Python
A natural language interface for computers
A Gradio web UI for Large Language Models with support for multiple inference backends.
The simplest, fastest repository for training/finetuning medium-sized GPTs.
An open platform for training, serving, and evaluating large language models. Release repo for Vicuna and Chatbot Arena.
A high-throughput and memory-efficient inference and serving engine for LLMs
A toolkit for developing and comparing reinforcement learning algorithms.
Composable transformations of Python+NumPy programs: differentiate, vectorize, JIT to GPU/TPU, and more
PyTorch Tutorial for Deep Learning Researchers
Code and documentation to train Stanford's Alpaca models, and generate the data.
⚡ A Fast, Extensible Progress Bar for Python and CLI
🤗 Diffusers: State-of-the-art diffusion models for image, video, and audio generation in PyTorch and FLAX.
Interactive deep learning book with multi-framework code, math, and discussions. Adopted at 500 universities from 70 countries including Stanford, MIT, Harvard, and Cambridge.
JARVIS, a system to connect LLMs with ML community. Paper: https://arxiv.org/pdf/2303.17580.pdf
Finetune Llama 3.3, DeepSeek-R1, Reasoning, Phi-4 & Gemma 2 LLMs 2x faster with 70% less memory
Image-to-Image Translation in PyTorch
Open-Sora: Democratizing Efficient Video Production for All
Code for the paper "Language Models are Unsupervised Multitask Learners"
Repository to track the progress in Natural Language Processing (NLP), including the datasets and the current state-of-the-art for the most common NLP tasks.