![awesome logo](https://raw.githubusercontent.com/github/explore/80688e429a7d4ef2fca1e82350fe8e3517d3494d/topics/awesome/awesome.png)
Starred repositories
A latent text-to-image diffusion model
21 Lessons, Get Started Building with Generative AI 🔗 https://microsoft.github.io/generative-ai-for-beginners/
The repository provides code for running inference with the SegmentAnything Model (SAM), links for downloading the trained model checkpoints, and example notebooks that show how to use the model.
Google Research
CLIP (Contrastive Language-Image Pretraining), Predict the most relevant text snippet given an image
Grounded SAM: Marrying Grounding DINO with Segment Anything & Stable Diffusion & Recognize Anything - Automatically Detect , Segment and Generate Anything
Natural Language Processing Tutorial for Deep Learning Researchers
《李宏毅深度学习教程》(李宏毅老师推荐👍,苹果书🍎),PDF下载地址:https://github.com/datawhalechina/leedl-tutorial/releases
The repository provides code for running inference with the Meta Segment Anything Model 2 (SAM 2), links for downloading the trained model checkpoints, and example notebooks that show how to use th…
AISystem 主要是指AI系统,包括AI芯片、AI编译器、AI推理和训练框架等AI全栈底层技术
Implementation of Dreambooth (https://arxiv.org/abs/2208.12242) with Stable Diffusion
Using Low-rank adaptation to quickly fine-tune diffusion models.
Inpaint anything using Segment Anything and inpainting models.
A unified framework for 3D content generation.
The image prompt adapter is designed to enable a pretrained text-to-image diffusion model to generate images with image prompt.
CoTracker is a model for tracking any point (pixel) on a video.
Notebooks using the Hugging Face libraries 🤗
An open-source project dedicated to tracking and segmenting any objects in videos, either automatically or interactively. The primary algorithms utilized include the Segment Anything Model (SAM) fo…
Let us democratise high-resolution generation! (CVPR 2024)
PyTorch implementation for Score-Based Generative Modeling through Stochastic Differential Equations (ICLR 2021, Oral)
[ICML 2024] Mastering Text-to-Image Diffusion: Recaptioning, Planning, and Generating with Multimodal LLMs (RPG)
Unofficial implementation of "Prompt-to-Prompt Image Editing with Cross Attention Control" with Stable Diffusion
[ICCV 2023 Oral] "FateZero: Fusing Attentions for Zero-shot Text-based Video Editing"
Official Pytorch Implementation for "MultiDiffusion: Fusing Diffusion Paths for Controlled Image Generation" presenting "MultiDiffusion" (ICML 2023)
Code for the paper "Visual Anagrams: Generating Multi-View Optical Illusions with Diffusion Models"
bert-base-chinese example
[ECCV 2024] HiDiffusion: Increases the resolution and speed of your diffusion model by only adding a single line of code!