Stars
We present StableAnimator, the first end-to-end ID-preserving video diffusion framework, which synthesizes high-quality videos without any post-processing, conditioned on a reference image and a se…
EchoMimic: Lifelike Audio-Driven Portrait Animations through Editable Landmark Conditioning
[CVPR'24] DiffSHEG: A Diffusion-Based Approach for Real-Time Speech-driven Holistic 3D Expression and Gesture Generation
chinese speech pretrained models
Public code release associated with SceneScript.
A modern Neovim configuration with full battery for Python, Lua, C++, Markdown, LaTeX, and more...
Blazing fast Neovim framework providing solid defaults and a beautiful UI, enhancing your neovim experience.
Over 385 terminal color schemes/themes for iTerm/iTerm2. Includes ports to Terminal, Konsole, PuTTY, Xresources, XRDB, Remmina, Termite, XFCE, Tilda, FreeBSD VT, Terminator, Kitty, MobaXterm, LXTer…
Out of time: automated lip sync in the wild
A launch point for your personal nvim configuration
SAiD: Blendshape-based Audio-Driven Speech Animation with Diffusion
Code for the paper "End-to-end Learning for 3D Facial Animation from Speech"
DiffuseStyleGesture: Stylized Audio-Driven Co-Speech Gesture Generation with Diffusion Models (IJCAI 2023) | The DiffuseStyleGesture+ entry to the GENEA Challenge 2023 (ICMI 2023, Reproducibility A…
[ICCV-2023] The official repo for the paper "LivelySpeaker: Towards Semantic-aware Co-Speech Gesture Generation".
💬 An extensive collection of exceptional resources dedicated to the captivating world of talking face synthesis! ⭐ If you find this repo useful, please give it a star! 🤩
This is a medical bot built using Llama2 and Sentence Transformers. The bot is powered by Langchain and Chainlit. The bot runs on a decent CPU machine with a minimum of 16GB of RAM.
This is a RAG implementation using Open Source stack. BioMistral 7B has been used to build this app along with PubMedBert as an embedding model, Qdrant as a self hosted Vector DB, and Langchain & L…
PantoMatrix: Generating Face and Body Animation from Speech
This is the official repository for TalkSHOW: Generating Holistic 3D Human Motion from Speech [CVPR2023].
Expressive Body Capture: 3D Hands, Face, and Body from a Single Image
This is the codebase for SHOW in Generating Holistic 3D Human Motion from Speech [CVPR2023],
[CVPR 2024 Highlight] Feature 3DGS: Supercharging 3D Gaussian Splatting to Enable Distilled Feature Fields