Stars
Official implementation of "HowToCaption: Prompting LLMs to Transform Video Annotations at Scale." ECCV 2024
Awesome list for research on CLIP (Contrastive Language-Image Pre-Training).
ICCV 2023 Papers: Discover cutting-edge research from ICCV 2023, the leading computer vision conference. Stay updated on the latest in computer vision and deep learning, with code included. ⭐ suppo…
PyTorch code and models for the DINOv2 self-supervised learning method.
CLIP (Contrastive Language-Image Pretraining), Predict the most relevant text snippet given an image
An open source implementation of CLIP.
Chinese version of CLIP which achieves Chinese cross-modal retrieval and representation generation.
[CVPR2023] All in One: Exploring Unified Video-Language Pre-training
Implementation of DALL-E 2, OpenAI's updated text-to-image synthesis neural network, in Pytorch
Code release for SLIP Self-supervision meets Language-Image Pre-training
LAVIS - A One-stop Library for Language-Vision Intelligence
PyTorch code for BLIP: Bootstrapping Language-Image Pre-training for Unified Vision-Language Understanding and Generation
Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalities
Homepage for STAT 157 at UC Berkeley
Interactive code for image similarity using SIFT algorithm
Experiments with supervised contrastive learning methods with different loss functions
Implementing Siamese networks with a contrastive loss for similarity learning
PyTorch implementation of MoCo: https://arxiv.org/abs/1911.05722
CrossCLR: Cross-modal Contrastive Learning For Multi-modal Video Representations, ICCV 2021
Video Contrastive Learning with Global Context, ICCVW 2021
Models and examples built with TensorFlow