Skip to content

[CVPR 2024] Alpha-CLIP: A CLIP Model Focusing on Wherever You Want

License

Notifications You must be signed in to change notification settings

qingpowuwu/AlphaCLIP

ย 
ย 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 

Repository files navigation

Alpha-CLIP

This repository is the official implementation of AlphaCLIP

Alpha-CLIP: A CLIP Model Focusing on Wherever You Want
Zeyi Sun*, Ye Fang*, Tong Wu, Pan Zhang, Yuhang Zang, Shu Kong, Yuanjun Xiong, Dahua Lin, Jiaqi Wang

*Equal Contribution

Demo Alpha-CLIP with Stable Diffusion: Hugging Face Spaces Open in OpenXLab

Demo Alpha-CLIP with LLaVA: Hugging Face Spaces Open in OpenXLab

๐Ÿ“œ News

๐Ÿš€ [2024/7/19] We have launched training code as well as data MaskImageNet!

๐Ÿš€ [2024/3/4] CLIP-L/14@336px finetuned on GRIT-20M is available, checkout model-zoo!

๐Ÿš€ [2024/2/27] Our paper Alpha-CLIP is accepted by CVPR'24!

๐Ÿš€ [2024/1/2] Zero-shot testing code for Imagenet-S Classification and Referring Expression Comprehension are released!

๐Ÿš€ [2023/12/27] Web demo and local demo of Alpha-CLIP with LLaVA are released!

๐Ÿš€ [2023/12/7] Web demo and local demo of Alpha-CLIP with Stable Diffusion are released!

๐Ÿš€ [2023/12/7] The paper and project page are released!

๐Ÿ’ก Highlights

  • ๐Ÿ”ฅ 3.93% improved zero-shot ImageNet classification accuracy when providing foreground alpha-map.
  • ๐Ÿ”ฅ Plug-in and play with region focus in any work that use CLIP vision encoder.
  • ๐Ÿ”ฅ A strong visual encoder as versatile tool when foreground mask is available.

๐Ÿ‘จโ€๐Ÿ’ป Todo

  • Training code for Alpha-CLIP and MaskImageNet data.
  • Evaluation code for Alpha-CLIP
  • Zero-shot evaluation for Imagenet-S Classification and REC tasks.
  • Web demo and local demo of Alpha-CLIP with LLaVA
  • Web demo and local demo of Alpha-CLIP with Stable Diffusion
  • Usage example notebook of Alpha-CLIP
  • Checkpoints of Alpha-CLIP

๐Ÿ› ๏ธ Usage

ไธ‹่ฝฝ MaskImageNet ๆ•ฐๆฎ้›†

cd AlphaCLIP
mkdir MaskImageNet
wget https://huggingface.co/datasets/Zery/MaskImageNet/resolve/main/M_ImageNet_top_460k.json?download=true
wget https://huggingface.co/datasets/Zery/MaskImageNet/resolve/main/imagenet_919.json?download=true

Installation

our model is based on CLIP, please first prepare environment for CLIP, then directly install Alpha-CLIP.

pip install -e .

install loralib

pip install loralib

how to use

Download model from model-zoo and place it under checkpoints.

import alpha_clip
clip_model, preprocessor = alpha_clip.load("ViT-B/16", alpha_vision_ckpt_pth="pretrained_models/clip_b16_grit1m_fultune_8xe.pth", device="cpu"), 
image_features = clip_model.visual(image, alpha)

alpha need to be normalized via transforms when using binary_mask in (0, 1)

mask_transform = transforms.Compose([
    transforms.ToTensor(), 
    transforms.Resize((224, 224)),
    transforms.Normalize(0.5, 0.26)
])
alpha = mask_transform(binary_mask * 255)

Training

Please refer to here

Zero-shot Prediction

import torch
import alpha_clip
from PIL import Image
import numpy as np
from torchvision import transforms

# load model and prepare mask transform
device = "cuda" if torch.cuda.is_available() else "cpu"
model, preprocess = alpha_clip.load("ViT-L/14", alpha_vision_ckpt_pth="./checkpoints/clip_l14_grit20m_fultune_2xe.pth", device=device)  # change to your own ckpt path
mask_transform = transforms.Compose([
    transforms.ToTensor(), 
    transforms.Resize((224, 224)), # change to (336,336) when using ViT-L/14@336px
    transforms.Normalize(0.5, 0.26)
])

# prepare image and mask
img_pth = './examples/image.png'
mask_pth = './examples/dress_mask.png' # image-type mask

image = Image.open(img_pth).convert('RGB')
mask = np.array(Image.open(mask_pth)) 
# get `binary_mask` array (2-dimensional bool matrix)
if len(mask.shape) == 2: binary_mask = (mask == 255)
if len(mask.shape) == 3: binary_mask = (mask[:, :, 0] == 255)

alpha = mask_transform((binary_mask * 255).astype(np.uint8))
alpha = alpha.half().cuda().unsqueeze(dim=0)

# calculate image and text features
image = preprocess(image).unsqueeze(0).half().to(device)
text = alpha_clip.tokenize(["a goegously dressed woman", "a purple sleeveness dress", "bouquet of pink flowers"]).to(device)

with torch.no_grad():
    image_features = model.visual(image, alpha)
    text_features = model.encode_text(text)

# normalize
image_features = image_features / image_features.norm(dim=-1, keepdim=True)
text_features = text_features / text_features.norm(dim=-1, keepdim=True)

## print the result
similarity = (100.0 * image_features @ text_features.T).softmax(dim=-1)
print("Label probs:", similarity.cpu().numpy()) # prints: [[9.388e-05 9.995e-01 2.415e-04]]

Note: Using .half() for tensor or .float() for model to maintain type consistency.

More usage examples are available๏ผš

  • Visualization of attention map: notebook
  • Alpha-CLIP used in BLIP-Diffusion: notebook
  • Alpha-CLIP used in SD_ImageVar: demo
  • Alpha-CLIP used in LLaVA-1.5: code demo
  • Alpha-CLIP evaluation code for Image Recognition: code

โญ Demos

โค๏ธ Acknowledgments

  • CLIP: The codebase we built upon. Thanks for their wonderful work.
  • LAVIS: The amazing open-sourced multimodality learning codebase, where we test Alpha-CLIP in BLIP-2 and BLIP-Diffusion.
  • Point-E: Wonderful point-cloud generation model, where we test Alpha-CLIP for 3D generation task.
  • LLaVA: Wounderful MLLM that use CLIP as visual bacbone where we test the effectiveness of Alpha-CLIP.

โœ’๏ธ Citation

If you find our work helpful for your research, please consider giving a star โญ and citation ๐Ÿ“

@misc{sun2023alphaclip,
      title={Alpha-CLIP: A CLIP Model Focusing on Wherever You Want}, 
      author={Zeyi Sun and Ye Fang and Tong Wu and Pan Zhang and Yuhang Zang and Shu Kong and Yuanjun Xiong and Dahua Lin and Jiaqi Wang},
      year={2023},
      eprint={2312.03818},
      archivePrefix={arXiv},
      primaryClass={cs.CV}
}

License

Code License Data License Usage and License Notices: The data and checkpoint is intended and licensed for research use only. They are also restricted to uses that follow the license agreement of CLIP. The dataset is CC BY NC 4.0 (allowing only non-commercial use) and models trained using the dataset should not be used outside of research purposes.

About

[CVPR 2024] Alpha-CLIP: A CLIP Model Focusing on Wherever You Want

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Jupyter Notebook 91.5%
  • Python 8.3%
  • Shell 0.2%