Skip to content

Contrastive Language-Image pretraining based retrieval for the deep fashion dataset for image-to-text and text-to-image search for fashion.

Notifications You must be signed in to change notification settings

sujitnoronha/CLIP_deepfashion

Repository files navigation

CLIP for Deep Fashion

CLIP

This repository contains an implementation of CLIP-based text-to-image and image-to-text retrieval using the Deep Fashion dataset. The goal of this project is to gain a better understanding of how CLIP works by training a model to embed images and their corresponding descriptions in a shared semantic space. By doing so, the model can learn to bring similar images and descriptions closer together while pushing dissimilar ones further apart. This approach allows for efficient retrieval of images based on textual queries or vice versa. As part of this project, we plan to continue exploring different datasets and fine-tuning our model to achieve even better performance. We welcome contributions from the community and look forward to collaborating with others interested in this exciting area of research.

Using tensorflow similarity

References: https://openai.com/research/clip https://github.com/tensorflow/similarity/blob/master/examples/multimodal_example.ipynb

deep fashion dataset: https://github.com/yumingj/DeepFashion-MultiModal

About

Contrastive Language-Image pretraining based retrieval for the deep fashion dataset for image-to-text and text-to-image search for fashion.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published