Skip to content

Code for ICLR 2024 paper: Measuring Vision-Language STEM Skills of Neural Models

License

Notifications You must be signed in to change notification settings

stemdataset/STEM

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

2 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Measuring Vision-Language STEM Skills of Neural Models

The code for ICLR 2024 paper: Measuring Vision-Language STEM Skills of Neural Models.

📃 [Paper] • 💻 [Github] • 🤗 [Dataset] • 🏆 [Leaderboard] • 📽 [Slides] • 📋 [Poster]

Setup Environment

We recommend using Anaconda to create a new environment and install the required packages. You can create a new environment and install the required packages using the following commands:

conda create -n clip python=3.10
conda activate clip
conda install pytorch==1.12.1 torchvision==0.13.1 torchaudio==0.12.1 -c pytorch
pip install git+https://github.com/openai/CLIP.git
pip install transformers==4.18.0
pip install datasets

Run the Code

You can run the inference code using the following command:

bash run_eval_clip.sh ${eval_split}

where ${eval_split} is the evaluation split you want to evaluate on. The evaluation splits are valid, test. The results will be saved in results/clip_${model}_${eval_split}/. You can submit the preds.txt to the leaderboard for the test split evaluation.

Citation

@inproceedings{shen2024measuring,
  title={Measuring Vision-Language STEM Skills of Neural Models},
  author={Shen, Jianhao and Yuan, Ye and Mirzoyan, Srbuhi and Zhang, Ming and Wang, Chenguang},
  booktitle={ICLR},
  year={2024}
}

About

Code for ICLR 2024 paper: Measuring Vision-Language STEM Skills of Neural Models

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published