-
Huazhong University of Science and Technology
- Wuhan, China
Highlights
- Pro
Stars
Code for our paper in submission "Towards Assessing the Synthetic-to-Measured Adversarial Vulnerability of SAR ATR".
Exploring Visual Prompts for Adapting Large-Scale Models
One Prompt Word is Enough to Boost Adversarial Robustness for Pre-trained Vision-Language Models
Chinese version of CLIP which achieves Chinese cross-modal retrieval and representation generation.
Code repository for CVPR2024 paper 《Pre-trained Model Guided Fine-Tuning for Zero-Shot Adversarial Robustness》
A Survey on multimodal learning research.
TransferAttack is a pytorch framework to boost the adversarial transferability for image classification.
DynamicPose, a simple and robust framework for animating human images.
[NeurIPS 2022] Make Sharpness-Aware Minimization Stronger: A Sparsified Perturbation Approach -- Official Implementation
Advanced AI Explainability for computer vision. Support for CNNs, Vision Transformers, Classification, Object detection, Segmentation, Image similarity and more.
[ICCV 2021- Oral] Official PyTorch implementation for Generic Attention-model Explainability for Interpreting Bi-Modal and Encoder-Decoder Transformers, a novel method to visualize any Transformer-…
Code for the paper "Speckle-Variant Attack: Toward Transferable Adversarial Attack to SAR Target Recognition".
High-Resolution Image Synthesis with Latent Diffusion Models
official PyTorch implement of Towards Adversarial Attack on Vision-Language Pre-training Models
Chinese and English multimodal conversational language model | 多模态中英双语对话语言模型
[TLLM'23] PandaGPT: One Model To Instruction-Follow Them All
Open-sourced codes for MiniGPT-4 and MiniGPT-v2 (https://minigpt-4.github.io, https://minigpt-v2.github.io/)
🦦 Otter, a multi-modal model based on OpenFlamingo (open-sourced version of DeepMind's Flamingo), trained on MIMIC-IT and showcasing improved instruction-following and in-context learning ability.
An Open-Source Package for Textual Adversarial Attack.
Set-level Guidance Attack: Boosting Adversarial Transferability of Vision-Language Pre-training Models. [ICCV 2023 Oral]
Code for ALBEF: a new vision-language pre-training method
Image-to-Image Translation in PyTorch