-
Macau University
- Taipa University Road, Macau, China
-
02:58
(UTC -12:00)
Highlights
- Pro
Stars
[CVPR 2024] Adaptive Multi-Modal Cross-Entropy Loss for Stereo Matching
Corruption and Perturbation Robustness (ICLR 2019)
Code for NeurIPS 2024 "Dual-frame Fluid Motion Estimation with Test-time Optimization and Zero-divergence Loss"
Code release for the ECCV 2024 paper 'Fully Test-Time Adaptation for Monocular 3D Object Detection'
[CVPR 2024] Bridging the Gap: A Unified Video Comprehension Framework for Moment Retrieval and Highlight Detection
[CVPR 2024 Award Candidate] Producing and Leveraging Online Map Uncertainty in Trajectory Prediction
[TPAMI 2024 & CVPR 2022] Attention Concatenation Volume for Accurate and Efficient Stereo Matching
[CVPR 2024] The official repo for FlashAvatar
Code for "Detector-Free Structure from Motion", CVPR 2024
[CVPR 2024] Make-Your-Anchor: A Diffusion-based 2D Avatar Generation Framework.
[CVPR 2024] FRESCO: Spatial-Temporal Correspondence for Zero-Shot Video Translation
[CVPR 2024 Oral] Official repository of FMA-Net
[ICCV 2023 Oral] Text-to-Image Diffusion Models are Zero-Shot Video Generators
Using Low-rank adaptation to quickly fine-tune diffusion models.
Official repo for VGen: a holistic video generation ecosystem for video generation building on diffusion models
VideoCrafter2: Overcoming Data Limitations for High-Quality Video Diffusion Models
[SIGGRAPH Asia 2024 (Journal Track)]StyleCrafter: Enhancing Stylized Text-to-Video Generation with Style Adapter
[ECCV 2024 Oral] MotionDirector: Motion Customization of Text-to-Video Diffusion Models.
Implementation of Video Diffusion Models, Jonathan Ho's new paper extending DDPMs to Video Generation - in Pytorch
[CSUR] A Survey on Video Diffusion Models
[TPAMI 2024 & CVPR 2023] PyTorch code for DGM4: Detecting and Grounding Multi-Modal Media Manipulation and beyond
Source code for paper "A Spark of Vision-Language Intelligence: 2-Dimensional Autoregressive Transformer for Efficient Finegrained Image Generation"
[Paper][AAAI 2025] MyGO: Discrete Modality Information as Fine-Grained Tokens for Multi-modal Knowledge Graph Completion
HART: Efficient Visual Generation with Hybrid Autoregressive Transformer
A comprehensive collection of IQA papers
I2-Multimedia-Lab / SaTQA
Forked from Srache/SaTQA[AAAI 2024] Transformer-Based No-Reference Image Quality Assessment via Supervised Contrastive Learning
Official repository for our paper titled "Learning Generalizable Perceptual Representations for Data-Efficient No-Reference Image Quality Assessment".