-
Xidian University
- Xi'an, China
-
09:42
(UTC +08:00)
Lists (4)
Sort Name ascending (A-Z)
Stars
[IEEE TCSVT'24] Study of Subjective and Objective Naturalness Assessment of AI-Generated Images
🦁 Lion, new optimizer discovered by Google Brain using genetic algorithms that is purportedly better than Adam(w), in Pytorch
解决Cursor在免费订阅期间出现以下提示的问题: Too many free trial accounts used on this machine. Please upgrade to pro. We have this limit in place to prevent abuse. Please let us know if you believe this is a mistake.
Get up and running with Llama 3.3, Mistral, Gemma 2, and other large language models.
MiniCPM-V 2.6: A GPT-4V Level MLLM for Single Image, Multi Image and Video on Your Phone
This list of writing prompts covers a range of topics and tasks, including brainstorming research ideas, improving language and style, conducting literature reviews, and developing research plans.
Code for Learning Subject-Aware Cropping by Outpainting Professional Photos
Official PyTorch implementation of the “Spatial-Semantic Collaborative Cropping for User Generated Content”. (AAAI24)
Edit anything in images powered by segment-anything, ControlNet, StableDiffusion, etc. (ACM MM)
⏰ Collaboratively track deadlines of conferences recommended by CCF (Website, Python Cli, Wechat Applet) / If you find it useful, please star this project, thanks~
[TPAMI] Multi-modality Multi-attribute Contrastive Pre-training for Image Aesthetics Computing
Semantics-Aware Image Aesthetics Assessment using Tag Matching and Contrastive Ranking (ACM MM2024)
Multi-Level Transitional Contrast Learning for Personalized Image Aesthetics Assessment (TMM2023)
👁️ 🖼️ 🔥PyTorch Toolbox for Image Quality Assessment, including PSNR, SSIM, LPIPS, FID, NIQE, NRQM(Ma), MUSIQ, TOPIQ, NIMA, DBCNN, BRISQUE, PI and more...
【CVPR 2024 Highlight】Monkey (LMM): Image Resolution and Text Label Are Important Things for Large Multi-modal Models
[LMM + AIGC] What do we expect from LMMs as AIGI evaluators and how do they perform?
③[ICML2024] [IQA, IAA, VQA] All-in-one Foundation Model for visual scoring. Can efficiently fine-tune to downstream datasets.
②[CVPR 2024] Low-level visual instruction tuning, with a 200K dataset and a model zoo for fine-tuned checkpoints.
①[ICLR2024 Spotlight] (GPT-4V/Gemini-Pro/Qwen-VL-Plus+16 OS MLLMs) A benchmark for multi-modality LLMs (MLLMs) on low-level vision and visual quality assessment.
An expert benchmark aiming to comprehensively evaluate the aesthetic perception capacities of MLLMs.
[ACMMM 2024] AesExpert: Towards Multi-modality Foundation Model for Image Aesthetics Perception
[TCSVT 2024] Official code release of our paper "Towards Explainable Image Aesthetics Assessment With Attribute-Oriented Critiques Generation"