Stars
Unified Efficient Fine-Tuning of 100+ LLMs (ACL 2024)
Official pytorch implementation of Paper - 🍝 PASTA: Pathology-Aware MRI to PET Cross-modal Translation with Diffusion Models - MICCAI 2024
A high-throughput and memory-efficient inference and serving engine for LLMs
High-Resolution Image Synthesis with Latent Diffusion Models
LAVIS - A One-stop Library for Language-Vision Intelligence
Official repository of the MIRAGE benchmark
🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.
Medical Image Segmentation with Diffusion Model
🧑🚀 全世界最好的LLM资料总结 | Summary of the world's best LLM resources.
CS 自学指南(Java编程语言、数据库、数据结构与算法、计算机组成原理、操作系统、计算机网络、英语、简历、面试)
detrex is a research platform for DETR-based object detection, segmentation, pose estimation and other visual recognition tasks.
[CVPR 2024 Oral] MemSAM: Taming Segment Anything Model for Echocardiography Video Segmentation.
[NeurIPS'23 Oral] Visual Instruction Tuning (LLaVA) built towards GPT-4V level capabilities and beyond.
[CVPR 2024 Oral] InternVL Family: A Pioneering Open-Source Alternative to GPT-4o. 接近GPT-4o表现的开源多模态对话模型
Official Code for our CVPR 2024 Paper "Diversified and Personalized Multi-rater Medical Image Segmentation" (Highlight)
Implementations of recent research prototypes/demonstrations using MONAI.
Emote Portrait Alive: Generating Expressive Portrait Videos with Audio2Video Diffusion Model under Weak Conditions
DSMIL: Dual-stream multiple instance learning networks for tumor detection in Whole Slide Image
✨✨Latest Advances on Multimodal Large Language Models
🚀 跃问YueWen 多模态大模型逆向API【特长:超强多模态】,支持高速流式输出、联网搜索、长文档解读、图像解析、多轮对话,零配置部署,多路token支持,自动清理会话痕迹,仅供测试,如需商用请前往官方开放平台。
The official repo of Qwen-VL (通义千问-VL) chat & pretrained large vision language model proposed by Alibaba Cloud.
[CVPR 2024] VoCo: A Simple-yet-Effective Volume Contrastive Learning Framework for 3D Medical Image Analysis
PPL Quantization Tool (PPQ) is a powerful offline neural network quantization tool.
Implementation of the LLaMA language model based on nanoGPT. Supports flash attention, Int8 and GPTQ 4bit quantization, LoRA and LLaMA-Adapter fine-tuning, pre-training. Apache 2.0-licensed.