- Republic of Korea, Seoul
-
00:00
(UTC +09:00)
Stars
Official repository of ’Visual-RFT: Visual Reinforcement Fine-Tuning’
This repository provides valuable reference for researchers in the field of multimodality, please start your exploratory travel in RL-based Reasoning MLLMs!
[TMLR 2025🔥] A survey for the autoregressive models in vision.
MedRAX: Medical Reasoning Agent for Chest X-ray
🤗 smolagents: a barebones library for agents. Agents write python code to call tools and orchestrate other agents.
MedXpertQA: Benchmarking Expert-Level Medical Reasoning and Understanding
A Python package for evaluating radiology report generation using multiple standard and medical-specific metrics.
[MICCAI 2024] RadiomicsFill-Mammo: Synthetic Mammogram Mass Manipulation with Radiomics Features
[ECCV 2024] Be-Your-Outpainter https://arxiv.org/abs/2403.13745
A course on aligning smol models.
MMedPO: Aligning Medical Vision-Language Models with Clinical-Aware Multimodal Preference Optimization
NeurIPS'24 DB (Spotlight) | Instruction Tuning Large Language Models to Understand Electronic Health Records
Project Page for "LISA: Reasoning Segmentation via Large Language Model"
[MM2024, oral] "Self-Supervised Visual Preference Alignment" https://arxiv.org/abs/2404.10501
Paper Reproduction Google SCoRE(Training Language Models to Self-Correct via Reinforcement Learning)
"Repository for implementing the Back-in-Time Diffusion method for detecting medical deepfakes in CT and MRI scans, including training and evaluation tools."
Smooth Diffusion: Crafting Smooth Latent Spaces in Diffusion Models arXiv 2023 / CVPR 2024
A collection of LLM papers, blogs, and projects, with a focus on OpenAI o1 🍓 and reasoning techniques.
g1: Using Llama-3.1 70b on Groq to create o1-like reasoning chains
Beyond Hallucinations: Enhancing LVLMs through Hallucination-Aware Direct Preference Optimization
[Nature Reviews Bioengineering🔥] Application of Large Language Models in Medicine. A curated list of practical guide resources of Medical LLMs (Medical LLMs Tree, Tables, and Papers)
Train transformer language models with reinforcement learning.
This repository contains the code for SFT, RLHF, and DPO, designed for vision-based LLMs, including the LLaVA models and the LLaMA-3.2-vision models.