Stars
Implementation of Vision Transformer, a simple way to achieve SOTA in vision classification with only a single transformer encoder, in Pytorch
Open source implementation of "Vision Transformers Need Registers"
MiniCPM-V 2.6: A GPT-4V Level MLLM for Single Image, Multi Image and Video on Your Phone
Chatbot Arena meets multi-modality! Multi-Modality Arena allows you to benchmark vision-language models side-by-side while providing images as inputs. Supports MiniGPT-4, LLaMA-Adapter V2, LLaVA, B…
[NeurIPS 2024] Knowledge Circuits in Pretrained Transformers
Repository for Label Words are Anchors: An Information Flow Perspective for Understanding In-Context Learning
Official Pytorch implementation of "Interpreting and Editing Vision-Language Representations to Mitigate Hallucinations"
Locating and editing factual associations in GPT (NeurIPS 2022)
Collection of Reverse Engineering in Large Model
[NLPCC 2022] Kformer: Knowledge Injection in Transformer Feed-Forward Layers
code for EMNLP 2024 paper: Neuron-Level Knowledge Attribution in Large Language Models
The SVO-Probes Dataset for Verb Understanding
Evaluating Vision & Language Pretraining Models with Objects, Attributes and Relations. [EMNLP 2022]
A curated list of awesome advice for computer science Ph.D. applicants.
Official Repository of Multi-Object Hallucination in Vision-Language Models (NeurIPS 2024)
[ACL'24] Beyond One-Preference-Fits-All Alignment: Multi-Objective Direct Preference Optimization
High Accuracy and efficiency multi-task fine-tuning framework for Code LLMs. This work has been accepted by KDD 2024.
A Wiki for students who want to transfer to another university
OLMoE: Open Mixture-of-Experts Language Models
This repository contains demos I made with the Transformers library by HuggingFace.
[TMLR] Public code repo for paper "A Single Transformer for Scalable Vision-Language Modeling"
Coding a Multimodal (Vision) Language Model from scratch in PyTorch with full explanation: https://www.youtube.com/watch?v=vAmKB7iPkWw
Multi-Objective Reinforcement Learning algorithms implementations.
Efficient implementations of state-of-the-art linear attention models in Pytorch and Triton
The repository provides code for running inference with the Meta Segment Anything Model 2 (SAM 2), links for downloading the trained model checkpoints, and example notebooks that show how to use th…