-
Peking University
- Beijing, China
- https://www.microsoft.com/en-us/research/people/wangliang/
Highlights
Stars
A bibliography and survey of the papers surrounding o1
Educational framework exploring ergonomic, lightweight multi-agent orchestration. Managed by OpenAI Solution team.
Homepage for ProLong (Princeton long-context language models) and paper "How to Train Long-Context Language Models (Effectively)"
A collection of LLM papers, blogs, and projects, with a focus on OpenAI o1 🍓 and reasoning techniques.
Finetune Llama 3.3, Mistral, Phi, Qwen 2.5 & Gemma LLMs 2-5x faster with 70% less memory
Memory optimization and training recipes to extrapolate language models' context length to 1 million tokens, with minimal hardware.
Large World Model -- Modeling Text and Video with Millions Context
LOFT: A 1 Million+ Token Long-Context Benchmark
This repo contains the source code for RULER: What’s the Real Context Size of Your Long-Context Language Models?
Repository for Meta Chameleon, a mixed-modal early-fusion foundation model from FAIR.
A programming framework for agentic AI 🤖 PyPi: autogen-agentchat Discord: https://aka.ms/autogen-discord Office Hour: https://aka.ms/autogen-officehour
⚡FlashRAG: A Python Toolkit for Efficient RAG Research
[MLSys 2024 Best Paper Award] AWQ: Activation-aware Weight Quantization for LLM Compression and Acceleration
Generative Representational Instruction Tuning
Perplexica is an AI-powered search engine. It is an Open source alternative to Perplexity AI
LongEmbed: Extending Embedding Models for Long Context Retrieval (EMNLP 2024)
A high-throughput and memory-efficient inference and serving engine for LLMs
Robust recipes to align language models with human and AI preferences
Awesome-LLM-RAG: a curated list of advanced retrieval augmented generation (RAG) in Large Language Models
The official PyTorch implementation of Google's Gemma models
📰 Must-read papers and blogs on LLM based Long Context Modeling 🔥
Doing simple retrieval from LLM models at various context lengths to measure accuracy
SmartPlay is a benchmark for Large Language Models (LLMs). Uses a variety of games to test various important LLM capabilities as agents. SmartPlay is designed to be easy to use, and to support futu…
ICLR2024 Spotlight: curation/training code, metadata, distribution and pre-trained models for MetaCLIP; CVPR 2024: MoDE: CLIP Data Experts via Clustering