-
University of Michigan
- San José, California, USA
-
16:51
(UTC -08:00) - http://ziqiaoma.com/
- @ziqiao_ma
- https://scholar.google.com/citations?user=WbybssYAAAAJ&hl=en
Highlights
- Pro
Stars
Code for Paper: Training Software Engineering Agents and Verifiers with SWE-Gym
Animate124: Animating One Image to 4D Dynamic Scene
Search docs.voxel51.com with an LLM!
PyTorch codes for "LST: Ladder Side-Tuning for Parameter and Memory Efficient Transfer Learning"
[NeurIPS 2023] Self-Chained Image-Language Model for Video Localization and Question Answering
Official Code for IJCAI 2023 Paper: Towards Collaborative Plan Acquisition through Theory of Mind Modeling in Situated Dialogue
Repo for the ACL 2023 Paper "NLP Reproducibility For All: Understanding Experiences of Beginners."
Transformer related optimization, including BERT, GPT
Tool Learning for Big Models, Open-Source Solutions of ChatGPT-Plugins
Official implementation of the paper "Uncovering the Disentanglement Capability in Text-to-Image Diffusion Models
A curated list of practical guide resources of LLMs (LLMs Tree, Examples, Papers)
The repository provides code for running inference with the SegmentAnything Model (SAM), links for downloading the trained model checkpoints, and example notebooks that show how to use the model.
A comprehensive list of papers using large language/multi-modal models for Robotics/RL, including papers, codes, and related websites
Code for EMNLP 2022 Paper DANLI: Deliberative Agent for Following Natural Language Instructions
A curated list of Decision Transformer resources (continually updated)
Code release for "Masked-attention Mask Transformer for Universal Image Segmentation"
Diagnostic tests for linguistic capacities in language models
Episodic Transformer (E.T.) is a novel attention-based architecture for vision-and-language navigation. E.T. is based on a multimodal transformer that encodes language inputs and the full episode h…
Repository containing code for "How to Train BERT with an Academic Budget" paper
Supervision Exists Everywhere: A Data Efficient Contrastive Language-Image Pre-training Paradigm
METER: A Multimodal End-to-end TransformER Framework