-
LoRA-Pro-fix Public
Forked from mrflogs/LoRA-ProOfficial code for our paper, "LoRA-Pro: Are Low-Rank Adapters Properly Optimized? "
-
Twin-Merging Public
[NeurIPS2024] Twin-Merging: Dynamic Integration of Modular Expertise in Model Merging
-
-
CustomLLMFinetuningHandbook Public
an example to fine-tuning a Language Learning Model (LLM) from data preparation to deployment.
-
-
axolotl Public
Forked from axolotl-ai-cloud/axolotlaxolotl customize
-
Awesome-Model-Fusion Public
This GitHub repository summarizes most recent papers and resources related to the model fusion.
7 UpdatedAug 10, 2024 -
-
lm-evaluation-harness-fast Public
Forked from EleutherAI/lm-evaluation-harnessspeedup for lm-evaluation-harness; support tensor-parallel inference and data-parallel inference; support gptq, bitsandbytes, peft and exllamav2.
-
HELM-Extended-Local Public
Forked from stanford-crfm/helmsupport for various name in http/local run; support for gptq/bnb/tensorparallel in local run
-
vllm Public
Forked from vllm-project/vllmA high-throughput and memory-efficient inference and serving engine for LLMs
-
dpo Public
Forked from huggingface/alignment-handbookRobust recipes for to align language models with human and AI preferences
-
-
trl Public
Forked from huggingface/trlTrain transformer language models with reinforcement learning.
-
peft Public
Forked from huggingface/peft🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.
-
MIRACLE Public
[EMNLP2023]: MIRACLE: Towards Personalized Dialogue Generation with Latent-Space Multiple Personal Attribute Control
-
-
-