Skip to content

Commit

Permalink
Fix lora loading haotian-liu#1075
Browse files Browse the repository at this point in the history
  • Loading branch information
haotian-liu committed Feb 6, 2024
1 parent 159c8b0 commit 04fb03d
Showing 1 changed file with 2 additions and 1 deletion.
3 changes: 2 additions & 1 deletion llava/model/builder.py
Original file line number Diff line number Diff line change
Expand Up @@ -50,7 +50,8 @@ def load_pretrained_model(model_path, model_base, model_name, load_8bit=False, l
if 'lora' in model_name.lower() and model_base is None:
warnings.warn('There is `lora` in model name but no `model_base` is provided. If you are loading a LoRA model, please provide the `model_base` argument. Detailed instruction: https://github.com/haotian-liu/LLaVA#launch-a-model-worker-lora-weights-unmerged.')
if 'lora' in model_name.lower() and model_base is not None:
lora_cfg_pretrained = AutoConfig.from_pretrained(model_path)
from llava.model.language_model.llava_llama import LlavaConfig
lora_cfg_pretrained = LlavaConfig.from_pretrained(model_path)
tokenizer = AutoTokenizer.from_pretrained(model_base, use_fast=False)
print('Loading LLaVA from base model...')
model = LlavaLlamaForCausalLM.from_pretrained(model_base, low_cpu_mem_usage=True, config=lora_cfg_pretrained, **kwargs)
Expand Down

0 comments on commit 04fb03d

Please sign in to comment.