Skip to content

Commit

Permalink
new llama-2 default settings (axolotl-ai-cloud#370)
Browse files Browse the repository at this point in the history
* new default settings

* fix whitespace

* rm max packed sequence length

---------

Co-authored-by: Mads Henrichsen <[email protected]>
  • Loading branch information
mhenrichsen and Mads Henrichsen authored Aug 14, 2023
1 parent 919246f commit fdffef5
Show file tree
Hide file tree
Showing 2 changed files with 7 additions and 8 deletions.
7 changes: 3 additions & 4 deletions examples/llama-2/lora.yml
Original file line number Diff line number Diff line change
Expand Up @@ -15,7 +15,7 @@ val_set_size: 0.01
output_dir: ./lora-out

sequence_len: 4096
max_packed_sequence_len: 4096
sample_packing: true

adapter: lora
lora_model_dir:
Expand Down Expand Up @@ -49,8 +49,8 @@ early_stopping_patience:
resume_from_checkpoint:
local_rank:
logging_steps: 1
xformers_attention: true
flash_attention:
xformers_attention:
flash_attention: true

warmup_steps: 10
eval_steps: 20
Expand All @@ -64,4 +64,3 @@ special_tokens:
bos_token: "<s>"
eos_token: "</s>"
unk_token: "<unk>"
pad_token: "<pad>"
8 changes: 4 additions & 4 deletions examples/llama-2/qlora.yml
Original file line number Diff line number Diff line change
Expand Up @@ -18,7 +18,8 @@ adapter: qlora
lora_model_dir:

sequence_len: 4096
max_packed_sequence_len: 4096
sample_packing: true

lora_r: 32
lora_alpha: 16
lora_dropout: 0.05
Expand Down Expand Up @@ -50,8 +51,8 @@ early_stopping_patience:
resume_from_checkpoint:
local_rank:
logging_steps: 1
xformers_attention: true
flash_attention:
xformers_attention:
flash_attention: true

warmup_steps: 10
eval_steps: 20
Expand All @@ -65,4 +66,3 @@ special_tokens:
bos_token: "<s>"
eos_token: "</s>"
unk_token: "<unk>"
pad_token: "<pad>"

0 comments on commit fdffef5

Please sign in to comment.