Skip to content

Commit

Permalink
Release v1.5 training scripts and dataset.
Browse files Browse the repository at this point in the history
  • Loading branch information
haotian-liu committed Oct 11, 2023
1 parent ecb5984 commit f290dff
Showing 1 changed file with 1 addition and 1 deletion.
2 changes: 1 addition & 1 deletion README.md
Original file line number Diff line number Diff line change
Expand Up @@ -233,7 +233,7 @@ Training script with DeepSpeed ZeRO-3: [`finetune.sh`](https://github.com/haotia
New options to note:

- `--mm_projector_type mlp2x_gelu`: the two-layer MLP vision-language connector.
- `--image_aspect_ratio pad`: it slightly reduces hallucination.
- `--image_aspect_ratio pad`: this pads the non-square images to square, instead of cropping them; it slightly reduces hallucination.
- `--group_by_modality_length True`: this should only be used when your instruction tuning dataset contains both language (e.g. ShareGPT) and multimodal (e.g. LLaVA-Instruct). It makes the training sampler only sample a single modality (either image or language) during training, which we observe to speed up training by ~25%, and does not affect the final outcome.

## Evaluation
Expand Down

0 comments on commit f290dff

Please sign in to comment.