Download Mistral weights
Mistral 7B is Apache 2.0 licensed and can be used without restrictions. It:
- Outperforms Llama 2 13B on all benchmarks
- Outperforms Llama 1 34B on many benchmarks
- Approaches CodeLlama 7B performance on code, while remaining good at English tasks
- Uses Grouped-query attention (GQA) for faster inference
Uses Sliding Window Attention (SWA) to handle longer sequences at smaller cost. This project's implementation does not use Sliding Window Attention, so the context length is limited to 4096 tokens.
Details about the data used to train the model or training procedure have not been made public.
In order to use the Mistral 7B model checkpoint, which requires about 14 Gb of disk space, download the weights and convert the checkpoint to the lit-gpt format:
pip install huggingface_hub
python scripts/download.py --repo_id mistralai/Mistral-7B-Instruct-v0.1
python scripts/convert_hf_checkpoint.py --checkpoint_dir checkpoints/mistralai/Mistral-7B-Instruct-v0.1
You're done! To execute the model just run:
pip install sentencepiece
python chat/base.py --checkpoint_dir checkpoints/mistralai/Mistral-7B-Instruct-v0.1