We provide several profiling tools to benchmark our models.
Download the dataset below or create your own dataset.
wget https://huggingface.co/datasets/anon8231489123/ShareGPT_Vicuna_unfiltered/resolve/main/ShareGPT_V3_unfiltered_cleaned_split.json
Profiling your model with profile_throughput.py
python profile_throughput.py \
ShareGPT_V3_unfiltered_cleaned_split.json \
/path/to/your/model \
--concurrency 64
profile_generation.py
perform benchmark with dummy data.
python profile_generation.py \
/path/to/your/model \
--concurrency 8 --input_seqlen 0 --output_seqlen 2048
Tools above profile models with Python API. profile_serving.py
is used to do benchmark on serving.
wget https://huggingface.co/datasets/anon8231489123/ShareGPT_Vicuna_unfiltered/resolve/main/ShareGPT_V3_unfiltered_cleaned_split.json
python profile_serving.py \
${TritonServerAddress} \
/path/to/tokenizer \
ShareGPT_V3_unfiltered_cleaned_split.json \
--concurrency 64