-
Notifications
You must be signed in to change notification settings - Fork 118
/
eval.sh
executable file
·59 lines (50 loc) · 2.82 KB
/
eval.sh
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
#!/bin/bash
# python eval/perplexity.py -m meta-llama/Llama-2-7b-hf --dataset pg19 --split test --feature text --save-tokenized output/pg19-test-tokenized
PG19="--tokenized emozilla/pg19-test-tokenized"
# python eval/perplexity.py -m meta-llama/Llama-2-7b-hf --dataset tau/scrolls --subset gov_report --split test --feature input --save-tokenized output/govreport-test-tokenized
GOVREPORT="--tokenized emozilla/govreport-test-tokenized --dataset-min-tokens 16384 --samples 50"
# python eval/perplexity.py -m meta-llama/Llama-2-7b-hf --dataset hoskinson-center/proof-pile --split test --feature text --save-tokenized output/proofpile-test-tokenized
PROOFPILE="--tokenized emozilla/proofpile-test-tokenized --dataset-min-tokens 32768 --samples 50"
PROOFPILE_LONG_SMALL="--tokenized emozilla/proofpile-test-tokenized --dataset-min-tokens 131072 --samples 10 --truncate"
# python eval/perplexity.py -m mistralai/Mistral-7B-v0.1 --dataset hoskinson-center/proof-pile --split test --feature text --save-tokenized output/proofpile-test-tokenized-mistral
PROOFPILE_LONG_SMALL_MISTRAL="--tokenized emozilla/proofpile-test-tokenized-mistral --dataset-min-tokens 131072 --samples 10 --truncate --split train"
CUSTOM="--custom-model-together"
python eval/perplexity.py \
${PROOFPILE_LONG_SMALL} ${CUSTOM} \
--output-file data/proofpile-long-small.csv \
--min-tokens 2048 --max-tokens 131072 --tokens-step 2048 --aggressive-memory \
-m NousResearch/CodeLlama-13b-hf \
-m NousResearch/Yarn-Llama-2-13b-64k \
-m NousResearch/Yarn-Llama-2-13b-128k \
-m togethercomputer/LLaMA-2-7B-32K \
-m NousResearch/CodeLlama-7b-hf \
-m NousResearch/Yarn-Llama-2-7b-64k \
-m NousResearch/Yarn-Llama-2-7b-128k
python eval/perplexity.py \
${GOVREPORT} ${CUSTOM} \
--output-file data/govreport.csv \
--min-tokens 32768 --max-tokens 32768 \
-m NousResearch/CodeLlama-13b-hf \
-m NousResearch/Yarn-Llama-2-13b-64k \
-m NousResearch/Yarn-Llama-2-13b-128k \
-m togethercomputer/LLaMA-2-7B-32K \
-m NousResearch/CodeLlama-7b-hf \
-m NousResearch/Yarn-Llama-2-7b-64k \
-m NousResearch/Yarn-Llama-2-7b-128k
python eval/perplexity.py \
${PROOFPILE_LONG_SMALL} ${CUSTOM} \
--output-file data/proofpile-long-small-8k.csv \
--min-tokens 2048 --max-tokens 16384 --tokens-step 2048 \
-m emozilla/Yarn-Llama-2-7b-8k \
-m emozilla/NTK-Llama-2-7b-8k \
-m conceptofmind/LLongMA-2-7b
python eval/perplexity.py \
${PROOFPILE_LONG_SMALL_MISTRAL} \
--output-file data/proofpile-long-small-mistral.csv \
--flash-attention --custom-model-mistral \
--min-tokens 2048 --max-tokens 131072 --tokens-step 2048 --aggressive-memory \
--sliding-window-attention 131072 \
-m NousResearch/Yarn-Mistral-7b-64k \
-m NousResearch/Yarn-Mistral-7b-128k \
-m amazon/MistralLite \
-m mistralai/Mistral-7B-v0.1