Skip to content

Commit

Permalink
benchmark on a100
Browse files Browse the repository at this point in the history
  • Loading branch information
lvhan028 committed Dec 4, 2023
1 parent 4811d84 commit 717f058
Show file tree
Hide file tree
Showing 5 changed files with 150 additions and 7 deletions.
11 changes: 9 additions & 2 deletions benchmark/benchmark-7b.sh
Original file line number Diff line number Diff line change
@@ -1,4 +1,9 @@
#!/bin/bash
if [ - z "$1" ]
then
echo
fi

tp=1
model_name=llama2
model_path=/workspace/models-140/llama2/huggingface/llama-2-7b-chat/
Expand All @@ -16,12 +21,12 @@ apt-get install crudini
crudini --set ${config_path} llama max_context_token_num 4
crudini --set ${config_path} llama cache_chunk_size -1
crudini --set ${config_path} llama cache_max_entry_count 1000
crudini --set ${config_path} llama max_batch_size 256
crudini --set ${config_path} llama max_batch_size 128
# end of update config

benchmark_rpm () {
output_path=$1
mkdir -p ${output_path}
mkdir -p "${output_path}"

batches=(64 128)
for batch in "${batches[@]}"
Expand All @@ -40,6 +45,8 @@ benchmark_rpm () {

benchmark_generation () {
output_path=$1
mkdir -p "${output_path}"

python3 benchmark/profile_generation.py \
${turbomind_model_path} \
--concurrency 1 16 32 64 \
Expand Down
6 changes: 4 additions & 2 deletions benchmark/benchmark_13b.sh
Original file line number Diff line number Diff line change
Expand Up @@ -16,12 +16,12 @@ apt-get install crudini
crudini --set ${config_path} llama max_context_token_num 4
crudini --set ${config_path} llama cache_chunk_size -1
crudini --set ${config_path} llama cache_max_entry_count 500
crudini --set ${config_path} llama max_batch_size 256
crudini --set ${config_path} llama max_batch_size 128
# end of update config

benchmark_rpm () {
output_path=$1
mkdir -p ${output_path}
mkdir -p "${output_path}"

batches=(64 128)
for batch in "${batches[@]}"
Expand All @@ -40,6 +40,8 @@ benchmark_rpm () {

benchmark_generation () {
output_path=$1
mkdir -p "${output_path}"

python3 benchmark/profile_generation.py \
${turbomind_model_path} \
--concurrency 1 16 32 64 \
Expand Down
6 changes: 4 additions & 2 deletions benchmark/benchmark_20b.sh
Original file line number Diff line number Diff line change
Expand Up @@ -16,12 +16,12 @@ apt-get install crudini
crudini --set ${config_path} llama max_context_token_num 4
crudini --set ${config_path} llama cache_chunk_size -1
crudini --set ${config_path} llama cache_max_entry_count 700
crudini --set ${config_path} llama max_batch_size 256
crudini --set ${config_path} llama max_batch_size 128
# end of update config

benchmark_rpm () {
output_path=$1
mkdir -p ${output_path}
mkdir -p "${output_path}"

batches=(64 128)
for batch in "${batches[@]}"
Expand All @@ -40,6 +40,8 @@ benchmark_rpm () {

benchmark_generation () {
output_path=$1
mkdir -p "${output_path}"

python3 benchmark/profile_generation.py \
${turbomind_model_path} \
--concurrency 1 16 32 64 \
Expand Down
4 changes: 3 additions & 1 deletion benchmark/benchmark_70b.sh
Original file line number Diff line number Diff line change
Expand Up @@ -21,7 +21,7 @@ crudini --set ${config_path} llama max_batch_size 256

benchmark_rpm () {
output_path=$1
mkdir -p ${output_path}
mkdir -p "${output_path}"

batches=(64 128 256)
for batch in "${batches[@]}"
Expand All @@ -40,6 +40,8 @@ benchmark_rpm () {

benchmark_generation () {
output_path=$1
mkdir -p "${output_path}"

python3 benchmark/profile_generation.py \
${turbomind_model_path} \
--concurrency 1 64 128 256 \
Expand Down
Loading

0 comments on commit 717f058

Please sign in to comment.