-
Notifications
You must be signed in to change notification settings - Fork 275
/
Copy pathbenchmark_docsum.yaml
87 lines (76 loc) · 2.5 KB
/
benchmark_docsum.yaml
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
# Copyright (C) 2025 Intel Corporation
# SPDX-License-Identifier: Apache-2.0
deploy:
device: gaudi
version: 1.2.0
modelUseHostPath: /mnt/models
HUGGINGFACEHUB_API_TOKEN: "" # mandatory
node: [1]
namespace: ""
node_name: []
timeout: 1000 # timeout in seconds for services to be ready, default 30 minutes
interval: 5 # interval in seconds between service ready checks, default 5 seconds
services:
backend:
resources:
enabled: False
cores_per_instance: "16"
memory_capacity: "8000Mi"
replicaCount: [1]
teirerank:
enabled: False
llm:
engine: vllm # or tgi
model_id: "meta-llama/Llama-3.2-3B-Instruct" # mandatory
replicaCount:
without_teirerank: [1] # When teirerank.enabled is False
resources:
enabled: False
cards_per_instance: 1
model_params:
vllm: # VLLM specific parameters
batch_params:
enabled: True
max_num_seqs: "8" # Each value triggers an LLM service upgrade
token_params:
enabled: True
max_input_length: ""
max_total_tokens: ""
max_batch_total_tokens: ""
max_batch_prefill_tokens: ""
tgi: # TGI specific parameters
batch_params:
enabled: True
max_batch_size: [1] # Each value triggers an LLM service upgrade
token_params:
enabled: False
max_input_length: "1280"
max_total_tokens: "2048"
max_batch_total_tokens: "65536"
max_batch_prefill_tokens: "4096"
docsum-ui:
replicaCount: [1]
whisper:
replicaCount: [1]
llm-uservice:
model_id: "meta-llama/Llama-3.2-3B-Instruct" # mandatory
replicaCount: [1]
nginx:
replicaCount: [1]
benchmark:
# http request behavior related fields
user_queries: [16]
concurrency: [4]
load_shape_type: "constant" # "constant" or "poisson"
poisson_arrival_rate: 1.0 # only used when load_shape_type is "poisson"
warmup_iterations: 10
seed: 1024
collect_service_metric: True
# workload, all of the test cases will run for benchmark
bench_target: ["docsumfixed"] # specify the bench_target for benchmark
dataset: "/home/sdp/upload.txt" # specify the absolute path to the dataset file
summary_type: "stuff"
stream: True
llm:
# specify the llm output token size
max_token_size: [1024]