-
Notifications
You must be signed in to change notification settings - Fork 16
/
example_config.yaml
69 lines (65 loc) · 1.93 KB
/
example_config.yaml
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
---
s3_config:
- access_key: abc
secret_key: as
region: eu-central-1
endpoint: https://my.rgw.endpoint:8080
skipSSLverify: false
- access_key: def
secret_key: as
region: eu-central-2
endpoint: https://my.rgw.endpoint:8080
skipSSLverify: false
- access_key: ghi
secret_key: as
region: eu-central-3
endpoint: https://my.rgw.endpoint:8080
skipSSLverify: false
# For generating annotations when we start/stop testcases
# https://grafana.com/docs/http_api/annotations/#create-annotation
grafana_config:
endpoint: http://grafana
username: admin
password: grafana
tests:
- name: My first example test
read_weight: 20
existing_read_weight: 0
write_weight: 80
delete_weight: 0
list_weight: 0
objects:
size_min: 5
size_max: 100
part_size: 0
# distribution: constant, random, sequential
size_distribution: random
unit: KB
number_min: 10
number_max: 10
# distribution: constant, random, sequential
number_distribution: constant
buckets:
number_min: 1
number_max: 10
# distribution: constant, random, sequential
number_distribution: constant
# Name prefix for buckets and objects
bucket_prefix: 1255gosbench-
object_prefix: obj
# End after a set amount of time
# Runtime in time.Duration - do not forget the unit please
# stop_with_runtime: 60s # Example with 60 seconds runtime
stop_with_runtime:
# End after a set amount of operations (per worker)
stop_with_ops: 10
# Number of s3 performance test servers to run in parallel
workers: 2
# Set wheter workers share the same buckets or not
# If set to True - bucket names will have the worker # appended
workers_share_buckets: True
# Number of requests processed in parallel by each worker
parallel_clients: 3
# Remove all generated buckets and its content after run
clean_after: True
...