@@ -68,6 +68,8 @@ To check available options use `-h`, `--help` switches:
68
68
``` bash
69
69
docker run benchmark_client --help
70
70
71
+ Client 2.7
72
+ NO_PROXY=localhost no_proxy=localhost python3 /ovms_benchmark_client/main.py --help
71
73
usage: main.py [-h] [-i ID] [-c CONCURRENCY] [-a SERVER_ADDRESS]
72
74
[-p GRPC_PORT] [-r REST_PORT] [-l] [-b [BS ...]]
73
75
[-s [SHAPE ...]] [-d [DATA ...]] [-j] [-m MODEL_NAME]
@@ -77,7 +79,7 @@ usage: main.py [-h] [-i ID] [-c CONCURRENCY] [-a SERVER_ADDRESS]
77
79
[--max_value MAX_VALUE] [--min_value MIN_VALUE] [--xrand XRAND]
78
80
[--dump_png] [--step_timeout STEP_TIMEOUT]
79
81
[--metadata_timeout METADATA_TIMEOUT] [-Y DB_ENDPOINT]
80
- [-y [DB_METADATA ...]] [--print_all] [-ps PRINT_SUMMARY ] [--print_time]
82
+ [-y [DB_METADATA ...]] [--print_all] [-ps] [--print_time]
81
83
[--report_warmup] [--certs_dir CERTS_DIR] [-q STATEFUL_LENGTH]
82
84
[--stateful_id STATEFUL_ID] [--stateful_hop STATEFUL_HOP]
83
85
[--sync_interval SYNC_INTERVAL]
@@ -93,6 +95,8 @@ The version can be checked by using `--internal_version` switch as follows:
93
95
` ` ` bash
94
96
docker run benchmark_client --internal_version
95
97
98
+ Client 2.7
99
+ NO_PROXY= localhost no_proxy= localhost python3 /ovms_benchmark_client/main.py --internal_version
96
100
2.7
97
101
` ` `
98
102
@@ -105,11 +109,11 @@ docker run --network host benchmark_client -a localhost -r 8000 --list_models
105
109
106
110
Client 2.7
107
111
NO_PROXY= localhost no_proxy= localhost python3 /ovms_benchmark_client/main.py -a localhost -r 8000 --list_models
108
- XI worker: try to send request to endpoint: http://localhost:8000/v1/config
109
- XI worker: received status code is 200.
110
- XI worker: found models and their status:
111
- XI worker: model: resnet50-binary-0001, version: 1 - AVAILABLE
112
-
112
+ XW worker: Finished execution. If you want to run inference remove --list_models.
113
+ XI worker: try to send request to endpoint: http://localhost:8000/v1/config
114
+ XI worker: received status code is 200.
115
+ XI worker: found models and their status:
116
+ XI worker: model: resnet50-binary-0001, version: 1 - AVAILABLE
113
117
` ` `
114
118
# # Sample benchmarks
115
119
@@ -123,25 +127,24 @@ docker run --network host benchmark_client -a localhost -r 8000 -l -m resnet50-b
123
127
124
128
Client 2.7
125
129
NO_PROXY= localhost no_proxy= localhost python3 /ovms_benchmark_client/main.py -a localhost -r 8000 -l -m resnet50-binary-0001 -p 9000 -i id
126
- XW id: Finished execution. If you want to run inference remove --list_models.
127
- XI id: try to send request to endpoint: http://localhost:8000/v1/config
128
- XI id: received status code is 200.
129
- XI id: found models and their status:
130
- XI id: model: resnet50-binary-0001, version: 1 - AVAILABLE
131
- XI id: request for metadata of model resnet50-binary-0001...
132
- XI id: Metadata for model resnet50-binary-0001 is downloaded...
133
- XI id: set version of model resnet50-binary-0001: 1
134
- XI id: inputs:
135
- XI id: 0:
136
- XI id: name: 0
137
- XI id: dtype: DT_FLOAT
138
- XI id: tensorShape: {' dim' : [{' size' : ' 1' }, {' size' : ' 3' }, {' size' : ' 224' }, {' size' : ' 224' }]}
139
- XI id: outputs:
140
- XI id: 1463:
141
- XI id: name: 1463
142
- XI id: dtype: DT_FLOAT
143
- XI id: tensorShape: {' dim' : [{' size' : ' 1' }, {' size' : ' 1000' }]}
144
-
130
+ XW id: Finished execution. If you want to run inference remove --list_models.
131
+ XI id: try to send request to endpoint: http://localhost:8000/v1/config
132
+ XI id: received status code is 200.
133
+ XI id: found models and their status:
134
+ XI id: model: resnet50-binary-0001, version: 1 - AVAILABLE
135
+ XI id: request for metadata of model resnet50-binary-0001...
136
+ XI id: Metadata for model resnet50-binary-0001 is downloaded...
137
+ XI id: set version of model resnet50-binary-0001: 1
138
+ XI id: inputs:
139
+ XI id: 0:
140
+ XI id: name: 0
141
+ XI id: dtype: DT_FLOAT
142
+ XI id: tensorShape: {' dim' : [{' size' : ' 1' }, {' size' : ' 3' }, {' size' : ' 224' }, {' size' : ' 224' }]}
143
+ XI id: outputs:
144
+ XI id: 1463:
145
+ XI id: name: 1463
146
+ XI id: dtype: DT_FLOAT
147
+ XI id: tensorShape: {' dim' : [{' size' : ' 1' }, {' size' : ' 1000' }]}
145
148
` ` `
146
149
Be sure the model name specified is identical to the model name shown when using
147
150
the ` --list_models` parameter. A model version is not required but it can be added
@@ -159,127 +162,126 @@ docker run --network host benchmark_client -a localhost -r 8000 -m resnet50-bina
159
162
160
163
Client 2.7
161
164
NO_PROXY=localhost no_proxy=localhost python3 /ovms_benchmark_client/main.py -a localhost -r 8000 -m resnet50-binary-0001 -p 9000 -n 8 --report_warmup --print_all
162
- XI worker: request for metadata of model resnet50-binary-0001...
163
- XI worker: Metadata for model resnet50-binary-0001 is downloaded...
164
- XI worker: set version of model resnet50-binary-0001: 1
165
- XI worker: inputs:
166
- XI worker: 0:
167
- XI worker: name: 0
168
- XI worker: dtype: DT_FLOAT
169
- XI worker: tensorShape: {' dim' : [{' size' : ' 1' }, {' size' : ' 3' }, {' size' : ' 224' }, {' size' : ' 224' }]}
170
- XI worker: outputs:
171
- XI worker: 1463:
172
- XI worker: name: 1463
173
- XI worker: dtype: DT_FLOAT
174
- XI worker: tensorShape: {' dim' : [{' size' : ' 1' }, {' size' : ' 1000' }]}
175
- XI worker: new random range: 0.0, 255.0
176
- XI worker: batchsize sequence: [1]
177
- XI worker: dataset length (0): 1
178
- XI worker: --> dim: 1
179
- XI worker: --> dim: 3
180
- XI worker: --> dim: 224
181
- XI worker: --> dim: 224
182
- XI worker: Generated data shape: (1, 3, 224, 224)
183
- XI worker: start workload...
184
- XI worker: stop warmup: 374943.047975389
185
- XI worker: stop window: inf
186
- XI worker: Workload started!
187
- XI worker: Warmup normally stopped: 374943.075028319
188
- XI worker: Window normally start: 374943.07504102
189
- XI worker: Window stopped: 374943.354446821
190
- XI worker: total_duration: 0.3065074360347353
191
- XI worker: total_batches: 8
192
- XI worker: total_frames: 8
193
- XI worker: start_timestamp: 374943.047975189
194
- XI worker: stop_timestamp: 374943.354482625
195
- XI worker: pass_batches: 8
196
- XI worker: fail_batches: 0
197
- XI worker: pass_frames: 8
198
- XI worker: fail_frames: 0
199
- XI worker: first_latency: 0.027021727000828832
200
- XI worker: pass_max_latency: 0.0449919409584254
201
- XI worker: fail_max_latency: 0.0
202
- XI worker: brutto_batch_rate: 26.100508697261724
203
- XI worker: brutto_frame_rate: 26.100508697261724
204
- XI worker: netto_batch_rate: 26.127558971388062
205
- XI worker: netto_frame_rate: 26.127558971388062
206
- XI worker: frame_passrate: 1.0
207
- XI worker: batch_passrate: 1.0
208
- XI worker: mean_latency: 0.0382737630061456
209
- XI worker: mean_latency2: 0.0015027467953827884
210
- XI worker: stdev_latency: 0.006153524252994284
211
- XI worker: cv_latency: 0.16077656780197483
212
- XI worker: pass_mean_latency: 0.0382737630061456
213
- XI worker: pass_mean_latency2: 0.0015027467953827884
214
- XI worker: pass_stdev_latency: 0.006153524252994284
215
- XI worker: pass_cv_latency: 0.16077656780197483
216
- XI worker: fail_mean_latency: 0.0
217
- XI worker: fail_mean_latency2: 0.0
218
- XI worker: fail_stdev_latency: 0.0
219
- XI worker: fail_cv_latency: 0.0
220
- XI worker: window_total_duration: 0.27940580097492784
221
- XI worker: window_total_batches: 8
222
- XI worker: window_total_frames: 8
223
- XI worker: window_start_timestamp: 374943.07504102
224
- XI worker: window_stop_timestamp: 374943.354446821
225
- XI worker: window_pass_batches: 8
226
- XI worker: window_fail_batches: 0
227
- XI worker: window_pass_frames: 8
228
- XI worker: window_fail_frames: 0
229
- XI worker: window_first_latency: 0.027021727000828832
230
- XI worker: window_pass_max_latency: 0.0449919409584254
231
- XI worker: window_fail_max_latency: 0.0
232
- XI worker: window_brutto_batch_rate: 28.632190069374655
233
- XI worker: window_brutto_frame_rate: 28.632190069374655
234
- XI worker: window_netto_batch_rate: 26.127558971388062
235
- XI worker: window_netto_frame_rate: 26.127558971388062
236
- XI worker: window_frame_passrate: 1.0
237
- XI worker: window_batch_passrate: 1.0
238
- XI worker: window_mean_latency: 0.0382737630061456
239
- XI worker: window_mean_latency2: 0.0015027467953827884
240
- XI worker: window_stdev_latency: 0.006153524252994284
241
- XI worker: window_cv_latency: 0.16077656780197483
242
- XI worker: window_pass_mean_latency: 0.0382737630061456
243
- XI worker: window_pass_mean_latency2: 0.0015027467953827884
244
- XI worker: window_pass_stdev_latency: 0.006153524252994284
245
- XI worker: window_pass_cv_latency: 0.16077656780197483
246
- XI worker: window_fail_mean_latency: 0.0
247
- XI worker: window_fail_mean_latency2: 0.0
248
- XI worker: window_fail_stdev_latency: 0.0
249
- XI worker: window_fail_cv_latency: 0.0
250
- XI worker: window_hist_latency_4: 2
251
- XI worker: window_hist_latency_9: 1
252
- XI worker: window_hist_latency_8: 5
253
- XI worker: warmup_total_duration: 0.02705443004379049
254
- XI worker: warmup_total_batches: 0
255
- XI worker: warmup_total_frames: 0
256
- XI worker: warmup_start_timestamp: 374943.047973889
257
- XI worker: warmup_stop_timestamp: 374943.075028319
258
- XI worker: warmup_pass_batches: 0
259
- XI worker: warmup_fail_batches: 0
260
- XI worker: warmup_pass_frames: 0
261
- XI worker: warmup_fail_frames: 0
262
- XI worker: warmup_first_latency: inf
263
- XI worker: warmup_pass_max_latency: 0.0
264
- XI worker: warmup_fail_max_latency: 0.0
265
- XI worker: warmup_brutto_batch_rate: 0.0
266
- XI worker: warmup_brutto_frame_rate: 0.0
267
- XI worker: warmup_netto_batch_rate: 0.0
268
- XI worker: warmup_netto_frame_rate: 0.0
269
- XI worker: warmup_frame_passrate: 0.0
270
- XI worker: warmup_batch_passrate: 0.0
271
- XI worker: warmup_mean_latency: 0.0
272
- XI worker: warmup_mean_latency2: 0.0
273
- XI worker: warmup_stdev_latency: 0.0
274
- XI worker: warmup_cv_latency: 0.0
275
- XI worker: warmup_pass_mean_latency: 0.0
276
- XI worker: warmup_pass_mean_latency2: 0.0
277
- XI worker: warmup_pass_stdev_latency: 0.0
278
- XI worker: warmup_pass_cv_latency: 0.0
279
- XI worker: warmup_fail_mean_latency: 0.0
280
- XI worker: warmup_fail_mean_latency2: 0.0
281
- XI worker: warmup_fail_stdev_latency: 0.0
282
- XI worker: warmup_fail_cv_latency: 0.0
165
+ XI worker: request for metadata of model resnet50-binary-0001...
166
+ XI worker: Metadata for model resnet50-binary-0001 is downloaded...
167
+ XI worker: set version of model resnet50-binary-0001: 1
168
+ XI worker: inputs:
169
+ XI worker: 0:
170
+ XI worker: name: 0
171
+ XI worker: dtype: DT_FLOAT
172
+ XI worker: tensorShape: {' dim' : [{' size' : ' 1' }, {' size' : ' 3' }, {' size' : ' 224' }, {' size' : ' 224' }]}
173
+ XI worker: outputs:
174
+ XI worker: 1463:
175
+ XI worker: name: 1463
176
+ XI worker: dtype: DT_FLOAT
177
+ XI worker: tensorShape: {' dim' : [{' size' : ' 1' }, {' size' : ' 1000' }]}
178
+ XI worker: new random range: 0.0, 255.0
179
+ XI worker: batchsize sequence: [1]
180
+ XI worker: dataset length (0): 1
181
+ XI worker: --> dim: 1
182
+ XI worker: --> dim: 3
183
+ XI worker: --> dim: 224
184
+ XI worker: --> dim: 224
185
+ XI worker: Generated data shape: (1, 3, 224, 224)
186
+ XI worker: start workload...
187
+ XI worker: stop warmup: 9408188.83686497
188
+ XI worker: stop window: inf
189
+ XI worker: Workload started!
190
+ XI worker: Warmup normally stopped: 9408188.848778868
191
+ XI worker: Window normally start: 9408188.848811286
192
+ XI worker: Window stopped: 9408188.893217305
193
+ XI worker: total_duration: 0.0563836432993412
194
+ XI worker: total_batches: 8
195
+ XI worker: total_frames: 8
196
+ XI worker: start_timestamp: 9408188.836864596
197
+ XI worker: stop_timestamp: 9408188.89324824
198
+ XI worker: pass_batches: 8
199
+ XI worker: fail_batches: 0
200
+ XI worker: pass_frames: 8
201
+ XI worker: fail_frames: 0
202
+ XI worker: first_latency: 0.011858431622385979
203
+ XI worker: pass_max_latency: 0.011858431622385979
204
+ XI worker: fail_max_latency: 0.0
205
+ XI worker: brutto_batch_rate: 141.88512007867135
206
+ XI worker: brutto_frame_rate: 141.88512007867135
207
+ XI worker: netto_batch_rate: 142.7839056346449
208
+ XI worker: netto_frame_rate: 142.7839056346449
209
+ XI worker: frame_passrate: 1.0
210
+ XI worker: batch_passrate: 1.0
211
+ XI worker: mean_latency: 0.00700359046459198
212
+ XI worker: mean_latency2: 5.376289226632219e-05
213
+ XI worker: stdev_latency: 0.002170855331568294
214
+ XI worker: cv_latency: 0.309963202809113
215
+ XI worker: pass_mean_latency: 0.00700359046459198
216
+ XI worker: pass_mean_latency2: 5.376289226632219e-05
217
+ XI worker: pass_stdev_latency: 0.002170855331568294
218
+ XI worker: pass_cv_latency: 0.309963202809113
219
+ XI worker: fail_mean_latency: 0.0
220
+ XI worker: fail_mean_latency2: 0.0
221
+ XI worker: fail_stdev_latency: 0.0
222
+ XI worker: fail_cv_latency: 0.0
223
+ XI worker: window_total_duration: 0.044406019151210785
224
+ XI worker: window_total_batches: 8
225
+ XI worker: window_total_frames: 8
226
+ XI worker: window_start_timestamp: 9408188.848811286
227
+ XI worker: window_stop_timestamp: 9408188.893217305
228
+ XI worker: window_pass_batches: 8
229
+ XI worker: window_fail_batches: 0
230
+ XI worker: window_pass_frames: 8
231
+ XI worker: window_fail_frames: 0
232
+ XI worker: window_first_latency: 0.011858431622385979
233
+ XI worker: window_pass_max_latency: 0.011858431622385979
234
+ XI worker: window_fail_max_latency: 0.0
235
+ XI worker: window_brutto_batch_rate: 180.15575710037206
236
+ XI worker: window_brutto_frame_rate: 180.15575710037206
237
+ XI worker: window_netto_batch_rate: 142.7839056346449
238
+ XI worker: window_netto_frame_rate: 142.7839056346449
239
+ XI worker: window_frame_passrate: 1.0
240
+ XI worker: window_batch_passrate: 1.0
241
+ XI worker: window_mean_latency: 0.00700359046459198
242
+ XI worker: window_mean_latency2: 5.376289226632219e-05
243
+ XI worker: window_stdev_latency: 0.002170855331568294
244
+ XI worker: window_cv_latency: 0.309963202809113
245
+ XI worker: window_pass_mean_latency: 0.00700359046459198
246
+ XI worker: window_pass_mean_latency2: 5.376289226632219e-05
247
+ XI worker: window_pass_stdev_latency: 0.002170855331568294
248
+ XI worker: window_pass_cv_latency: 0.309963202809113
249
+ XI worker: window_fail_mean_latency: 0.0
250
+ XI worker: window_fail_mean_latency2: 0.0
251
+ XI worker: window_fail_stdev_latency: 0.0
252
+ XI worker: window_fail_cv_latency: 0.0
253
+ XI worker: window_hist_latency_1: 1
254
+ XI worker: window_hist_latency_0: 7
255
+ XI worker: warmup_total_duration: 0.011916300281882286
256
+ XI worker: warmup_total_batches: 0
257
+ XI worker: warmup_total_frames: 0
258
+ XI worker: warmup_start_timestamp: 9408188.836862568
259
+ XI worker: warmup_stop_timestamp: 9408188.848778868
260
+ XI worker: warmup_pass_batches: 0
261
+ XI worker: warmup_fail_batches: 0
262
+ XI worker: warmup_pass_frames: 0
263
+ XI worker: warmup_fail_frames: 0
264
+ XI worker: warmup_first_latency: inf
265
+ XI worker: warmup_pass_max_latency: 0.0
266
+ XI worker: warmup_fail_max_latency: 0.0
267
+ XI worker: warmup_brutto_batch_rate: 0.0
268
+ XI worker: warmup_brutto_frame_rate: 0.0
269
+ XI worker: warmup_netto_batch_rate: 0.0
270
+ XI worker: warmup_netto_frame_rate: 0.0
271
+ XI worker: warmup_frame_passrate: 0.0
272
+ XI worker: warmup_batch_passrate: 0.0
273
+ XI worker: warmup_mean_latency: 0.0
274
+ XI worker: warmup_mean_latency2: 0.0
275
+ XI worker: warmup_stdev_latency: 0.0
276
+ XI worker: warmup_cv_latency: 0.0
277
+ XI worker: warmup_pass_mean_latency: 0.0
278
+ XI worker: warmup_pass_mean_latency2: 0.0
279
+ XI worker: warmup_pass_stdev_latency: 0.0
280
+ XI worker: warmup_pass_cv_latency: 0.0
281
+ XI worker: warmup_fail_mean_latency: 0.0
282
+ XI worker: warmup_fail_mean_latency2: 0.0
283
+ XI worker: warmup_fail_stdev_latency: 0.0
284
+ XI worker: warmup_fail_cv_latency: 0.0
283
285
` ` `
284
286
285
287
# # Dynamic models benchmarking
0 commit comments