Skip to content

Commit de28e17

Browse files
authored
[llm_bench] Add setting of max_num_batched_tokens for SchedulerConfig (#2316)
1 parent 0047931 commit de28e17

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

tools/llm_bench/llm_bench_utils/ov_utils.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -168,7 +168,7 @@ def create_text_gen_model(model_path, device, memory_monitor, **kwargs):
168168
def get_scheduler_config_genai(user_config, config_name="CB config"):
169169
import openvino_genai
170170

171-
default_cb_config = {"cache_size": 1}
171+
default_cb_config = {"cache_size": 1, "max_num_batched_tokens": 65535}
172172
scheduler_config = openvino_genai.SchedulerConfig()
173173
scheduler_params = user_config or default_cb_config
174174
if scheduler_params:

0 commit comments

Comments
 (0)