We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent a3436f5 commit 8a14e60Copy full SHA for 8a14e60
script/app-mlperf-inference-nvidia/customize.py
@@ -703,6 +703,8 @@ def preprocess(i):
703
else:
704
run_config += f" --tensor_parallelism={tmp_tp_size}"
705
run_config += f" --pipeline_parallelism={tmp_pp_size}"
706
+ if is_true(env.get('MLC_CUSTOM_CONFIG', '')):
707
+ run_config += f" --llm_gen_config_path={env['MLC_MLPERF_INFERENCE_NVIDIA_CODE_PATH']}/code/llama2-70b/tensorrt/generation_config.json"
708
709
enable_sort = env.get('MLC_MLPERF_NVIDIA_HARNESS_ENABLE_SORT')
710
if enable_sort and not is_false(enable_sort):
0 commit comments