Skip to content

Commit 3513d99

Browse files
committed
fix: hardcode the model max length for vllm
1 parent 5f0743d commit 3513d99

File tree

1 file changed

+2
-1
lines changed

1 file changed

+2
-1
lines changed

bigcodebench/provider/vllm.py

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -41,7 +41,8 @@ def __init__(self, name: str, lora_path: str, dataset: str, tp: int, **kwargs) -
4141
local_lora_path,
4242
)
4343

44-
self.llm = LLM(model=name, enable_lora=True if self.lora_request else False, **kwargs)
44+
# max_model_len is set to max_new_tokens * 10
45+
self.llm = LLM(model=name, max_model_len=self.max_new_tokens * 10, enable_lora=True if self.lora_request else False, **kwargs)
4546
self.llm.set_tokenizer(tokenizer=self.tokenizer)
4647

4748
def is_direct_completion(self) -> bool:

0 commit comments

Comments
 (0)