"add support for vllm api stop parameter"

This commit is contained in:
zhaonx
2024-04-30 17:17:09 +08:00
parent d27e6a46b4
commit 42edc81585
4 changed files with 12 additions and 3 deletions

View File

@@ -96,6 +96,7 @@ class VllmEngine(BaseEngine):
repetition_penalty = input_kwargs.pop("repetition_penalty", None)
max_length = input_kwargs.pop("max_length", None)
max_new_tokens = input_kwargs.pop("max_new_tokens", None)
stop = input_kwargs.pop("stop", None)
generating_args = self.generating_args.copy()
generating_args.update(
@@ -105,6 +106,7 @@ class VllmEngine(BaseEngine):
top_k=top_k or generating_args["top_k"],
num_return_sequences=num_return_sequences or 1,
repetition_penalty=repetition_penalty or generating_args["repetition_penalty"],
stop=stop or generating_args["stop"]
)
)
@@ -125,6 +127,7 @@ class VllmEngine(BaseEngine):
stop_token_ids=[self.tokenizer.eos_token_id] + self.tokenizer.additional_special_tokens_ids,
max_tokens=generating_args["max_new_tokens"],
skip_special_tokens=True,
stop=generating_args["stop"],
)
if self.processor is not None and image is not None: