"add support for vllm api stop parameter"
This commit is contained in:
@@ -96,6 +96,7 @@ class VllmEngine(BaseEngine):
|
||||
repetition_penalty = input_kwargs.pop("repetition_penalty", None)
|
||||
max_length = input_kwargs.pop("max_length", None)
|
||||
max_new_tokens = input_kwargs.pop("max_new_tokens", None)
|
||||
stop = input_kwargs.pop("stop", None)
|
||||
|
||||
generating_args = self.generating_args.copy()
|
||||
generating_args.update(
|
||||
@@ -105,6 +106,7 @@ class VllmEngine(BaseEngine):
|
||||
top_k=top_k or generating_args["top_k"],
|
||||
num_return_sequences=num_return_sequences or 1,
|
||||
repetition_penalty=repetition_penalty or generating_args["repetition_penalty"],
|
||||
stop=stop or generating_args["stop"]
|
||||
)
|
||||
)
|
||||
|
||||
@@ -125,6 +127,7 @@ class VllmEngine(BaseEngine):
|
||||
stop_token_ids=[self.tokenizer.eos_token_id] + self.tokenizer.additional_special_tokens_ids,
|
||||
max_tokens=generating_args["max_new_tokens"],
|
||||
skip_special_tokens=True,
|
||||
stop=generating_args["stop"],
|
||||
)
|
||||
|
||||
if self.processor is not None and image is not None:
|
||||
|
||||
Reference in New Issue
Block a user