From f42c45e5c628fa8b1c4176183dd1d7fc9805f396 Mon Sep 17 00:00:00 2001 From: maleksan85 Date: Fri, 17 Jan 2025 18:20:53 -0600 Subject: [PATCH] restoring default in main vllm code for detokenize --- vllm/sampling_params.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/vllm/sampling_params.py b/vllm/sampling_params.py index 62f6a20e13d1b..c1a7bfd0c82b4 100644 --- a/vllm/sampling_params.py +++ b/vllm/sampling_params.py @@ -195,7 +195,7 @@ class SamplingParams( # NOTE: This parameter is only exposed at the engine level for now. # It is not exposed in the OpenAI API server, as the OpenAI API does # not support returning only a list of token IDs. - detokenize: bool = False + detokenize: bool = True skip_special_tokens: bool = True spaces_between_special_tokens: bool = True # Optional[List[LogitsProcessor]] type. We use Any here because