parent
b1c255630d
commit
7c3604fb68
@ -109,7 +109,7 @@ class ChatCompletionRequest(OpenAIBaseModel):
|
||||
frequency_penalty: Optional[float] = 0.0
|
||||
logit_bias: Optional[Dict[str, float]] = None
|
||||
logprobs: Optional[bool] = False
|
||||
top_logprobs: Optional[int] = None
|
||||
top_logprobs: Optional[int] = 0
|
||||
max_tokens: Optional[int] = None
|
||||
n: Optional[int] = 1
|
||||
presence_penalty: Optional[float] = 0.0
|
||||
@ -192,8 +192,7 @@ class ChatCompletionRequest(OpenAIBaseModel):
|
||||
# doc: end-chat-completion-extra-params
|
||||
|
||||
def to_sampling_params(self) -> SamplingParams:
|
||||
if self.logprobs and not self.top_logprobs:
|
||||
raise ValueError("Top logprobs must be set when logprobs is.")
|
||||
# We now allow logprobs being true without top_logrobs.
|
||||
|
||||
logits_processors = None
|
||||
if self.logit_bias:
|
||||
|
||||
@ -286,7 +286,7 @@ class OpenAIServingChat(OpenAIServing):
|
||||
logprobs = self._create_logprobs(
|
||||
token_ids=delta_token_ids,
|
||||
top_logprobs=top_logprobs,
|
||||
num_output_top_logprobs=request.logprobs,
|
||||
num_output_top_logprobs=request.top_logprobs,
|
||||
initial_text_offset=len(previous_texts[i]),
|
||||
)
|
||||
else:
|
||||
@ -373,7 +373,7 @@ class OpenAIServingChat(OpenAIServing):
|
||||
logprobs = self._create_logprobs(
|
||||
token_ids=token_ids,
|
||||
top_logprobs=top_logprobs,
|
||||
num_output_top_logprobs=request.logprobs,
|
||||
num_output_top_logprobs=request.top_logprobs,
|
||||
)
|
||||
else:
|
||||
logprobs = None
|
||||
|
||||
Loading…
Reference in New Issue
Block a user