Skip to content

Commit 8809a34

Browse files
author
Omar Abdelwahab
committed
Removed additional dead code
1 parent f608004 commit 8809a34

File tree

2 files changed

+169
-172
lines changed

2 files changed

+169
-172
lines changed

llama_stack/apis/inference/inference.py

Lines changed: 0 additions & 28 deletions
Original file line numberDiff line numberDiff line change
@@ -351,34 +351,6 @@ class CompletionRequest(BaseModel):
351351
logprobs: LogProbConfig | None = None
352352

353353

354-
# @json_schema_type
355-
# class CompletionResponse(MetricResponseMixin):
356-
# """Response from a completion request.
357-
358-
# :param content: The generated completion text
359-
# :param stop_reason: Reason why generation stopped
360-
# :param logprobs: Optional log probabilities for generated tokens
361-
# """
362-
363-
# content: str
364-
# stop_reason: StopReason
365-
# logprobs: list[TokenLogProbs] | None = None
366-
367-
368-
# @json_schema_type
369-
# class CompletionResponseStreamChunk(MetricResponseMixin):
370-
# """A chunk of a streamed completion response.
371-
372-
# :param delta: New content generated since last chunk. This can be one or more tokens.
373-
# :param stop_reason: Optional reason why generation stopped, if complete
374-
# :param logprobs: Optional log probabilities for generated tokens
375-
# """
376-
377-
# delta: str
378-
# stop_reason: StopReason | None = None
379-
# logprobs: list[TokenLogProbs] | None = None
380-
381-
382354
class SystemMessageBehavior(Enum):
383355
"""Config for how to override the default system prompt.
384356

0 commit comments

Comments
 (0)