vllm.entrypoints.openai.responses.api_router ¶
_convert_stream_to_sse_events async ¶
_convert_stream_to_sse_events(
generator: AsyncGenerator[
StreamingResponsesResponse, None
],
) -> AsyncGenerator[str, None]
Convert the generator to a stream of events in SSE format
Source code in vllm/entrypoints/openai/responses/api_router.py
attach_router ¶
cancel_responses async ¶
cancel_responses(response_id: str, raw_request: Request)
Source code in vllm/entrypoints/openai/responses/api_router.py
create_responses async ¶
create_responses(
request: ResponsesRequest, raw_request: Request
)
Source code in vllm/entrypoints/openai/responses/api_router.py
responses ¶
responses(
request: Request,
) -> OpenAIServingResponses | None
retrieve_responses async ¶
retrieve_responses(
response_id: str,
raw_request: Request,
starting_after: int | None = None,
stream: bool | None = False,
)