Skip to content

vllm.renderers.protocol

RendererLike

Bases: Protocol

Source code in vllm/renderers/protocol.py
class RendererLike(Protocol):
    @classmethod
    def from_config(
        cls,
        config: "ModelConfig",
        tokenizer_kwargs: dict[str, Any],
    ) -> "RendererLike":
        raise NotImplementedError

    @property
    def tokenizer(self) -> TokenizerLike | None:
        raise NotImplementedError

    def get_tokenizer(self) -> TokenizerLike:
        tokenizer = self.tokenizer
        if tokenizer is None:
            raise ValueError("Tokenizer not available when `skip_tokenizer_init=True`")

        return tokenizer

    def render_messages(
        self,
        messages: list["ChatCompletionMessageParam"],
        **kwargs,
    ) -> tuple[list["ConversationMessage"], TextPrompt | TokensPrompt]:
        raise NotImplementedError

    async def render_messages_async(
        self,
        messages: list["ChatCompletionMessageParam"],
        **kwargs,
    ) -> tuple[list["ConversationMessage"], TextPrompt | TokensPrompt]:
        return self.render_messages(messages, **kwargs)

tokenizer property

tokenizer: TokenizerLike | None

from_config classmethod

from_config(
    config: ModelConfig, tokenizer_kwargs: dict[str, Any]
) -> RendererLike
Source code in vllm/renderers/protocol.py
@classmethod
def from_config(
    cls,
    config: "ModelConfig",
    tokenizer_kwargs: dict[str, Any],
) -> "RendererLike":
    raise NotImplementedError

get_tokenizer

get_tokenizer() -> TokenizerLike
Source code in vllm/renderers/protocol.py
def get_tokenizer(self) -> TokenizerLike:
    tokenizer = self.tokenizer
    if tokenizer is None:
        raise ValueError("Tokenizer not available when `skip_tokenizer_init=True`")

    return tokenizer

render_messages

render_messages(
    messages: list[ChatCompletionMessageParam], **kwargs
) -> tuple[
    list[ConversationMessage], TextPrompt | TokensPrompt
]
Source code in vllm/renderers/protocol.py
def render_messages(
    self,
    messages: list["ChatCompletionMessageParam"],
    **kwargs,
) -> tuple[list["ConversationMessage"], TextPrompt | TokensPrompt]:
    raise NotImplementedError

render_messages_async async

render_messages_async(
    messages: list[ChatCompletionMessageParam], **kwargs
) -> tuple[
    list[ConversationMessage], TextPrompt | TokensPrompt
]
Source code in vllm/renderers/protocol.py
async def render_messages_async(
    self,
    messages: list["ChatCompletionMessageParam"],
    **kwargs,
) -> tuple[list["ConversationMessage"], TextPrompt | TokensPrompt]:
    return self.render_messages(messages, **kwargs)