vllm.model_executor.guided_decoding.outlines_decoding
GuidedDecodingMode
¶
Bases: Enum
Source code in vllm/model_executor/guided_decoding/outlines_decoding.py
_get_guide_and_mode
¶
_get_guide_and_mode(
guided_params: GuidedDecodingParams,
) -> Union[
tuple[str, GuidedDecodingMode], tuple[None, None]
]
Source code in vllm/model_executor/guided_decoding/outlines_decoding.py
_get_logits_processor
¶
_get_logits_processor(
guide: str,
tokenizer: PreTrainedTokenizerBase,
mode: GuidedDecodingMode,
whitespace_pattern: Union[str, None],
reasoner: Optional[ReasoningParser],
) -> Union[JSONLogitsProcessor, RegexLogitsProcessor]
Source code in vllm/model_executor/guided_decoding/outlines_decoding.py
get_local_outlines_guided_decoding_logits_processor
¶
get_local_outlines_guided_decoding_logits_processor(
guided_params: GuidedDecodingParams,
tokenizer: PreTrainedTokenizerBase,
reasoner: Optional[ReasoningParser],
) -> Union[JSONLogitsProcessor, RegexLogitsProcessor, None]
Given an OpenAI-compatible request, check for guided decoding parameters and get the necessary logits processor for the given guide.
Source code in vllm/model_executor/guided_decoding/outlines_decoding.py
get_outlines_guided_decoding_logits_processor
async
¶
get_outlines_guided_decoding_logits_processor(
guided_params: GuidedDecodingParams,
tokenizer: PreTrainedTokenizerBase,
reasoner: Optional[ReasoningParser],
) -> Union[JSONLogitsProcessor, RegexLogitsProcessor, None]
Given an OpenAI-compatible request, check for guided decoding parameters and get the necessary logits processor for the given guide.