vllm.entrypoints.openai.serving_classification ¶
   ClassificationMixin ¶
  Bases: OpenAIServing
Source code in vllm/entrypoints/openai/serving_classification.py
   _build_render_config ¶
 _build_render_config(
    request: ClassificationRequest,
) -> RenderConfig
  _build_response ¶
 _build_response(
    ctx: ServeContext,
) -> ClassificationResponse | ErrorResponse
Convert model outputs to a formatted classification response with probabilities and labels.
Source code in vllm/entrypoints/openai/serving_classification.py
   _preprocess  async  ¶
 _preprocess(ctx: ServeContext) -> ErrorResponse | None
Process classification inputs: tokenize text, resolve adapters, and prepare model-specific inputs.
Source code in vllm/entrypoints/openai/serving_classification.py
   ServingClassification ¶
  Bases: ClassificationMixin
Source code in vllm/entrypoints/openai/serving_classification.py
   __init__ ¶
 __init__(
    engine_client: EngineClient,
    models: OpenAIServingModels,
    *,
    request_logger: RequestLogger | None,
    log_error_stack: bool = False,
) -> None
Source code in vllm/entrypoints/openai/serving_classification.py
   _create_pooling_params ¶
 _create_pooling_params(
    ctx: ClassificationServeContext,
) -> PoolingParams | ErrorResponse
Source code in vllm/entrypoints/openai/serving_classification.py
   create_classify  async  ¶
 create_classify(
    request: ClassificationRequest, raw_request: Request
) -> ClassificationResponse | ErrorResponse