Skip to content

[LMI v20] ModuleNotFoundError: vllm.entrypoints.openai.protocol breaks chat completions #5719

@remilstl

Description

@remilstl

Cross-posting from deepjavalibrary/djl-serving#3000 as this affects the AWS LMI DLC images and i'm not getting any traction there.

In LMI v19/v20 (DJL 0.36.0), any request using messages format fails with ModuleNotFoundError: No module named 'vllm.entrypoints.openai.protocol' because vllm_chat_utils.py references a module path that was refactored in vLLM 0.13.0+.

Affected images: 763104351884.dkr.ecr.ca-central-1.amazonaws.com/djl-inference:0.36.0-lmi20.0.0-cu128-v1.0

I am able to get some predictions by formatting my requests in sagemaker with "inputs" instead of "messages" but it seems to have the side effects of breaking some other parameters such as guided_choices, log probs and so on.

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions