vllm load model error

#3
by Dharma0818 - opened

When I use vllm to load the gemma:2-27b-chinese-chat model, I receive the following error message. What is causing this issue?

ERROR 07-12 10:40:51 serving_chat.py:238] Error in applying chat template from request: 'system_message' is undefined
INFO:     100.89.61.1:55874 - "POST /v1/chat/completions HTTP/1.1" 400 Bad Request

the same situation in xinference "'system_message' is undefined"
在 xinference 加速推理框架中,同样提示“'system_message' is undefined”

Sign up or log in to comment