Replies: 2 comments 1 reply
-
same issue !!! |
Beta Was this translation helpful? Give feedback.
0 replies
-
The default value of |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Description:
I'm encountering an issue when trying to serve two different models using vLLM. The behavior differs between the v1 engine and when forcing v0 engine via
VLLM_USE_V1=0
.VLLM_USE_V1=0
:However, when running without
VLLM_USE_V1=0
:Environment:
Additional Context:
--enforce-eager
modeQuestions:
Beta Was this translation helpful? Give feedback.
All reactions