-
Bug
-
Resolution: Unresolved
-
Undefined
-
None
-
4.20
-
Quality / Stability / Reliability
-
False
-
-
None
-
None
-
None
-
None
-
None
-
None
-
OCP Node Sprint 275 (green)
-
1
-
None
-
None
-
None
Description of problem:
When using the latest vllm container image, the model fails to serve due to their backend library (triton) not recognizing CUDA_VISIBLE_DEVICES set to the MIG UUID
Version-Release number of selected component (if applicable):
How reproducible:
Steps to Reproduce:
Pods serving the models using vllm will fail if CUDA_VISIBILE_DEVICES is not represented in the int form due to https://github.com/liammcinroy/vllm/pull/1 https://github.com/vllm-project/vllm/issues/6551 https://github.com/vllm-project/vllm/issues/13815
Actual results:
Expected results:
vllm to work with DAS Operator
Additional info: