-
Story
-
Resolution: Done
-
Undefined
-
None
-
None
-
8
-
False
-
-
False
-
-
During the investigation of AI Workloads RHELPERF-82 I have noticed issues with building the vLLM Inference Engine. The vLLM github repo provides a variety of 'Dockerfiles' and my use of 'podman build' with several of these (Dockerfile, Dockerfile.cpu, Dockerfile.arm) have led to unexpected results. This STORY will document them an lead to submission of upstream issues.
- clones
-
RHELPERF-82 Evaluate multi-arch AI Workloads for power usage
-
- In Progress
-