-
Epic
-
Resolution: Done
-
High
-
None
-
None
-
None
Epic Goal
- Run an AI inference workload based on vLLM server with confidential GPU on baremetal
Why is this important?
- Protecting AI model and inference data
Scenarios
- As an AI application owner, I should be able to deploy my inference workload in a confidential GPU environment
- ...
- ...
Acceptance Criteria
(The Epic is complete when...)
- Ability to execute vLLM inference server using confidential GPU on baremetal
- ..
- ..
Additional context:
- depends on
-
KATA-4035 Providing podVM image with Nvidia Confidential GPU support for Azure
-
- New
-
-
KATA-4111 kata containers rpm: Create custom initrds for the TDX kata-nvidia-gpu runtime class
-
- In Progress
-
-
KATA-3701 kata containers rpm: Create custom initrds for the SNP kata-nvidia-gpu runtime class
-
- Closed
-