Uploaded image for project: 'Openshift sandboxed containers'
  1. Openshift sandboxed containers
  2. KATA-4074

NVIDIA GTC demo using VLLM server

XMLWordPrintable

    • Icon: Epic Epic
    • Resolution: Done
    • Icon: High High
    • None
    • None
    • None
    • NVIDIA GTC demo using VLLM server
    • Product / Portfolio Work
    • False
    • Hide

      None

      Show
      None
    • False
    • Not Selected
    • Done
    • KATA-3747 - Confidential Containers on Bare Metal [Technology Preview]
    • KATA-3747Confidential Containers on Bare Metal [Technology Preview]
    • 0% To Do, 0% In Progress, 100% Done

      Epic Goal

      • Run an AI inference workload based on vLLM server with confidential GPU on baremetal

      Why is this important?

      • Protecting AI model and inference data

      Scenarios

      1. As an AI application owner, I should be able to deploy my inference workload in a confidential GPU environment
      2. ...
      3. ...

      Acceptance Criteria 

      (The Epic is complete when...)

      1. Ability to execute vLLM inference server using confidential GPU on baremetal
      2. ..
      3. ..

      Additional context:

              bpradipt Pradipta Banerjee
              bpradipt Pradipta Banerjee
              Votes:
              0 Vote for this issue
              Watchers:
              2 Start watching this issue

                Created:
                Updated:
                Resolved: