Uploaded image for project: 'Hybrid Cloud Console'
  1. Hybrid Cloud Console
  2. RHCLOUD-43238

SPIKE: Deploy LLM using llm-d

XMLWordPrintable

    • Product / Portfolio Work
    • False
    • Hide

      None

      Show
      None
    • False
    • None
    • Unset
    • None

      Deploy an LLM using llm-d

      llm-d is a platform for inference serving that can be bigger than just using VLLM (in fact, it is used under the hood), but it can bring us some key features to scale up our AI platform. We need to verify the identify the pros and cons to install/maintain/usage.

      To perform this spike, it's important to follow a GitOps approach, since ArgoCD will be our deployment platform on the HCMAI cluster.

              rh-ee-jbarea Juan Manuel Barea Martinez
              rh-ee-jbarea Juan Manuel Barea Martinez
              Votes:
              0 Vote for this issue
              Watchers:
              2 Start watching this issue

                Created:
                Updated:
                Resolved: