Uploaded image for project: 'Red Hat Enterprise Linux AI'
  1. Red Hat Enterprise Linux AI
  2. RHELAI-2667

Upstream vLLM MP backend on Gaudi 3 cards

XMLWordPrintable

    • Icon: Spike Spike
    • Resolution: Unresolved
    • Icon: Undefined Undefined
    • None
    • RHELAI 1.4 GA
    • InstructLab - Training
    • None
    • False
    • Hide

      None

      Show
      None
    • False

      Intel's vllm-fork v0.5.3 only supports the Ray distributed executor for multi-card inference. Having merged into upstream vLLM, the new 0.6.2 release may have relaxed this requirement, unblocking multi-card inference for RHEL AI.

       

      Investigate the status of this dependency relaxation in upstream vLLM for Gaudi inference.

              rhn-support-jkunstle James Kunstle
              rhn-support-jkunstle James Kunstle
              Votes:
              0 Vote for this issue
              Watchers:
              1 Start watching this issue

                Created:
                Updated: