XMLWordPrintable

    • False
    • Hide

      None

      Show
      None
    • False

      Inspired by

      caveat:

      ollama: A container based on the Ollama container image that comprises the Ollama web server and is additonally configured to leverage GPUs by setting nvidia.com/gpu: 1 in the container's resource request. Due to that configuration in the devfile, the ollama container (and therewith the entire pod) is being deployed on an OpenShift worker node that hosts a GPU, which significantly accelerates the inference step of the local LLM and hence tremendously improves the performance of the personal AI assistant for developers.
      

      No airgap, s390x support

              ibuziuk@redhat.com Ilya Buziuk
              ibuziuk@redhat.com Ilya Buziuk
              Votes:
              0 Vote for this issue
              Watchers:
              1 Start watching this issue

                Created:
                Updated:
                Resolved: