XMLWordPrintable

    • False
    • None
    • False

      Inspired by

      caveat:

      ollama: A container based on the Ollama container image that comprises the Ollama web server and is additonally configured to leverage GPUs by setting nvidia.com/gpu: 1 in the container's resource request. Due to that configuration in the devfile, the ollama container (and therewith the entire pod) is being deployed on an OpenShift worker node that hosts a GPU, which significantly accelerates the inference step of the local LLM and hence tremendously improves the performance of the personal AI assistant for developers.
      

      No airgap, s390x support

            ibuziuk@redhat.com Ilya Buziuk
            ibuziuk@redhat.com Ilya Buziuk
            Votes:
            0 Vote for this issue
            Watchers:
            1 Start watching this issue

              Created:
              Updated: