• Icon: Sub-task Sub-task
    • Resolution: Unresolved
    • Icon: Undefined Undefined
    • None
    • None
    • Kiali
    • None
    • False
    • Hide

      None

      Show
      None
    • False

      Create a standalone Python-based service (e.g., Flask or FastAPI) that exposes endpoints to serve responses from different LLMs, starting with the fine-tuned Kiali model. The service must support:

      • Multi-model backend (via dynamic loading or routing)
      • Token usage and latency metrics
      • Configurable inference parameters
      • Optional disk-based metric logging

        1. Screenshot From 2025-06-23 09-33-30.png
          358 kB
          Alberto Jesus Gutierrez Juanes

              agutierr@redhat.com Alberto Jesus Gutierrez Juanes
              agutierr@redhat.com Alberto Jesus Gutierrez Juanes
              Votes:
              0 Vote for this issue
              Watchers:
              1 Start watching this issue

                Created:
                Updated: