-
Story
-
Resolution: Done
-
Undefined
-
None
-
None
-
None
-
AIPCC Accelerators 13, AIPCC Accelerators 14
vLLM in CUDA does not use it, it uses vllm's fork of flash-attn:
We don't need to build it anymore for RHAIIS, so we need to update collections/rhaiis/cuda-ubi9/requirements.txt