inductor/test_gpu_cpp_wrapper tests are failing because of NotImplementedError: Could not run 'torchvision::roi_align' with arguments from the 'CUDA' backend on main branch
Tests Failing:
test_roi_align_cuda_gpu_wrapper, test_roi_align_cuda_dynamic_shapes_gpu_wrapper
Env details:
PyTorch version: 2.10.0
Branch: main
OS: RHEL 9.6
CPU: Intel
python version: 3.12
commit id : 6de6685797cabc6256df76803f3a5f772d5275a7 (tag: trunk/6de6685797cabc6256df76803f3a5f772d5275a7, origin/main, origin/HEAD)
Steps to repro:
Log in to H200.
Login to quay.io: podman login quay.io
Pull base image: podman pull quay.io/aipcc/pytorch:rhel_cuda_build_without_pins
Run the image and specify the GPU to be used: podman run -it <IMAGE_NAME>
Run the PyTorch UT: TEST_CONFIG=cpu python3 test/run_test.py -i inductor.test_gpu_cpp_wrapper
Expected result: UTs should run fine.
Actual result:
NotImplementedError: Could not run 'torchvision::roi_align' with arguments from the 'CUDA' backend. This could be because the operator doesn't exist for this backend, or was omitted during the selective/custom build process (if using custom build). The 'torchvision::roi_align' operator is only available for CPU, Meta, QuantizedCPU, and other backends, but not for CUDA backend.
Logs are attached below
- is related to
-
AIPCC-8151 [QA][PyTorch UT][GPU] inductor/test_gpu_cpp_wrapper and inductor/test_compile_subprocess modules are failing due to NotImplemented Error
-
- Closed
-