Skip to content

SYCL :: EnqueueNativeCommand/custom-command-multiple-dev-cuda.cpp failing in nightly on CUDA #14715

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Closed
sarnex opened this issue Jul 23, 2024 · 2 comments · Fixed by #14717
Closed
Labels
bug Something isn't working confirmed cuda CUDA back-end

Comments

@sarnex
Copy link
Contributor

sarnex commented Jul 23, 2024

Describe the bug

Failed Tests (1):
  SYCL :: EnqueueNativeCommand/custom-command-multiple-dev-cuda.cpp

Command Output (stdout):
--
# RUN: at line 2
/__w/llvm/llvm/toolchain/bin//clang++   -fsycl -fsycl-targets=nvptx64-nvidia-cuda  /__w/llvm/llvm/llvm/sycl/test-e2e/EnqueueNativeCommand/custom-command-multiple-dev-cuda.cpp -o /__w/llvm/llvm/build-e2e/EnqueueNativeCommand/Output/custom-command-multiple-dev-cuda.cpp.tmp.out -lcuda
# executed command: /__w/llvm/llvm/toolchain/bin//clang++ -fsycl -fsycl-targets=nvptx64-nvidia-cuda /__w/llvm/llvm/llvm/sycl/test-e2e/EnqueueNativeCommand/custom-command-multiple-dev-cuda.cpp -o /__w/llvm/llvm/build-e2e/EnqueueNativeCommand/Output/custom-command-multiple-dev-cuda.cpp.tmp.out -lcuda
# note: command had no output on stdout or stderr
# RUN: at line 3
env SYCL_PI_CUDA_ENABLE_IMAGE_SUPPORT=1 ONEAPI_DEVICE_SELECTOR=cuda:gpu  /__w/llvm/llvm/build-e2e/EnqueueNativeCommand/Output/custom-command-multiple-dev-cuda.cpp.tmp.out
# executed command: env SYCL_PI_CUDA_ENABLE_IMAGE_SUPPORT=1 ONEAPI_DEVICE_SELECTOR=cuda:gpu /__w/llvm/llvm/build-e2e/EnqueueNativeCommand/Output/custom-command-multiple-dev-cuda.cpp.tmp.out
# .---command stdout------------
# | 1 devices found
# | Using on device 0: NVIDIA GeForce RTX 3090
# | Checking result on host: FAILED
# | 0 ?= 1
# `-----------------------------
# error: command failed with exit status: 1

https://github.com/intel/llvm/actions/runs/10051976562/job/27782653780

To reproduce

No response

Environment

No response

Additional context

No response

@sarnex sarnex added bug Something isn't working cuda CUDA back-end labels Jul 23, 2024
@sarnex
Copy link
Contributor Author

sarnex commented Jul 23, 2024

@hdelan Looks like this test was added in #14136. Can you please take a look at this?

@hdelan
Copy link
Contributor

hdelan commented Jul 23, 2024

Have made a patch here to not run for single device context #14717 . Although this should still work. I'll try to repro and see what is happening. Potentially there is a race condition somewhere.

@sarnex sarnex closed this as completed in 619185f Jul 29, 2024
AlexeySachkov pushed a commit to AlexeySachkov/llvm that referenced this issue Nov 26, 2024
Wrong command type was being used for enqueue native command, causing a
race condition as the host task codepath was being taken, where another
path should have been taken. This fixes
intel#14715

@sarnex
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working confirmed cuda CUDA back-end
Projects
None yet
Development

Successfully merging a pull request may close this issue.

2 participants