Skip to content
This repository was archived by the owner on Mar 28, 2023. It is now read-only.

Commit a93bc8c

Browse files
author
Hugh Delaney
committed
Fixing merge conflict
2 parents b9e3bd9 + cc4bf25 commit a93bc8c

File tree

168 files changed

+5756
-1163
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

168 files changed

+5756
-1163
lines changed

.github/CODEOWNERS

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1,9 +1,8 @@
1-
* @vladimirlaz @romanovvlad @bader
1+
* @pvchupin
22

33
# Use runtime team as the umbrella for most of the tests
44
/SYCL/ @intel/llvm-reviewers-runtime
55

6-
76
# SYCL sub-directory matchers are grouped by code owner first, followed by
87
# alphabetical order within the group. Please, keep this ordering.
98

.github/workflows/clang-format.yml

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -20,6 +20,7 @@ jobs:
2020
- name: Run clang-format for the patch
2121
shell: bash {0}
2222
run: |
23+
git config --global --add safe.directory /__w/llvm-test-suite/llvm-test-suite
2324
git clang-format ${GITHUB_SHA}^1
2425
git diff > ./clang-format.patch
2526

SYCL/AtomicRef/add.cpp

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,7 +1,7 @@
11
// See https://github.com/intel/llvm-test-suite/issues/867 for detailed status
22
// UNSUPPORTED: hip
33

4-
// RUN: %clangxx -fsycl -fsycl-unnamed-lambda -fsycl-device-code-split=per_kernel -fsycl-targets=%sycl_triple %s -o %t.out -Xsycl-target-backend=nvptx64-nvidia-cuda --cuda-gpu-arch=sm_70
4+
// RUN: %clangxx -fsycl -fsycl-unnamed-lambda -fsycl-device-code-split=per_kernel -fsycl-targets=%sycl_triple %s -o %t.out
55
// RUN: %HOST_RUN_PLACEHOLDER %t.out
66
// RUN: %GPU_RUN_PLACEHOLDER %t.out
77
// RUN: %CPU_RUN_PLACEHOLDER %t.out

SYCL/AtomicRef/add.h

Lines changed: 0 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -287,10 +287,6 @@ template <access::address_space space, typename T, typename Difference = T,
287287
void add_test_scopes(queue q, size_t N) {
288288
std::vector<memory_scope> scopes =
289289
q.get_device().get_info<info::device::atomic_memory_scope_capabilities>();
290-
if (std::find(scopes.begin(), scopes.end(), memory_scope::system) !=
291-
scopes.end()) {
292-
add_test<space, T, Difference, order, memory_scope::system>(q, N);
293-
}
294290
if (std::find(scopes.begin(), scopes.end(), memory_scope::work_group) !=
295291
scopes.end()) {
296292
add_test<space, T, Difference, order, memory_scope::work_group>(q, N);

SYCL/AtomicRef/add_generic.cpp

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -1,11 +1,11 @@
1-
// RUN: %clangxx -fsycl -fsycl-unnamed-lambda -fsycl-targets=%sycl_triple %s -o %t.out -Xsycl-target-backend=nvptx64-nvidia-cuda --cuda-gpu-arch=sm_70
1+
// RUN: %clangxx -fsycl -fsycl-unnamed-lambda -fsycl-targets=%sycl_triple %s -o %t.out
22
// RUN: %HOST_RUN_PLACEHOLDER %t.out
33
// RUN: %GPU_RUN_PLACEHOLDER %t.out
44
// RUN: %CPU_RUN_PLACEHOLDER %t.out
55
// RUN: %ACC_RUN_PLACEHOLDER %t.out
66

7-
// CUDA and HIP backends have had no support for the generic address space yet
8-
// XFAIL: cuda || hip
7+
// HIP backend has no support for the generic address space yet
8+
// XFAIL: hip
99

1010
#include "add.h"
1111

SYCL/AtomicRef/add_generic_local.cpp

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -1,12 +1,12 @@
1-
// RUN: %clangxx -fsycl -fsycl-unnamed-lambda -fsycl-targets=%sycl_triple %s -o %t.out -Xsycl-target-backend=nvptx64-nvidia-cuda --cuda-gpu-arch=sm_70
1+
// RUN: %clangxx -fsycl -fsycl-unnamed-lambda -fsycl-targets=%sycl_triple %s -o %t.out
22
// RUN: %HOST_RUN_PLACEHOLDER %t.out
33
// RUN: %GPU_RUN_PLACEHOLDER %t.out
44
// RUN: %CPU_RUN_PLACEHOLDER %t.out
55
// RUN: %ACC_RUN_PLACEHOLDER %t.out
66

7-
// CUDA and HIP backends have had no support for the generic address space yet.
7+
// HIP backend has no support for the generic address space yet.
88
// Host does not support barrier.
9-
// XFAIL: cuda || hip || host
9+
// XFAIL: hip || host
1010

1111
#define TEST_GENERIC_IN_LOCAL 1
1212

SYCL/AtomicRef/add_generic_local_native_fp.cpp

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -1,13 +1,13 @@
1-
// RUN: %clangxx -fsycl -fsycl-unnamed-lambda -fsycl-targets=%sycl_triple %s -o %t.out -Xsycl-target-backend=nvptx64-nvidia-cuda --cuda-gpu-arch=sm_70
1+
// RUN: %clangxx -fsycl -fsycl-unnamed-lambda -fsycl-targets=%sycl_triple %s -o %t.out
22
// RUN: %HOST_RUN_PLACEHOLDER %t.out
33
// RUN: %GPU_RUN_PLACEHOLDER %t.out
44
// RUN: %CPU_RUN_PLACEHOLDER %t.out
55
// RUN: %ACC_RUN_PLACEHOLDER %t.out
66

7-
// CUDA and HIP backends have had no support for the generic address space yet.
7+
// HIP backend has no support for the generic address space yet.
88
// Host does not support barrier. HIP does not support native floating point
99
// atomics
10-
// XFAIL: cuda, hip, host
10+
// XFAIL: hip, host
1111

1212
#define SYCL_USE_NATIVE_FP_ATOMICS
1313
#define FP_TESTS_ONLY

SYCL/AtomicRef/add_generic_native_fp.cpp

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -1,12 +1,12 @@
1-
// RUN: %clangxx -fsycl -fsycl-unnamed-lambda -fsycl-targets=%sycl_triple %s -o %t.out -Xsycl-target-backend=nvptx64-nvidia-cuda --cuda-gpu-arch=sm_70
1+
// RUN: %clangxx -fsycl -fsycl-unnamed-lambda -fsycl-targets=%sycl_triple %s -o %t.out
22
// RUN: %HOST_RUN_PLACEHOLDER %t.out
33
// RUN: %GPU_RUN_PLACEHOLDER %t.out
44
// RUN: %CPU_RUN_PLACEHOLDER %t.out
55
// RUN: %ACC_RUN_PLACEHOLDER %t.out
66

7-
// CUDA and HIP backends have had no support for the generic address space yet.
7+
// HIP backend has had no support for the generic address space yet.
88
// HIP does not support native floating point atomics
9-
// XFAIL: cuda, hip
9+
// XFAIL: hip
1010

1111
#define SYCL_USE_NATIVE_FP_ATOMICS
1212
#define FP_TESTS_ONLY

SYCL/AtomicRef/add_local.cpp

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -1,12 +1,12 @@
1-
// RUN: %clangxx -fsycl -fsycl-unnamed-lambda -fsycl-targets=%sycl_triple %s -o %t.out -Xsycl-target-backend=nvptx64-nvidia-cuda --cuda-gpu-arch=sm_70
1+
// RUN: %clangxx -fsycl -fsycl-unnamed-lambda -fsycl-targets=%sycl_triple %s -o %t.out
22
// RUN: %HOST_RUN_PLACEHOLDER %t.out
33
// RUN: %GPU_RUN_PLACEHOLDER %t.out
44
// RUN: %CPU_RUN_PLACEHOLDER %t.out
55
// RUN: %ACC_RUN_PLACEHOLDER %t.out
66

7-
// Barrier is not supported on host. HIP and ACC do not support floating
8-
// point atomics.
9-
// XFAIL: host, hip, acc
7+
// Barrier is not supported on host. HIP does not support floating point
8+
// atomics.
9+
// XFAIL: host, hip
1010

1111
#include "add.h"
1212

SYCL/AtomicRef/add_local_native_fp.cpp

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,4 @@
1-
// RUN: %clangxx -fsycl -fsycl-unnamed-lambda -fsycl-targets=%sycl_triple %s -o %t.out -Xsycl-target-backend=nvptx64-nvidia-cuda --cuda-gpu-arch=sm_70
1+
// RUN: %clangxx -fsycl -fsycl-unnamed-lambda -fsycl-targets=%sycl_triple %s -o %t.out
22
// RUN: %HOST_RUN_PLACEHOLDER %t.out
33
// RUN: %GPU_RUN_PLACEHOLDER %t.out
44
// RUN: %CPU_RUN_PLACEHOLDER %t.out

0 commit comments

Comments
 (0)