| 1 | // RUN: %clangxx_tsan %s -o %t |
| 2 | // RUN: %run %t 2>&1 | FileCheck %s |
| 3 | |
| 4 | // bench.h needs pthread barriers which are not available on OS X |
| 5 | // UNSUPPORTED: darwin |
| 6 | |
| 7 | #include "bench.h" |
| 8 | #include <memory.h> |
| 9 | |
| 10 | void thread(int tid) { |
| 11 | volatile long x = 0; |
| 12 | switch (bench_mode) { |
| 13 | case 0: |
| 14 | for (int i = 0; i < bench_niter; i++) |
| 15 | *(volatile char *)&x = 1; |
| 16 | break; |
| 17 | case 1: |
| 18 | for (int i = 0; i < bench_niter; i++) |
| 19 | *(volatile short *)&x = 1; |
| 20 | break; |
| 21 | case 2: |
| 22 | for (int i = 0; i < bench_niter; i++) |
| 23 | *(volatile int *)&x = 1; |
| 24 | break; |
| 25 | case 3: |
| 26 | for (int i = 0; i < bench_niter; i++) |
| 27 | *(volatile long *)&x = 1; |
| 28 | break; |
| 29 | case 4: |
| 30 | for (int i = 0; i < bench_niter; i++) |
| 31 | *(volatile char *)&x; |
| 32 | break; |
| 33 | case 5: |
| 34 | for (int i = 0; i < bench_niter; i++) |
| 35 | *(volatile short *)&x; |
| 36 | break; |
| 37 | case 6: |
| 38 | for (int i = 0; i < bench_niter; i++) |
| 39 | *(volatile int *)&x; |
| 40 | break; |
| 41 | case 7: |
| 42 | for (int i = 0; i < bench_niter; i++) |
| 43 | *(volatile long *)&x; |
| 44 | case 8: |
| 45 | for (int i = 0; i < bench_niter / 10; i++) { |
| 46 | ((volatile long *)&x)[0]; |
| 47 | ((volatile int *)&x)[0]; |
| 48 | ((volatile short *)&x)[2]; |
| 49 | ((volatile char *)&x)[6]; |
| 50 | ((volatile char *)&x)[7]; |
| 51 | ((volatile long *)&x)[0] = 1; |
| 52 | ((volatile int *)&x)[0] = 1; |
| 53 | ((volatile short *)&x)[2] = 1; |
| 54 | ((volatile char *)&x)[6] = 1; |
| 55 | ((volatile char *)&x)[7] = 1; |
| 56 | } |
| 57 | break; |
| 58 | case 9: { |
| 59 | volatile long size = sizeof(x); |
| 60 | for (int i = 0; i < bench_niter; i++) |
| 61 | memset(s: (void *)&x, c: i, n: size); |
| 62 | break; |
| 63 | } |
| 64 | case 10: { |
| 65 | volatile long data[2] = {}; |
| 66 | volatile long size = sizeof(data) - 2; |
| 67 | for (int i = 0; i < bench_niter; i++) |
| 68 | memset(s: ((char *)data) + 1, c: i, n: size); |
| 69 | break; |
| 70 | } |
| 71 | case 11: { |
| 72 | volatile long data[2] = {}; |
| 73 | for (int i = 0; i < bench_niter / 8 / 3; i++) { |
| 74 | for (int off = 0; off < 8; off++) { |
| 75 | __sanitizer_unaligned_store16(p: ((char *)data) + off, x: i); |
| 76 | __sanitizer_unaligned_store32(p: ((char *)data) + off, x: i); |
| 77 | __sanitizer_unaligned_store64(p: ((char *)data) + off, x: i); |
| 78 | } |
| 79 | } |
| 80 | break; |
| 81 | } |
| 82 | #if TSAN_VECTORIZE |
| 83 | case 12: { |
| 84 | // The compiler wants to optimize all this away. |
| 85 | // Use volatile to prevent optimization, but then use kBlock |
| 86 | // to avoid the additional non-vector load in the inner loop. |
| 87 | // Also use only even indexes to prevent compiler from |
| 88 | // inserting memset. |
| 89 | const int kBlock = 128; |
| 90 | __m128i data[kBlock * 2]; |
| 91 | __m128i *volatile vptr = data; |
| 92 | for (int i = 0; i < bench_niter / kBlock; i++) { |
| 93 | __m128i *ptr = vptr; |
| 94 | for (int j = 0; j < kBlock; j++) |
| 95 | _mm_store_si128(&ptr[j * 2], _mm_setzero_si128()); |
| 96 | } |
| 97 | break; |
| 98 | } |
| 99 | #endif |
| 100 | } |
| 101 | } |
| 102 | |
| 103 | void bench() { |
| 104 | start_thread_group(nth: bench_nthread, f: thread); |
| 105 | } |
| 106 | |
| 107 | // CHECK: DONE |
| 108 | |