| 1 | //===-- tsan_test_util_posix.cpp ------------------------------------------===// |
| 2 | // |
| 3 | // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. |
| 4 | // See https://llvm.org/LICENSE.txt for license information. |
| 5 | // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception |
| 6 | // |
| 7 | //===----------------------------------------------------------------------===// |
| 8 | // |
| 9 | // This file is a part of ThreadSanitizer (TSan), a race detector. |
| 10 | // |
| 11 | // Test utils, Linux, FreeBSD, NetBSD and Darwin implementation. |
| 12 | //===----------------------------------------------------------------------===// |
| 13 | |
| 14 | #include "sanitizer_common/sanitizer_atomic.h" |
| 15 | #include "tsan_interface.h" |
| 16 | #include "tsan_posix_util.h" |
| 17 | #include "tsan_rtl.h" |
| 18 | #include "tsan_test_util.h" |
| 19 | #include "tsan_report.h" |
| 20 | |
| 21 | #include <assert.h> |
| 22 | #include <pthread.h> |
| 23 | #include <stdio.h> |
| 24 | #include <stdint.h> |
| 25 | #include <string.h> |
| 26 | #include <unistd.h> |
| 27 | #include <errno.h> |
| 28 | |
| 29 | #define CALLERPC (__builtin_return_address(0)) |
| 30 | |
| 31 | static __thread bool expect_report; |
| 32 | static __thread bool expect_report_reported; |
| 33 | static __thread __tsan::ReportType expect_report_type; |
| 34 | |
| 35 | void ThreadSanitizer::TearDown() { |
| 36 | __tsan::ctx->racy_stacks.Reset(); |
| 37 | } |
| 38 | |
| 39 | static void *BeforeInitThread(void *param) { |
| 40 | (void)param; |
| 41 | return 0; |
| 42 | } |
| 43 | |
| 44 | static void AtExit() { |
| 45 | } |
| 46 | |
| 47 | void TestMutexBeforeInit() { |
| 48 | // Mutexes must be usable before __tsan_init(); |
| 49 | pthread_mutex_t mtx = PTHREAD_MUTEX_INITIALIZER; |
| 50 | __interceptor_pthread_mutex_lock(mutex: &mtx); |
| 51 | __interceptor_pthread_mutex_unlock(mutex: &mtx); |
| 52 | __interceptor_pthread_mutex_destroy(mutex: &mtx); |
| 53 | pthread_t thr; |
| 54 | __interceptor_pthread_create(thread: &thr, attr: 0, start_routine: BeforeInitThread, arg: 0); |
| 55 | __interceptor_pthread_join(thread: thr, value_ptr: 0); |
| 56 | atexit(AtExit); |
| 57 | } |
| 58 | |
| 59 | namespace __tsan { |
| 60 | bool OnReport(const ReportDesc *rep, bool suppressed) { |
| 61 | if (expect_report) { |
| 62 | if (rep->typ != expect_report_type) { |
| 63 | printf(format: "Expected report of type %d, got type %d\n" , |
| 64 | (int)expect_report_type, (int)rep->typ); |
| 65 | EXPECT_TRUE(false) << "Wrong report type" ; |
| 66 | return false; |
| 67 | } |
| 68 | } else { |
| 69 | EXPECT_TRUE(false) << "Unexpected report" ; |
| 70 | return false; |
| 71 | } |
| 72 | expect_report_reported = true; |
| 73 | return true; |
| 74 | } |
| 75 | } // namespace __tsan |
| 76 | |
| 77 | static void* allocate_addr(int size, int offset_from_aligned = 0) { |
| 78 | static uintptr_t foo; |
| 79 | static __tsan::atomic_uintptr_t uniq = {.val_dont_use: (uintptr_t)&foo}; // Some real address. |
| 80 | const int kAlign = 16; |
| 81 | CHECK(offset_from_aligned < kAlign); |
| 82 | size = (size + 2 * kAlign) & ~(kAlign - 1); |
| 83 | uintptr_t addr = atomic_fetch_add(a: &uniq, v: size, mo: __tsan::memory_order_relaxed); |
| 84 | return (void*)(addr + offset_from_aligned); |
| 85 | } |
| 86 | |
| 87 | MemLoc::MemLoc(int offset_from_aligned) |
| 88 | : loc_(allocate_addr(size: 16, offset_from_aligned)) { |
| 89 | } |
| 90 | |
| 91 | MemLoc::~MemLoc() { |
| 92 | } |
| 93 | |
| 94 | UserMutex::UserMutex(Type type) : alive_(), type_(type) {} |
| 95 | |
| 96 | UserMutex::~UserMutex() { CHECK(!alive_); } |
| 97 | |
| 98 | void UserMutex::Init() { |
| 99 | CHECK(!alive_); |
| 100 | alive_ = true; |
| 101 | if (type_ == Normal) |
| 102 | CHECK_EQ(__interceptor_pthread_mutex_init((pthread_mutex_t*)mtx_, 0), 0); |
| 103 | #ifndef __APPLE__ |
| 104 | else if (type_ == Spin) |
| 105 | CHECK_EQ(pthread_spin_init((pthread_spinlock_t*)mtx_, 0), 0); |
| 106 | #endif |
| 107 | else if (type_ == RW) |
| 108 | CHECK_EQ(__interceptor_pthread_rwlock_init((pthread_rwlock_t*)mtx_, 0), 0); |
| 109 | else |
| 110 | CHECK(0); |
| 111 | } |
| 112 | |
| 113 | void UserMutex::StaticInit() { |
| 114 | CHECK(!alive_); |
| 115 | CHECK(type_ == Normal); |
| 116 | alive_ = true; |
| 117 | pthread_mutex_t tmp = PTHREAD_MUTEX_INITIALIZER; |
| 118 | memcpy(dest: mtx_, src: &tmp, n: sizeof(tmp)); |
| 119 | } |
| 120 | |
| 121 | void UserMutex::Destroy() { |
| 122 | CHECK(alive_); |
| 123 | alive_ = false; |
| 124 | if (type_ == Normal) |
| 125 | CHECK_EQ(__interceptor_pthread_mutex_destroy((pthread_mutex_t*)mtx_), 0); |
| 126 | #ifndef __APPLE__ |
| 127 | else if (type_ == Spin) |
| 128 | CHECK_EQ(pthread_spin_destroy((pthread_spinlock_t*)mtx_), 0); |
| 129 | #endif |
| 130 | else if (type_ == RW) |
| 131 | CHECK_EQ(__interceptor_pthread_rwlock_destroy((pthread_rwlock_t*)mtx_), 0); |
| 132 | } |
| 133 | |
| 134 | void UserMutex::Lock() { |
| 135 | CHECK(alive_); |
| 136 | if (type_ == Normal) |
| 137 | CHECK_EQ(__interceptor_pthread_mutex_lock((pthread_mutex_t*)mtx_), 0); |
| 138 | #ifndef __APPLE__ |
| 139 | else if (type_ == Spin) |
| 140 | CHECK_EQ(pthread_spin_lock((pthread_spinlock_t*)mtx_), 0); |
| 141 | #endif |
| 142 | else if (type_ == RW) |
| 143 | CHECK_EQ(__interceptor_pthread_rwlock_wrlock((pthread_rwlock_t*)mtx_), 0); |
| 144 | } |
| 145 | |
| 146 | bool UserMutex::TryLock() { |
| 147 | CHECK(alive_); |
| 148 | if (type_ == Normal) |
| 149 | return __interceptor_pthread_mutex_trylock(mutex: (pthread_mutex_t*)mtx_) == 0; |
| 150 | #ifndef __APPLE__ |
| 151 | else if (type_ == Spin) |
| 152 | return pthread_spin_trylock(lock: (pthread_spinlock_t*)mtx_) == 0; |
| 153 | #endif |
| 154 | else if (type_ == RW) |
| 155 | return __interceptor_pthread_rwlock_trywrlock(rwlock: (pthread_rwlock_t*)mtx_) == 0; |
| 156 | return false; |
| 157 | } |
| 158 | |
| 159 | void UserMutex::Unlock() { |
| 160 | CHECK(alive_); |
| 161 | if (type_ == Normal) |
| 162 | CHECK_EQ(__interceptor_pthread_mutex_unlock((pthread_mutex_t*)mtx_), 0); |
| 163 | #ifndef __APPLE__ |
| 164 | else if (type_ == Spin) |
| 165 | CHECK_EQ(pthread_spin_unlock((pthread_spinlock_t*)mtx_), 0); |
| 166 | #endif |
| 167 | else if (type_ == RW) |
| 168 | CHECK_EQ(__interceptor_pthread_rwlock_unlock((pthread_rwlock_t*)mtx_), 0); |
| 169 | } |
| 170 | |
| 171 | void UserMutex::ReadLock() { |
| 172 | CHECK(alive_); |
| 173 | CHECK(type_ == RW); |
| 174 | CHECK_EQ(__interceptor_pthread_rwlock_rdlock((pthread_rwlock_t*)mtx_), 0); |
| 175 | } |
| 176 | |
| 177 | bool UserMutex::TryReadLock() { |
| 178 | CHECK(alive_); |
| 179 | CHECK(type_ == RW); |
| 180 | return __interceptor_pthread_rwlock_tryrdlock(rwlock: (pthread_rwlock_t*)mtx_) == 0; |
| 181 | } |
| 182 | |
| 183 | void UserMutex::ReadUnlock() { |
| 184 | CHECK(alive_); |
| 185 | CHECK(type_ == RW); |
| 186 | CHECK_EQ(__interceptor_pthread_rwlock_unlock((pthread_rwlock_t*)mtx_), 0); |
| 187 | } |
| 188 | |
| 189 | struct Event { |
| 190 | enum Type { |
| 191 | SHUTDOWN, |
| 192 | READ, |
| 193 | WRITE, |
| 194 | VPTR_UPDATE, |
| 195 | CALL, |
| 196 | RETURN, |
| 197 | MUTEX_CREATE, |
| 198 | MUTEX_DESTROY, |
| 199 | MUTEX_LOCK, |
| 200 | MUTEX_TRYLOCK, |
| 201 | MUTEX_UNLOCK, |
| 202 | MUTEX_READLOCK, |
| 203 | MUTEX_TRYREADLOCK, |
| 204 | MUTEX_READUNLOCK, |
| 205 | MEMCPY, |
| 206 | MEMSET |
| 207 | }; |
| 208 | Type type; |
| 209 | void *ptr; |
| 210 | uptr arg; |
| 211 | uptr arg2; |
| 212 | bool res; |
| 213 | bool expect_report; |
| 214 | __tsan::ReportType report_type; |
| 215 | |
| 216 | explicit Event(Type type, const void *ptr = 0, uptr arg = 0, uptr arg2 = 0) |
| 217 | : type(type), |
| 218 | ptr(const_cast<void *>(ptr)), |
| 219 | arg(arg), |
| 220 | arg2(arg2), |
| 221 | res(), |
| 222 | expect_report(), |
| 223 | report_type() {} |
| 224 | |
| 225 | void ExpectReport(__tsan::ReportType type) { |
| 226 | expect_report = true; |
| 227 | report_type = type; |
| 228 | } |
| 229 | }; |
| 230 | |
| 231 | struct ScopedThread::Impl { |
| 232 | pthread_t thread; |
| 233 | bool main; |
| 234 | bool detached; |
| 235 | __tsan::atomic_uintptr_t event; // Event* |
| 236 | |
| 237 | static void *ScopedThreadCallback(void *arg); |
| 238 | void send(Event *ev); |
| 239 | void HandleEvent(Event *ev); |
| 240 | }; |
| 241 | |
| 242 | void ScopedThread::Impl::HandleEvent(Event *ev) { |
| 243 | CHECK_EQ(expect_report, false); |
| 244 | expect_report = ev->expect_report; |
| 245 | expect_report_reported = false; |
| 246 | expect_report_type = ev->report_type; |
| 247 | switch (ev->type) { |
| 248 | case Event::READ: |
| 249 | case Event::WRITE: { |
| 250 | void (*tsan_mop)(void *addr, void *pc) = 0; |
| 251 | if (ev->type == Event::READ) { |
| 252 | switch (ev->arg /*size*/) { |
| 253 | case 1: |
| 254 | tsan_mop = __tsan_read1_pc; |
| 255 | break; |
| 256 | case 2: |
| 257 | tsan_mop = __tsan_read2_pc; |
| 258 | break; |
| 259 | case 4: |
| 260 | tsan_mop = __tsan_read4_pc; |
| 261 | break; |
| 262 | case 8: |
| 263 | tsan_mop = __tsan_read8_pc; |
| 264 | break; |
| 265 | case 16: |
| 266 | tsan_mop = __tsan_read16_pc; |
| 267 | break; |
| 268 | } |
| 269 | } else { |
| 270 | switch (ev->arg /*size*/) { |
| 271 | case 1: |
| 272 | tsan_mop = __tsan_write1_pc; |
| 273 | break; |
| 274 | case 2: |
| 275 | tsan_mop = __tsan_write2_pc; |
| 276 | break; |
| 277 | case 4: |
| 278 | tsan_mop = __tsan_write4_pc; |
| 279 | break; |
| 280 | case 8: |
| 281 | tsan_mop = __tsan_write8_pc; |
| 282 | break; |
| 283 | case 16: |
| 284 | tsan_mop = __tsan_write16_pc; |
| 285 | break; |
| 286 | } |
| 287 | } |
| 288 | CHECK_NE(tsan_mop, 0); |
| 289 | #if defined(__FreeBSD__) || defined(__APPLE__) || defined(__NetBSD__) |
| 290 | const int ErrCode = ESOCKTNOSUPPORT; |
| 291 | #else |
| 292 | const int ErrCode = ECHRNG; |
| 293 | #endif |
| 294 | errno = ErrCode; |
| 295 | tsan_mop(ev->ptr, (void *)ev->arg2); |
| 296 | CHECK_EQ(ErrCode, errno); // In no case must errno be changed. |
| 297 | break; |
| 298 | } |
| 299 | case Event::VPTR_UPDATE: |
| 300 | __tsan_vptr_update(vptr_p: (void**)ev->ptr, new_val: (void*)ev->arg); |
| 301 | break; |
| 302 | case Event::CALL: |
| 303 | __tsan_func_entry(call_pc: (void*)((uptr)ev->ptr)); |
| 304 | break; |
| 305 | case Event::RETURN: |
| 306 | __tsan_func_exit(); |
| 307 | break; |
| 308 | case Event::MUTEX_CREATE: |
| 309 | static_cast<UserMutex *>(ev->ptr)->Init(); |
| 310 | break; |
| 311 | case Event::MUTEX_DESTROY: |
| 312 | static_cast<UserMutex *>(ev->ptr)->Destroy(); |
| 313 | break; |
| 314 | case Event::MUTEX_LOCK: |
| 315 | static_cast<UserMutex *>(ev->ptr)->Lock(); |
| 316 | break; |
| 317 | case Event::MUTEX_TRYLOCK: |
| 318 | ev->res = static_cast<UserMutex *>(ev->ptr)->TryLock(); |
| 319 | break; |
| 320 | case Event::MUTEX_UNLOCK: |
| 321 | static_cast<UserMutex *>(ev->ptr)->Unlock(); |
| 322 | break; |
| 323 | case Event::MUTEX_READLOCK: |
| 324 | static_cast<UserMutex *>(ev->ptr)->ReadLock(); |
| 325 | break; |
| 326 | case Event::MUTEX_TRYREADLOCK: |
| 327 | ev->res = static_cast<UserMutex *>(ev->ptr)->TryReadLock(); |
| 328 | break; |
| 329 | case Event::MUTEX_READUNLOCK: |
| 330 | static_cast<UserMutex *>(ev->ptr)->ReadUnlock(); |
| 331 | break; |
| 332 | case Event::MEMCPY: |
| 333 | __interceptor_memcpy(ev->ptr, (void*)ev->arg, ev->arg2); |
| 334 | break; |
| 335 | case Event::MEMSET: |
| 336 | __interceptor_memset(ev->ptr, ev->arg, ev->arg2); |
| 337 | break; |
| 338 | default: CHECK(0); |
| 339 | } |
| 340 | if (expect_report && !expect_report_reported) { |
| 341 | printf(format: "Missed expected report of type %d\n" , (int)ev->report_type); |
| 342 | EXPECT_TRUE(false) << "Missed expected race" ; |
| 343 | } |
| 344 | expect_report = false; |
| 345 | } |
| 346 | |
| 347 | void *ScopedThread::Impl::ScopedThreadCallback(void *arg) { |
| 348 | __tsan_func_entry(CALLERPC); |
| 349 | Impl *impl = (Impl*)arg; |
| 350 | for (;;) { |
| 351 | Event *ev = |
| 352 | (Event *)atomic_load(a: &impl->event, mo: __tsan::memory_order_acquire); |
| 353 | if (ev == 0) { |
| 354 | sched_yield(); |
| 355 | continue; |
| 356 | } |
| 357 | if (ev->type == Event::SHUTDOWN) { |
| 358 | atomic_store(a: &impl->event, v: 0, mo: __tsan::memory_order_release); |
| 359 | break; |
| 360 | } |
| 361 | impl->HandleEvent(ev); |
| 362 | atomic_store(a: &impl->event, v: 0, mo: __tsan::memory_order_release); |
| 363 | } |
| 364 | __tsan_func_exit(); |
| 365 | return 0; |
| 366 | } |
| 367 | |
| 368 | void ScopedThread::Impl::send(Event *e) { |
| 369 | if (main) { |
| 370 | HandleEvent(ev: e); |
| 371 | } else { |
| 372 | CHECK_EQ(atomic_load(&event, __tsan::memory_order_relaxed), 0); |
| 373 | atomic_store(a: &event, v: (uintptr_t)e, mo: __tsan::memory_order_release); |
| 374 | while (atomic_load(a: &event, mo: __tsan::memory_order_acquire) != 0) |
| 375 | sched_yield(); |
| 376 | } |
| 377 | } |
| 378 | |
| 379 | ScopedThread::ScopedThread(bool detached, bool main) { |
| 380 | impl_ = new Impl; |
| 381 | impl_->main = main; |
| 382 | impl_->detached = detached; |
| 383 | atomic_store(a: &impl_->event, v: 0, mo: __tsan::memory_order_relaxed); |
| 384 | if (!main) { |
| 385 | pthread_attr_t attr; |
| 386 | pthread_attr_init(attr: &attr); |
| 387 | pthread_attr_setdetachstate( |
| 388 | attr: &attr, detachstate: detached ? PTHREAD_CREATE_DETACHED : PTHREAD_CREATE_JOINABLE); |
| 389 | pthread_attr_setstacksize(attr: &attr, stacksize: 64*1024); |
| 390 | __interceptor_pthread_create(thread: &impl_->thread, attr: &attr, |
| 391 | start_routine: ScopedThread::Impl::ScopedThreadCallback, arg: impl_); |
| 392 | } |
| 393 | } |
| 394 | |
| 395 | ScopedThread::~ScopedThread() { |
| 396 | if (!impl_->main) { |
| 397 | Event event(Event::SHUTDOWN); |
| 398 | impl_->send(e: &event); |
| 399 | if (!impl_->detached) |
| 400 | __interceptor_pthread_join(thread: impl_->thread, value_ptr: 0); |
| 401 | } |
| 402 | delete impl_; |
| 403 | } |
| 404 | |
| 405 | void ScopedThread::Detach() { |
| 406 | CHECK(!impl_->main); |
| 407 | CHECK(!impl_->detached); |
| 408 | impl_->detached = true; |
| 409 | __interceptor_pthread_detach(thread: impl_->thread); |
| 410 | } |
| 411 | |
| 412 | void ScopedThread::Access(void *addr, bool is_write, |
| 413 | int size, bool expect_race) { |
| 414 | Event event(is_write ? Event::WRITE : Event::READ, addr, size, |
| 415 | (uptr)CALLERPC); |
| 416 | if (expect_race) |
| 417 | event.ExpectReport(type: __tsan::ReportTypeRace); |
| 418 | impl_->send(e: &event); |
| 419 | } |
| 420 | |
| 421 | void ScopedThread::VptrUpdate(const MemLoc &vptr, |
| 422 | const MemLoc &new_val, |
| 423 | bool expect_race) { |
| 424 | Event event(Event::VPTR_UPDATE, vptr.loc(), (uptr)new_val.loc()); |
| 425 | if (expect_race) |
| 426 | event.ExpectReport(type: __tsan::ReportTypeRace); |
| 427 | impl_->send(e: &event); |
| 428 | } |
| 429 | |
| 430 | void ScopedThread::Call(void(*pc)()) { |
| 431 | Event event(Event::CALL, (void*)((uintptr_t)pc)); |
| 432 | impl_->send(e: &event); |
| 433 | } |
| 434 | |
| 435 | void ScopedThread::Return() { |
| 436 | Event event(Event::RETURN); |
| 437 | impl_->send(e: &event); |
| 438 | } |
| 439 | |
| 440 | void ScopedThread::Create(const UserMutex &m) { |
| 441 | Event event(Event::MUTEX_CREATE, &m); |
| 442 | impl_->send(e: &event); |
| 443 | } |
| 444 | |
| 445 | void ScopedThread::Destroy(const UserMutex &m) { |
| 446 | Event event(Event::MUTEX_DESTROY, &m); |
| 447 | impl_->send(e: &event); |
| 448 | } |
| 449 | |
| 450 | void ScopedThread::Lock(const UserMutex &m) { |
| 451 | Event event(Event::MUTEX_LOCK, &m); |
| 452 | impl_->send(e: &event); |
| 453 | } |
| 454 | |
| 455 | bool ScopedThread::TryLock(const UserMutex &m) { |
| 456 | Event event(Event::MUTEX_TRYLOCK, &m); |
| 457 | impl_->send(e: &event); |
| 458 | return event.res; |
| 459 | } |
| 460 | |
| 461 | void ScopedThread::Unlock(const UserMutex &m) { |
| 462 | Event event(Event::MUTEX_UNLOCK, &m); |
| 463 | impl_->send(e: &event); |
| 464 | } |
| 465 | |
| 466 | void ScopedThread::ReadLock(const UserMutex &m) { |
| 467 | Event event(Event::MUTEX_READLOCK, &m); |
| 468 | impl_->send(e: &event); |
| 469 | } |
| 470 | |
| 471 | bool ScopedThread::TryReadLock(const UserMutex &m) { |
| 472 | Event event(Event::MUTEX_TRYREADLOCK, &m); |
| 473 | impl_->send(e: &event); |
| 474 | return event.res; |
| 475 | } |
| 476 | |
| 477 | void ScopedThread::ReadUnlock(const UserMutex &m) { |
| 478 | Event event(Event::MUTEX_READUNLOCK, &m); |
| 479 | impl_->send(e: &event); |
| 480 | } |
| 481 | |
| 482 | void ScopedThread::Memcpy(void *dst, const void *src, int size, |
| 483 | bool expect_race) { |
| 484 | Event event(Event::MEMCPY, dst, (uptr)src, size); |
| 485 | if (expect_race) |
| 486 | event.ExpectReport(type: __tsan::ReportTypeRace); |
| 487 | impl_->send(e: &event); |
| 488 | } |
| 489 | |
| 490 | void ScopedThread::Memset(void *dst, int val, int size, |
| 491 | bool expect_race) { |
| 492 | Event event(Event::MEMSET, dst, val, size); |
| 493 | if (expect_race) |
| 494 | event.ExpectReport(type: __tsan::ReportTypeRace); |
| 495 | impl_->send(e: &event); |
| 496 | } |
| 497 | |