1//===-- tsan_dense_alloc.h --------------------------------------*- C++ -*-===//
2//
3// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
4// See https://llvm.org/LICENSE.txt for license information.
5// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
6//
7//===----------------------------------------------------------------------===//
8//
9// This file is a part of ThreadSanitizer (TSan), a race detector.
10//
11// A DenseSlabAlloc is a freelist-based allocator of fixed-size objects.
12// DenseSlabAllocCache is a thread-local cache for DenseSlabAlloc.
13// The only difference with traditional slab allocators is that DenseSlabAlloc
14// allocates/free indices of objects and provide a functionality to map
15// the index onto the real pointer. The index is u32, that is, 2 times smaller
16// than uptr (hense the Dense prefix).
17//===----------------------------------------------------------------------===//
18#ifndef TSAN_DENSE_ALLOC_H
19#define TSAN_DENSE_ALLOC_H
20
21#include "sanitizer_common/sanitizer_common.h"
22#include "tsan_defs.h"
23
24namespace __tsan {
25
26class DenseSlabAllocCache {
27 static const uptr kSize = 128;
28 typedef u32 IndexT;
29 uptr pos;
30 IndexT cache[kSize];
31 template <typename, uptr, uptr, u64>
32 friend class DenseSlabAlloc;
33};
34
35template <typename T, uptr kL1Size, uptr kL2Size, u64 kReserved = 0>
36class DenseSlabAlloc {
37 public:
38 typedef DenseSlabAllocCache Cache;
39 typedef typename Cache::IndexT IndexT;
40
41 static_assert((kL1Size & (kL1Size - 1)) == 0,
42 "kL1Size must be a power-of-two");
43 static_assert((kL2Size & (kL2Size - 1)) == 0,
44 "kL2Size must be a power-of-two");
45 static_assert((kL1Size * kL2Size) <= (1ull << (sizeof(IndexT) * 8)),
46 "kL1Size/kL2Size are too large");
47 static_assert(((kL1Size * kL2Size - 1) & kReserved) == 0,
48 "reserved bits don't fit");
49 static_assert(sizeof(T) > sizeof(IndexT),
50 "it doesn't make sense to use dense alloc");
51
52 DenseSlabAlloc(LinkerInitialized, const char *name) : name_(name) {}
53
54 explicit DenseSlabAlloc(const char *name)
55 : DenseSlabAlloc(LINKER_INITIALIZED, name) {
56 // It can be very large.
57 // Don't page it in for linker initialized objects.
58 internal_memset(map_, 0, sizeof(map_));
59 }
60
61 ~DenseSlabAlloc() {
62 for (uptr i = 0; i < kL1Size; i++) {
63 if (map_[i] != 0)
64 UnmapOrDie(map_[i], kL2Size * sizeof(T));
65 }
66 }
67
68 IndexT Alloc(Cache *c) {
69 if (c->pos == 0)
70 Refill(c);
71 return c->cache[--c->pos];
72 }
73
74 void Free(Cache *c, IndexT idx) {
75 DCHECK_NE(idx, 0);
76 if (c->pos == Cache::kSize)
77 Drain(c);
78 c->cache[c->pos++] = idx;
79 }
80
81 T *Map(IndexT idx) {
82 DCHECK_NE(idx, 0);
83 DCHECK_LE(idx, kL1Size * kL2Size);
84 return &map_[idx / kL2Size][idx % kL2Size];
85 }
86
87 void FlushCache(Cache *c) {
88 while (c->pos) Drain(c);
89 }
90
91 void InitCache(Cache *c) {
92 c->pos = 0;
93 internal_memset(c->cache, 0, sizeof(c->cache));
94 }
95
96 uptr AllocatedMemory() const {
97 return atomic_load_relaxed(&fillpos_) * kL2Size * sizeof(T);
98 }
99
100 template <typename Func>
101 void ForEach(Func func) {
102 Lock lock(&mtx_);
103 uptr fillpos = atomic_load_relaxed(&fillpos_);
104 for (uptr l1 = 0; l1 < fillpos; l1++) {
105 for (IndexT l2 = l1 == 0 ? 1 : 0; l2 < kL2Size; l2++) func(&map_[l1][l2]);
106 }
107 }
108
109 private:
110 T *map_[kL1Size];
111 Mutex mtx_;
112 // The freelist is organized as a lock-free stack of batches of nodes.
113 // The stack itself uses Block::next links, while the batch within each
114 // stack node uses Block::batch links.
115 // Low 32-bits of freelist_ is the node index, top 32-bits is ABA-counter.
116 atomic_uint64_t freelist_ = {0};
117 atomic_uintptr_t fillpos_ = {0};
118 const char *const name_;
119
120 struct Block {
121 IndexT next;
122 IndexT batch;
123 };
124
125 Block *MapBlock(IndexT idx) { return reinterpret_cast<Block *>(Map(idx)); }
126
127 static constexpr u64 kCounterInc = 1ull << 32;
128 static constexpr u64 kCounterMask = ~(kCounterInc - 1);
129
130 NOINLINE void Refill(Cache *c) {
131 // Pop 1 batch of nodes from the freelist.
132 IndexT idx;
133 u64 xchg;
134 u64 cmp = atomic_load(&freelist_, memory_order_acquire);
135 do {
136 idx = static_cast<IndexT>(cmp);
137 if (!idx)
138 return AllocSuperBlock(c);
139 Block *ptr = MapBlock(idx);
140 xchg = ptr->next | (cmp & kCounterMask);
141 } while (!atomic_compare_exchange_weak(&freelist_, &cmp, xchg,
142 memory_order_acq_rel));
143 // Unpack it into c->cache.
144 while (idx) {
145 c->cache[c->pos++] = idx;
146 idx = MapBlock(idx)->batch;
147 }
148 }
149
150 NOINLINE void Drain(Cache *c) {
151 // Build a batch of at most Cache::kSize / 2 nodes linked by Block::batch.
152 IndexT head_idx = 0;
153 for (uptr i = 0; i < Cache::kSize / 2 && c->pos; i++) {
154 IndexT idx = c->cache[--c->pos];
155 Block *ptr = MapBlock(idx);
156 ptr->batch = head_idx;
157 head_idx = idx;
158 }
159 // Push it onto the freelist stack.
160 Block *head = MapBlock(head_idx);
161 u64 xchg;
162 u64 cmp = atomic_load(&freelist_, memory_order_acquire);
163 do {
164 head->next = static_cast<IndexT>(cmp);
165 xchg = head_idx | (cmp & kCounterMask) + kCounterInc;
166 } while (!atomic_compare_exchange_weak(&freelist_, &cmp, xchg,
167 memory_order_acq_rel));
168 }
169
170 NOINLINE void AllocSuperBlock(Cache *c) {
171 Lock lock(&mtx_);
172 uptr fillpos = atomic_load_relaxed(&fillpos_);
173 if (fillpos == kL1Size) {
174 Printf("ThreadSanitizer: %s overflow (%zu*%zu). Dying.\n", name_, kL1Size,
175 kL2Size);
176 Die();
177 }
178 VPrintf(2, "ThreadSanitizer: growing %s: %zu out of %zu*%zu\n", name_,
179 fillpos, kL1Size, kL2Size);
180 T *batch = (T *)MmapOrDie(kL2Size * sizeof(T), name_);
181 map_[fillpos] = batch;
182 // Reserve 0 as invalid index.
183 for (IndexT i = fillpos ? 0 : 1; i < kL2Size; i++) {
184 new (batch + i) T;
185 c->cache[c->pos++] = i + fillpos * kL2Size;
186 if (c->pos == Cache::kSize)
187 Drain(c);
188 }
189 atomic_store_relaxed(&fillpos_, fillpos + 1);
190 CHECK(c->pos);
191 }
192};
193
194} // namespace __tsan
195
196#endif // TSAN_DENSE_ALLOC_H
197

source code of compiler-rt/lib/tsan/rtl/tsan_dense_alloc.h