1 | //===-- fuchsia.cpp ---------------------------------------------*- C++ -*-===// |
2 | // |
3 | // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. |
4 | // See https://llvm.org/LICENSE.txt for license information. |
5 | // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception |
6 | // |
7 | //===----------------------------------------------------------------------===// |
8 | |
9 | #include "platform.h" |
10 | |
11 | #if SCUDO_FUCHSIA |
12 | |
13 | #include "common.h" |
14 | #include "mutex.h" |
15 | #include "string_utils.h" |
16 | |
17 | #include <lib/sync/mutex.h> // for sync_mutex_t |
18 | #include <stdlib.h> // for getenv() |
19 | #include <zircon/compiler.h> |
20 | #include <zircon/process.h> |
21 | #include <zircon/sanitizer.h> |
22 | #include <zircon/status.h> |
23 | #include <zircon/syscalls.h> |
24 | |
25 | namespace scudo { |
26 | |
27 | uptr getPageSize() { return _zx_system_get_page_size(); } |
28 | |
29 | void NORETURN die() { __builtin_trap(); } |
30 | |
31 | // We zero-initialize the Extra parameter of map(), make sure this is consistent |
32 | // with ZX_HANDLE_INVALID. |
33 | static_assert(ZX_HANDLE_INVALID == 0, "" ); |
34 | |
35 | static void NORETURN dieOnError(zx_status_t Status, const char *FnName, |
36 | uptr Size) { |
37 | ScopedString Error; |
38 | Error.append("SCUDO ERROR: %s failed with size %zuKB (%s)" , FnName, |
39 | Size >> 10, zx_status_get_string(Status)); |
40 | outputRaw(Error.data()); |
41 | die(); |
42 | } |
43 | |
44 | static void *allocateVmar(uptr Size, MapPlatformData *Data, bool AllowNoMem) { |
45 | // Only scenario so far. |
46 | DCHECK(Data); |
47 | DCHECK_EQ(Data->Vmar, ZX_HANDLE_INVALID); |
48 | |
49 | const zx_status_t Status = _zx_vmar_allocate( |
50 | _zx_vmar_root_self(), |
51 | ZX_VM_CAN_MAP_READ | ZX_VM_CAN_MAP_WRITE | ZX_VM_CAN_MAP_SPECIFIC, 0, |
52 | Size, &Data->Vmar, &Data->VmarBase); |
53 | if (UNLIKELY(Status != ZX_OK)) { |
54 | if (Status != ZX_ERR_NO_MEMORY || !AllowNoMem) |
55 | dieOnError(Status, "zx_vmar_allocate" , Size); |
56 | return nullptr; |
57 | } |
58 | return reinterpret_cast<void *>(Data->VmarBase); |
59 | } |
60 | |
61 | void *map(void *Addr, uptr Size, const char *Name, uptr Flags, |
62 | MapPlatformData *Data) { |
63 | DCHECK_EQ(Size % getPageSizeCached(), 0); |
64 | const bool AllowNoMem = !!(Flags & MAP_ALLOWNOMEM); |
65 | |
66 | // For MAP_NOACCESS, just allocate a Vmar and return. |
67 | if (Flags & MAP_NOACCESS) |
68 | return allocateVmar(Size, Data, AllowNoMem); |
69 | |
70 | const zx_handle_t Vmar = (Data && Data->Vmar != ZX_HANDLE_INVALID) |
71 | ? Data->Vmar |
72 | : _zx_vmar_root_self(); |
73 | |
74 | zx_status_t Status; |
75 | zx_handle_t Vmo; |
76 | uint64_t VmoSize = 0; |
77 | if (Data && Data->Vmo != ZX_HANDLE_INVALID) { |
78 | // If a Vmo was specified, it's a resize operation. |
79 | CHECK(Addr); |
80 | DCHECK(Flags & MAP_RESIZABLE); |
81 | Vmo = Data->Vmo; |
82 | VmoSize = Data->VmoSize; |
83 | Status = _zx_vmo_set_size(Vmo, VmoSize + Size); |
84 | if (Status != ZX_OK) { |
85 | if (Status != ZX_ERR_NO_MEMORY || !AllowNoMem) |
86 | dieOnError(Status, "zx_vmo_set_size" , VmoSize + Size); |
87 | return nullptr; |
88 | } |
89 | } else { |
90 | // Otherwise, create a Vmo and set its name. |
91 | Status = _zx_vmo_create(Size, ZX_VMO_RESIZABLE, &Vmo); |
92 | if (UNLIKELY(Status != ZX_OK)) { |
93 | if (Status != ZX_ERR_NO_MEMORY || !AllowNoMem) |
94 | dieOnError(Status, "zx_vmo_create" , Size); |
95 | return nullptr; |
96 | } |
97 | _zx_object_set_property(Vmo, ZX_PROP_NAME, Name, strlen(Name)); |
98 | } |
99 | |
100 | uintptr_t P; |
101 | zx_vm_option_t MapFlags = |
102 | ZX_VM_PERM_READ | ZX_VM_PERM_WRITE | ZX_VM_ALLOW_FAULTS; |
103 | if (Addr) |
104 | DCHECK(Data); |
105 | const uint64_t Offset = |
106 | Addr ? reinterpret_cast<uintptr_t>(Addr) - Data->VmarBase : 0; |
107 | if (Offset) |
108 | MapFlags |= ZX_VM_SPECIFIC; |
109 | Status = _zx_vmar_map(Vmar, MapFlags, Offset, Vmo, VmoSize, Size, &P); |
110 | if (UNLIKELY(Status != ZX_OK)) { |
111 | if (Status != ZX_ERR_NO_MEMORY || !AllowNoMem) |
112 | dieOnError(Status, "zx_vmar_map" , Size); |
113 | return nullptr; |
114 | } |
115 | |
116 | if (Flags & MAP_PRECOMMIT) { |
117 | Status = _zx_vmar_op_range(Vmar, ZX_VMAR_OP_COMMIT, P, Size, |
118 | /*buffer=*/nullptr, /*buffer_size=*/0); |
119 | } |
120 | |
121 | // No need to track the Vmo if we don't intend on resizing it. Close it. |
122 | if (Flags & MAP_RESIZABLE) { |
123 | DCHECK(Data); |
124 | if (Data->Vmo == ZX_HANDLE_INVALID) |
125 | Data->Vmo = Vmo; |
126 | else |
127 | DCHECK_EQ(Data->Vmo, Vmo); |
128 | } else { |
129 | CHECK_EQ(_zx_handle_close(Vmo), ZX_OK); |
130 | } |
131 | if (UNLIKELY(Status != ZX_OK)) { |
132 | if (Status != ZX_ERR_NO_MEMORY || !AllowNoMem) |
133 | dieOnError(Status, "zx_vmar_op_range" , Size); |
134 | return nullptr; |
135 | } |
136 | |
137 | if (Data) |
138 | Data->VmoSize += Size; |
139 | |
140 | return reinterpret_cast<void *>(P); |
141 | } |
142 | |
143 | void unmap(void *Addr, uptr Size, uptr Flags, MapPlatformData *Data) { |
144 | if (Flags & UNMAP_ALL) { |
145 | DCHECK_NE(Data, nullptr); |
146 | const zx_handle_t Vmar = Data->Vmar; |
147 | DCHECK_NE(Vmar, _zx_vmar_root_self()); |
148 | // Destroying the vmar effectively unmaps the whole mapping. |
149 | CHECK_EQ(_zx_vmar_destroy(Vmar), ZX_OK); |
150 | CHECK_EQ(_zx_handle_close(Vmar), ZX_OK); |
151 | } else { |
152 | const zx_handle_t Vmar = (Data && Data->Vmar != ZX_HANDLE_INVALID) |
153 | ? Data->Vmar |
154 | : _zx_vmar_root_self(); |
155 | const zx_status_t Status = |
156 | _zx_vmar_unmap(Vmar, reinterpret_cast<uintptr_t>(Addr), Size); |
157 | if (UNLIKELY(Status != ZX_OK)) |
158 | dieOnError(Status, "zx_vmar_unmap" , Size); |
159 | } |
160 | if (Data) { |
161 | if (Data->Vmo != ZX_HANDLE_INVALID) |
162 | CHECK_EQ(_zx_handle_close(Data->Vmo), ZX_OK); |
163 | memset(Data, 0, sizeof(*Data)); |
164 | } |
165 | } |
166 | |
167 | void setMemoryPermission(UNUSED uptr Addr, UNUSED uptr Size, UNUSED uptr Flags, |
168 | UNUSED MapPlatformData *Data) { |
169 | const zx_vm_option_t Prot = |
170 | (Flags & MAP_NOACCESS) ? 0 : (ZX_VM_PERM_READ | ZX_VM_PERM_WRITE); |
171 | DCHECK(Data); |
172 | DCHECK_NE(Data->Vmar, ZX_HANDLE_INVALID); |
173 | const zx_status_t Status = _zx_vmar_protect(Data->Vmar, Prot, Addr, Size); |
174 | if (Status != ZX_OK) |
175 | dieOnError(Status, "zx_vmar_protect" , Size); |
176 | } |
177 | |
178 | void releasePagesToOS(UNUSED uptr BaseAddress, uptr Offset, uptr Size, |
179 | MapPlatformData *Data) { |
180 | // TODO: DCHECK the BaseAddress is consistent with the data in |
181 | // MapPlatformData. |
182 | DCHECK(Data); |
183 | DCHECK_NE(Data->Vmar, ZX_HANDLE_INVALID); |
184 | DCHECK_NE(Data->Vmo, ZX_HANDLE_INVALID); |
185 | const zx_status_t Status = |
186 | _zx_vmo_op_range(Data->Vmo, ZX_VMO_OP_DECOMMIT, Offset, Size, NULL, 0); |
187 | CHECK_EQ(Status, ZX_OK); |
188 | } |
189 | |
190 | const char *getEnv(const char *Name) { return getenv(Name); } |
191 | |
192 | // Note: we need to flag these methods with __TA_NO_THREAD_SAFETY_ANALYSIS |
193 | // because the Fuchsia implementation of sync_mutex_t has clang thread safety |
194 | // annotations. Were we to apply proper capability annotations to the top level |
195 | // HybridMutex class itself, they would not be needed. As it stands, the |
196 | // thread analysis thinks that we are locking the mutex and accidentally leaving |
197 | // it locked on the way out. |
198 | bool HybridMutex::tryLock() __TA_NO_THREAD_SAFETY_ANALYSIS { |
199 | // Size and alignment must be compatible between both types. |
200 | return sync_mutex_trylock(&M) == ZX_OK; |
201 | } |
202 | |
203 | void HybridMutex::lockSlow() __TA_NO_THREAD_SAFETY_ANALYSIS { |
204 | sync_mutex_lock(&M); |
205 | } |
206 | |
207 | void HybridMutex::unlock() __TA_NO_THREAD_SAFETY_ANALYSIS { |
208 | sync_mutex_unlock(&M); |
209 | } |
210 | |
211 | void HybridMutex::assertHeldImpl() __TA_NO_THREAD_SAFETY_ANALYSIS {} |
212 | |
213 | u64 getMonotonicTime() { return _zx_clock_get_monotonic(); } |
214 | u64 getMonotonicTimeFast() { return _zx_clock_get_monotonic(); } |
215 | |
216 | u32 getNumberOfCPUs() { return _zx_system_get_num_cpus(); } |
217 | |
218 | u32 getThreadID() { return 0; } |
219 | |
220 | bool getRandom(void *Buffer, uptr Length, UNUSED bool Blocking) { |
221 | static_assert(MaxRandomLength <= ZX_CPRNG_DRAW_MAX_LEN, "" ); |
222 | if (UNLIKELY(!Buffer || !Length || Length > MaxRandomLength)) |
223 | return false; |
224 | _zx_cprng_draw(Buffer, Length); |
225 | return true; |
226 | } |
227 | |
228 | void outputRaw(const char *Buffer) { |
229 | __sanitizer_log_write(Buffer, strlen(Buffer)); |
230 | } |
231 | |
232 | void setAbortMessage(const char *Message) {} |
233 | |
234 | } // namespace scudo |
235 | |
236 | #endif // SCUDO_FUCHSIA |
237 | |