1 | /* |
2 | * ompt-specific.h - header of OMPT internal functions implementation |
3 | */ |
4 | |
5 | //===----------------------------------------------------------------------===// |
6 | // |
7 | // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. |
8 | // See https://llvm.org/LICENSE.txt for license information. |
9 | // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception |
10 | // |
11 | //===----------------------------------------------------------------------===// |
12 | |
13 | #ifndef OMPT_SPECIFIC_H |
14 | #define OMPT_SPECIFIC_H |
15 | |
16 | #include "kmp.h" |
17 | |
18 | #if OMPT_SUPPORT |
19 | /***************************************************************************** |
20 | * forward declarations |
21 | ****************************************************************************/ |
22 | |
23 | /// Entrypoint used by libomptarget to register callbacks in libomp, if not |
24 | /// done already |
25 | void __ompt_force_initialization(); |
26 | |
27 | void __ompt_team_assign_id(kmp_team_t *team, ompt_data_t ompt_pid); |
28 | void __ompt_thread_assign_wait_id(void *variable); |
29 | |
30 | void __ompt_lw_taskteam_init(ompt_lw_taskteam_t *lwt, kmp_info_t *thr, int gtid, |
31 | ompt_data_t *ompt_pid, void *codeptr); |
32 | |
33 | void __ompt_lw_taskteam_link(ompt_lw_taskteam_t *lwt, kmp_info_t *thr, |
34 | int on_heap, bool always = false); |
35 | |
36 | void __ompt_lw_taskteam_unlink(kmp_info_t *thr); |
37 | |
38 | ompt_team_info_t *__ompt_get_teaminfo(int depth, int *size); |
39 | |
40 | ompt_data_t *__ompt_get_task_data(); |
41 | |
42 | ompt_data_t *__ompt_get_target_task_data(); |
43 | |
44 | ompt_task_info_t *__ompt_get_task_info_object(int depth); |
45 | |
46 | int __ompt_get_parallel_info_internal(int ancestor_level, |
47 | ompt_data_t **parallel_data, |
48 | int *team_size); |
49 | |
50 | int __ompt_get_task_info_internal(int ancestor_level, int *type, |
51 | ompt_data_t **task_data, |
52 | ompt_frame_t **task_frame, |
53 | ompt_data_t **parallel_data, int *thread_num); |
54 | |
55 | ompt_data_t *__ompt_get_thread_data_internal(); |
56 | |
57 | // __ompt_task_init: |
58 | // Initialize OMPT fields maintained by a task. This will only be called after |
59 | // ompt_start_tool, so we already know whether ompt is enabled or not. |
60 | |
61 | static inline void __ompt_task_init(kmp_taskdata_t *task, int tid) { |
62 | // The calls to __ompt_task_init already have the ompt_enabled condition. |
63 | task->ompt_task_info.task_data.value = 0; |
64 | task->ompt_task_info.frame.exit_frame = ompt_data_none; |
65 | task->ompt_task_info.frame.enter_frame = ompt_data_none; |
66 | task->ompt_task_info.frame.exit_frame_flags = |
67 | task->ompt_task_info.frame.enter_frame_flags = OMPT_FRAME_FLAGS_RUNTIME; |
68 | task->ompt_task_info.dispatch_chunk.start = 0; |
69 | task->ompt_task_info.dispatch_chunk.iterations = 0; |
70 | } |
71 | |
72 | /* |
73 | * Unused currently |
74 | static uint64_t __ompt_get_get_unique_id_internal(); |
75 | */ |
76 | |
77 | ompt_sync_region_t __ompt_get_barrier_kind(enum barrier_type, kmp_info_t *); |
78 | |
79 | /***************************************************************************** |
80 | * macros |
81 | ****************************************************************************/ |
82 | |
83 | #define OMPT_CUR_TASK_INFO(thr) (&((thr)->th.th_current_task->ompt_task_info)) |
84 | #define OMPT_CUR_TASK_DATA(thr) \ |
85 | (&((thr)->th.th_current_task->ompt_task_info.task_data)) |
86 | #define OMPT_CUR_TEAM_INFO(thr) (&((thr)->th.th_team->t.ompt_team_info)) |
87 | #define OMPT_CUR_TEAM_DATA(thr) \ |
88 | (&((thr)->th.th_team->t.ompt_team_info.parallel_data)) |
89 | |
90 | #define OMPT_HAVE_WEAK_ATTRIBUTE KMP_HAVE_WEAK_ATTRIBUTE |
91 | #define OMPT_HAVE_PSAPI KMP_HAVE_PSAPI |
92 | #define OMPT_STR_MATCH(haystack, needle) __kmp_str_match(haystack, 0, needle) |
93 | |
94 | inline void *__ompt_load_return_address(int gtid) { |
95 | kmp_info_t *thr = __kmp_threads[gtid]; |
96 | void *return_address = thr->th.ompt_thread_info.return_address; |
97 | thr->th.ompt_thread_info.return_address = NULL; |
98 | return return_address; |
99 | } |
100 | |
101 | /*#define OMPT_STORE_RETURN_ADDRESS(gtid) \ |
102 | if (ompt_enabled.enabled && gtid >= 0 && __kmp_threads[gtid] && \ |
103 | !__kmp_threads[gtid]->th.ompt_thread_info.return_address) \ |
104 | __kmp_threads[gtid]->th.ompt_thread_info.return_address = \ |
105 | __builtin_extract_return_addr(__builtin_return_address(0))*/ |
106 | #define OMPT_STORE_RETURN_ADDRESS(gtid) \ |
107 | OmptReturnAddressGuard ReturnAddressGuard{ \ |
108 | gtid, __builtin_extract_return_addr(__builtin_return_address(0))}; |
109 | #define OMPT_LOAD_RETURN_ADDRESS(gtid) __ompt_load_return_address(gtid) |
110 | #define OMPT_LOAD_OR_GET_RETURN_ADDRESS(gtid) \ |
111 | ((ompt_enabled.enabled && gtid >= 0 && __kmp_threads[gtid] && \ |
112 | __kmp_threads[gtid]->th.ompt_thread_info.return_address) \ |
113 | ? __ompt_load_return_address(gtid) \ |
114 | : __builtin_extract_return_addr(__builtin_return_address(0))) |
115 | |
116 | #define OMPT_GET_DISPATCH_CHUNK(chunk, lb, ub, incr) \ |
117 | do { \ |
118 | if (incr > 0) { \ |
119 | chunk.start = static_cast<uint64_t>(lb); \ |
120 | chunk.iterations = static_cast<uint64_t>(((ub) - (lb)) / (incr) + 1); \ |
121 | } else { \ |
122 | chunk.start = static_cast<uint64_t>(ub); \ |
123 | chunk.iterations = static_cast<uint64_t>(((lb) - (ub)) / -(incr) + 1); \ |
124 | } \ |
125 | } while (0) |
126 | |
127 | //****************************************************************************** |
128 | // inline functions |
129 | //****************************************************************************** |
130 | |
131 | inline kmp_info_t *ompt_get_thread_gtid(int gtid) { |
132 | return (gtid >= 0) ? __kmp_thread_from_gtid(gtid) : NULL; |
133 | } |
134 | |
135 | inline kmp_info_t *ompt_get_thread() { |
136 | int gtid = __kmp_get_gtid(); |
137 | return ompt_get_thread_gtid(gtid); |
138 | } |
139 | |
140 | inline void ompt_set_thread_state(kmp_info_t *thread, ompt_state_t state) { |
141 | if (thread) |
142 | thread->th.ompt_thread_info.state = state; |
143 | } |
144 | |
145 | inline const char *ompt_get_runtime_version() { |
146 | return &__kmp_version_lib_ver[KMP_VERSION_MAGIC_LEN]; |
147 | } |
148 | |
149 | inline ompt_work_t ompt_get_work_schedule(enum sched_type schedule) { |
150 | switch (SCHEDULE_WITHOUT_MODIFIERS(schedule)) { |
151 | case kmp_sch_static_chunked: |
152 | case kmp_sch_static_balanced: |
153 | case kmp_sch_static_greedy: |
154 | return ompt_work_loop_static; |
155 | case kmp_sch_dynamic_chunked: |
156 | case kmp_sch_static_steal: |
157 | return ompt_work_loop_dynamic; |
158 | case kmp_sch_guided_iterative_chunked: |
159 | case kmp_sch_guided_analytical_chunked: |
160 | case kmp_sch_guided_chunked: |
161 | case kmp_sch_guided_simd: |
162 | return ompt_work_loop_guided; |
163 | default: |
164 | return ompt_work_loop_other; |
165 | } |
166 | } |
167 | |
168 | class OmptReturnAddressGuard { |
169 | private: |
170 | bool SetAddress{false}; |
171 | int Gtid; |
172 | |
173 | public: |
174 | OmptReturnAddressGuard(int Gtid, void *ReturnAddress) : Gtid(Gtid) { |
175 | if (ompt_enabled.enabled && Gtid >= 0 && __kmp_threads[Gtid] && |
176 | !__kmp_threads[Gtid]->th.ompt_thread_info.return_address) { |
177 | SetAddress = true; |
178 | __kmp_threads[Gtid]->th.ompt_thread_info.return_address = ReturnAddress; |
179 | } |
180 | } |
181 | ~OmptReturnAddressGuard() { |
182 | if (SetAddress) |
183 | __kmp_threads[Gtid]->th.ompt_thread_info.return_address = NULL; |
184 | } |
185 | }; |
186 | |
187 | #endif // OMPT_SUPPORT |
188 | |
189 | // macros providing the OMPT callbacks for reduction clause |
190 | #if OMPT_SUPPORT && OMPT_OPTIONAL |
191 | #define OMPT_REDUCTION_DECL(this_thr, gtid) \ |
192 | ompt_data_t *my_task_data = OMPT_CUR_TASK_DATA(this_thr); \ |
193 | ompt_data_t *my_parallel_data = OMPT_CUR_TEAM_DATA(this_thr); \ |
194 | void *return_address = OMPT_LOAD_RETURN_ADDRESS(gtid); |
195 | #define OMPT_REDUCTION_BEGIN \ |
196 | if (ompt_enabled.enabled && ompt_enabled.ompt_callback_reduction) { \ |
197 | ompt_callbacks.ompt_callback(ompt_callback_reduction)( \ |
198 | ompt_sync_region_reduction, ompt_scope_begin, my_parallel_data, \ |
199 | my_task_data, return_address); \ |
200 | } |
201 | #define OMPT_REDUCTION_END \ |
202 | if (ompt_enabled.enabled && ompt_enabled.ompt_callback_reduction) { \ |
203 | ompt_callbacks.ompt_callback(ompt_callback_reduction)( \ |
204 | ompt_sync_region_reduction, ompt_scope_end, my_parallel_data, \ |
205 | my_task_data, return_address); \ |
206 | } |
207 | #else // OMPT_SUPPORT && OMPT_OPTIONAL |
208 | #define OMPT_REDUCTION_DECL(this_thr, gtid) |
209 | #define OMPT_REDUCTION_BEGIN |
210 | #define OMPT_REDUCTION_END |
211 | #endif // ! OMPT_SUPPORT && OMPT_OPTIONAL |
212 | |
213 | #endif |
214 | |