Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[OpenMP] Emit workshare events correctly for distribute-and-parallel loops #133323

Open
wants to merge 4 commits into
base: main
Choose a base branch
from
Open
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
1 change: 1 addition & 0 deletions openmp/runtime/src/dllexports
Original file line number Diff line number Diff line change
Expand Up @@ -351,6 +351,7 @@ kmpc_set_defaults 224
__kmpc_dist_for_static_init_4u 248
__kmpc_dist_for_static_init_8 249
__kmpc_dist_for_static_init_8u 250
__kmpc_dist_for_static_fini
__kmpc_dist_dispatch_init_4 251
__kmpc_dist_dispatch_init_4u 252
__kmpc_dist_dispatch_init_8 253
Expand Down
1 change: 1 addition & 0 deletions openmp/runtime/src/kmp.h
Original file line number Diff line number Diff line change
Expand Up @@ -4240,6 +4240,7 @@ KMP_EXPORT void KMPC_FOR_STATIC_INIT(ident_t *loc, kmp_int32 global_tid,
kmp_int chunk);

KMP_EXPORT void __kmpc_for_static_fini(ident_t *loc, kmp_int32 global_tid);
KMP_EXPORT void __kmpc_dist_for_static_fini(ident_t *loc, kmp_int32 gtid);

KMP_EXPORT void __kmpc_copyprivate(ident_t *loc, kmp_int32 global_tid,
size_t cpy_size, void *cpy_data,
Expand Down
35 changes: 35 additions & 0 deletions openmp/runtime/src/kmp_csupport.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -2032,6 +2032,41 @@ void __kmpc_for_static_fini(ident_t *loc, kmp_int32 global_tid) {
__kmp_pop_workshare(global_tid, ct_pdo, loc);
}

/*!
@ingroup WORK_SHARING
@param loc Source location
@param global_tid Global thread id

Mark the end of a statically scheduled distribute and parallel loop
*/
void __kmpc_dist_for_static_fini(ident_t *loc, kmp_int32 global_tid) {
KMP_POP_PARTITIONED_TIMER();
KE_TRACE(10, ("__kmpc_dist_for_static_fini called T#%d\n", global_tid));

#if OMPT_SUPPORT && OMPT_OPTIONAL
if (ompt_enabled.ompt_callback_work) {
// Workshare type is distribute and parallel loop.
// Emit ws-loop-end event for all threads.
// Emit distribute-end event for the primary threads.
ompt_team_info_t *team_info = __ompt_get_teaminfo(0, NULL);
ompt_task_info_t *task_info = __ompt_get_task_info_object(0);
int tid = __kmp_tid_from_gtid(global_tid);

ompt_callbacks.ompt_callback(ompt_callback_work)(
ompt_work_loop_static, ompt_scope_end, &(team_info->parallel_data),
&(task_info->task_data), 0, OMPT_GET_RETURN_ADDRESS(0));

if (tid == 0)
ompt_callbacks.ompt_callback(ompt_callback_work)(
ompt_work_distribute, ompt_scope_end, &(team_info->parallel_data),
&(task_info->task_data), 0, OMPT_GET_RETURN_ADDRESS(0));
}
#endif

if (__kmp_env_consistency_check)
__kmp_pop_workshare(global_tid, ct_pdo, loc);
}

// User routines which take C-style arguments (call by value)
// different from the Fortran equivalent routines

Expand Down
9 changes: 8 additions & 1 deletion openmp/runtime/src/kmp_sched.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -732,11 +732,18 @@ end:;
KE_TRACE(10, ("__kmpc_dist_for_static_init: T#%d return\n", gtid));
#if OMPT_SUPPORT && OMPT_OPTIONAL
if (ompt_enabled.ompt_callback_work || ompt_enabled.ompt_callback_dispatch) {
// Workshare type is distribute and parallel loop.
// Emit ws-loop-begin event for all threads.
// Emit distribute-begin event for the primary threads.
ompt_team_info_t *team_info = __ompt_get_teaminfo(0, NULL);
ompt_task_info_t *task_info = __ompt_get_task_info_object(0);
if (ompt_enabled.ompt_callback_work) {
if (tid == 0)
ompt_callbacks.ompt_callback(ompt_callback_work)(
ompt_work_distribute, ompt_scope_begin, &(team_info->parallel_data),
&(task_info->task_data), 0, codeptr);
ompt_callbacks.ompt_callback(ompt_callback_work)(
ompt_work_distribute, ompt_scope_begin, &(team_info->parallel_data),
ompt_work_loop_static, ompt_scope_begin, &(team_info->parallel_data),
&(task_info->task_data), 0, codeptr);
}
if (ompt_enabled.ompt_callback_dispatch) {
Expand Down