@@ -50,7 +50,7 @@ __kmp_linear_barrier_gather(enum barrier_type bt, kmp_info_t *this_thr, int gtid
50
50
void (*reduce)(void *, void *)
51
51
USE_ITT_BUILD_ARG(void * itt_sync_obj) )
52
52
{
53
- KMP_TIME_DEVELOPER_BLOCK (KMP_linear_gather);
53
+ KMP_TIME_DEVELOPER_PARTITIONED_BLOCK (KMP_linear_gather);
54
54
register kmp_team_t *team = this_thr->th .th_team ;
55
55
register kmp_bstate_t *thr_bar = & this_thr->th .th_bar [bt].bb ;
56
56
register kmp_info_t **other_threads = team->t .t_threads ;
@@ -130,7 +130,7 @@ __kmp_linear_barrier_release(enum barrier_type bt, kmp_info_t *this_thr, int gti
130
130
int propagate_icvs
131
131
USE_ITT_BUILD_ARG (void *itt_sync_obj) )
132
132
{
133
- KMP_TIME_DEVELOPER_BLOCK (KMP_linear_release);
133
+ KMP_TIME_DEVELOPER_PARTITIONED_BLOCK (KMP_linear_release);
134
134
register kmp_bstate_t *thr_bar = &this_thr->th .th_bar [bt].bb ;
135
135
register kmp_team_t *team;
136
136
@@ -149,7 +149,7 @@ __kmp_linear_barrier_release(enum barrier_type bt, kmp_info_t *this_thr, int gti
149
149
if (nproc > 1 ) {
150
150
#if KMP_BARRIER_ICV_PUSH
151
151
{
152
- KMP_TIME_DEVELOPER_BLOCK (USER_icv_copy);
152
+ KMP_TIME_DEVELOPER_PARTITIONED_BLOCK (USER_icv_copy);
153
153
if (propagate_icvs) {
154
154
ngo_load (&team->t .t_implicit_task_taskdata [0 ].td_icvs );
155
155
for (i=1 ; i<nproc; ++i) {
@@ -225,7 +225,7 @@ __kmp_tree_barrier_gather(enum barrier_type bt, kmp_info_t *this_thr, int gtid,
225
225
void (*reduce)(void *, void *)
226
226
USE_ITT_BUILD_ARG(void *itt_sync_obj) )
227
227
{
228
- KMP_TIME_DEVELOPER_BLOCK (KMP_tree_gather);
228
+ KMP_TIME_DEVELOPER_PARTITIONED_BLOCK (KMP_tree_gather);
229
229
register kmp_team_t *team = this_thr->th .th_team ;
230
230
register kmp_bstate_t *thr_bar = &this_thr->th .th_bar [bt].bb ;
231
231
register kmp_info_t **other_threads = team->t .t_threads ;
@@ -323,7 +323,7 @@ __kmp_tree_barrier_release(enum barrier_type bt, kmp_info_t *this_thr, int gtid,
323
323
int propagate_icvs
324
324
USE_ITT_BUILD_ARG (void *itt_sync_obj) )
325
325
{
326
- KMP_TIME_DEVELOPER_BLOCK (KMP_tree_release);
326
+ KMP_TIME_DEVELOPER_PARTITIONED_BLOCK (KMP_tree_release);
327
327
register kmp_team_t *team;
328
328
register kmp_bstate_t *thr_bar = &this_thr->th .th_bar [bt].bb ;
329
329
register kmp_uint32 nproc;
@@ -393,7 +393,7 @@ __kmp_tree_barrier_release(enum barrier_type bt, kmp_info_t *this_thr, int gtid,
393
393
394
394
#if KMP_BARRIER_ICV_PUSH
395
395
{
396
- KMP_TIME_DEVELOPER_BLOCK (USER_icv_copy);
396
+ KMP_TIME_DEVELOPER_PARTITIONED_BLOCK (USER_icv_copy);
397
397
if (propagate_icvs) {
398
398
__kmp_init_implicit_task (team->t .t_ident , team->t .t_threads [child_tid],
399
399
team, child_tid, FALSE );
@@ -426,7 +426,7 @@ __kmp_hyper_barrier_gather(enum barrier_type bt, kmp_info_t *this_thr, int gtid,
426
426
void (*reduce)(void *, void *)
427
427
USE_ITT_BUILD_ARG(void *itt_sync_obj) )
428
428
{
429
- KMP_TIME_DEVELOPER_BLOCK (KMP_hyper_gather);
429
+ KMP_TIME_DEVELOPER_PARTITIONED_BLOCK (KMP_hyper_gather);
430
430
register kmp_team_t *team = this_thr->th .th_team ;
431
431
register kmp_bstate_t *thr_bar = &this_thr->th .th_bar [bt].bb ;
432
432
register kmp_info_t **other_threads = team->t .t_threads ;
@@ -535,7 +535,7 @@ __kmp_hyper_barrier_release(enum barrier_type bt, kmp_info_t *this_thr, int gtid
535
535
int propagate_icvs
536
536
USE_ITT_BUILD_ARG (void *itt_sync_obj) )
537
537
{
538
- KMP_TIME_DEVELOPER_BLOCK (KMP_hyper_release);
538
+ KMP_TIME_DEVELOPER_PARTITIONED_BLOCK (KMP_hyper_release);
539
539
register kmp_team_t *team;
540
540
register kmp_bstate_t *thr_bar = & this_thr -> th.th_bar [ bt ].bb ;
541
541
register kmp_info_t **other_threads;
@@ -742,7 +742,7 @@ __kmp_hierarchical_barrier_gather(enum barrier_type bt, kmp_info_t *this_thr,
742
742
int gtid, int tid, void (*reduce) (void *, void *)
743
743
USE_ITT_BUILD_ARG(void * itt_sync_obj) )
744
744
{
745
- KMP_TIME_DEVELOPER_BLOCK (KMP_hier_gather);
745
+ KMP_TIME_DEVELOPER_PARTITIONED_BLOCK (KMP_hier_gather);
746
746
register kmp_team_t *team = this_thr->th .th_team ;
747
747
register kmp_bstate_t *thr_bar = & this_thr->th .th_bar [bt].bb ;
748
748
register kmp_uint32 nproc = this_thr->th .th_team_nproc ;
@@ -883,7 +883,7 @@ __kmp_hierarchical_barrier_release(enum barrier_type bt, kmp_info_t *this_thr, i
883
883
int propagate_icvs
884
884
USE_ITT_BUILD_ARG (void * itt_sync_obj) )
885
885
{
886
- KMP_TIME_DEVELOPER_BLOCK (KMP_hier_release);
886
+ KMP_TIME_DEVELOPER_PARTITIONED_BLOCK (KMP_hier_release);
887
887
register kmp_team_t *team;
888
888
register kmp_bstate_t *thr_bar = &this_thr->th .th_bar [bt].bb ;
889
889
register kmp_uint32 nproc;
@@ -1067,9 +1067,8 @@ int
1067
1067
__kmp_barrier (enum barrier_type bt, int gtid, int is_split, size_t reduce_size,
1068
1068
void *reduce_data, void (*reduce)(void *, void *))
1069
1069
{
1070
- KMP_TIME_DEVELOPER_BLOCK (KMP_barrier);
1071
- KMP_SET_THREAD_STATE_BLOCK (PLAIN_BARRIER);
1072
1070
KMP_TIME_PARTITIONED_BLOCK (OMP_plain_barrier);
1071
+ KMP_SET_THREAD_STATE_BLOCK (PLAIN_BARRIER);
1073
1072
register int tid = __kmp_tid_from_gtid (gtid);
1074
1073
register kmp_info_t *this_thr = __kmp_threads[gtid];
1075
1074
register kmp_team_t *team = this_thr->th .th_team ;
@@ -1333,7 +1332,8 @@ __kmp_barrier(enum barrier_type bt, int gtid, int is_split, size_t reduce_size,
1333
1332
void
1334
1333
__kmp_end_split_barrier (enum barrier_type bt, int gtid)
1335
1334
{
1336
- KMP_TIME_DEVELOPER_BLOCK (KMP_end_split_barrier);
1335
+ KMP_TIME_DEVELOPER_PARTITIONED_BLOCK (KMP_end_split_barrier);
1336
+ KMP_SET_THREAD_STATE_BLOCK (PLAIN_BARRIER);
1337
1337
int tid = __kmp_tid_from_gtid (gtid);
1338
1338
kmp_info_t *this_thr = __kmp_threads[gtid];
1339
1339
kmp_team_t *team = this_thr->th .th_team ;
@@ -1376,9 +1376,8 @@ __kmp_end_split_barrier(enum barrier_type bt, int gtid)
1376
1376
void
1377
1377
__kmp_join_barrier (int gtid)
1378
1378
{
1379
- KMP_TIME_PARTITIONED_BLOCK (OMP_fork_join_barrier );
1379
+ KMP_TIME_PARTITIONED_BLOCK (OMP_join_barrier );
1380
1380
KMP_SET_THREAD_STATE_BLOCK (FORK_JOIN_BARRIER);
1381
- KMP_TIME_DEVELOPER_BLOCK (KMP_join_barrier);
1382
1381
register kmp_info_t *this_thr = __kmp_threads[gtid];
1383
1382
register kmp_team_t *team;
1384
1383
register kmp_uint nproc;
@@ -1592,9 +1591,8 @@ __kmp_join_barrier(int gtid)
1592
1591
void
1593
1592
__kmp_fork_barrier (int gtid, int tid)
1594
1593
{
1595
- KMP_TIME_PARTITIONED_BLOCK (OMP_fork_join_barrier );
1594
+ KMP_TIME_PARTITIONED_BLOCK (OMP_fork_barrier );
1596
1595
KMP_SET_THREAD_STATE_BLOCK (FORK_JOIN_BARRIER);
1597
- KMP_TIME_DEVELOPER_BLOCK (KMP_fork_barrier);
1598
1596
kmp_info_t *this_thr = __kmp_threads[gtid];
1599
1597
kmp_team_t *team = (tid == 0 ) ? this_thr->th .th_team : NULL ;
1600
1598
#if USE_ITT_BUILD
@@ -1707,7 +1705,7 @@ __kmp_fork_barrier(int gtid, int tid)
1707
1705
the fixed ICVs in the master's thread struct, because it is not always the case that the
1708
1706
threads arrays have been allocated when __kmp_fork_call() is executed. */
1709
1707
{
1710
- KMP_TIME_DEVELOPER_BLOCK (USER_icv_copy);
1708
+ KMP_TIME_DEVELOPER_PARTITIONED_BLOCK (USER_icv_copy);
1711
1709
if (!KMP_MASTER_TID (tid)) { // master thread already has ICVs
1712
1710
// Copy the initial ICVs from the master's thread struct to the implicit task for this tid.
1713
1711
KA_TRACE (10 , (" __kmp_fork_barrier: T#%d(%d) is PULLing ICVs\n " , gtid, tid));
@@ -1762,7 +1760,7 @@ __kmp_fork_barrier(int gtid, int tid)
1762
1760
void
1763
1761
__kmp_setup_icv_copy (kmp_team_t *team, int new_nproc, kmp_internal_control_t *new_icvs, ident_t *loc )
1764
1762
{
1765
- KMP_TIME_DEVELOPER_BLOCK (KMP_setup_icv_copy);
1763
+ KMP_TIME_DEVELOPER_PARTITIONED_BLOCK (KMP_setup_icv_copy);
1766
1764
1767
1765
KMP_DEBUG_ASSERT (team && new_nproc && new_icvs);
1768
1766
KMP_DEBUG_ASSERT ((!TCR_4 (__kmp_init_parallel)) || new_icvs->nproc );
0 commit comments