Lines Matching defs:upper

4794     // Intel task just returns the upper bound normally
4833 // Intel task just sets the upper bound normally
4839 kmp_uint32 *upper = RCAST(kmp_uint32 *, task->shareds) + 1;
4840 *upper = (kmp_uint32)ub;
4842 kmp_uint64 *upper = RCAST(kmp_uint64 *, task->shareds) + 1;
4843 *upper = (kmp_uint64)ub;
4858 // ub Pointer to loop upper bound in task structure
4860 // ub_glob Global upper bound (used for lastprivate check)
4883 kmp_uint64 upper = task_bounds.get_ub();
4896 gtid, num_tasks, grainsize, extras, last_chunk, lower, upper,
4908 upper = lower + st * chunk_minus_1;
4909 if (upper > *ub) {
4910 upper = *ub;
4915 KMP_DEBUG_ASSERT(upper == *ub);
4916 if (upper == ub_glob)
4919 KMP_DEBUG_ASSERT((kmp_uint64)st > *ub - upper);
4920 if ((kmp_uint64)st > ub_glob - upper)
4923 KMP_DEBUG_ASSERT(upper + st < *ub);
4924 if (upper - ub_glob < (kmp_uint64)(-st))
4942 next_task_bounds.set_ub(upper + (st > 0 ? 1 : -1));
4944 next_task_bounds.set_ub(upper);
4951 "upper %lld stride %lld, (offsets %p %p)\n",
4952 gtid, i, next_task, lower, upper, st,
4961 lower, upper, st);
4967 lower = upper + st; // adjust lower bound for the next iteration
5060 // ub Pointer to loop upper bound in task structure
5062 // ub_glob Global upper bound (used for lastprivate check)
5137 // adjust lower bound (upper bound is not changed) for the 2nd half
5141 *ub = ub0; // adjust upper bound for the 1st half
5223 kmp_uint64 upper = task_bounds.get_ub();
5224 kmp_uint64 ub_glob = upper; // global upper used to calc lastprivate flag
5234 gtid, taskdata, lower, upper, st, grainsize, sched, modifier,
5239 tc = upper - lower + 1;
5241 tc = (lower - upper) / (-st) + 1;
5243 tc = (upper - lower) / st + 1;
5376 @param ub Pointer to loop upper bound in task structure
5402 @param ub Pointer to loop upper bound in task structure