Lines Matching refs:proc_id

231     for (PetscMPIInt proc_id = 0; proc_id < size; proc_id++) {  in MatIncreaseOverlap_MPISBAIJ_Once()  local
232 for (; j < C->rmap->range[proc_id + 1] / bs; j++) ctable[j] = proc_id; in MatIncreaseOverlap_MPISBAIJ_Once()
252 for (PetscMPIInt proc_id = 0; proc_id < size; proc_id++) { in MatIncreaseOverlap_MPISBAIJ_Once() local
253 iwork[proc_id] = 0; in MatIncreaseOverlap_MPISBAIJ_Once()
254 *data1_start[proc_id] = is_max; in MatIncreaseOverlap_MPISBAIJ_Once()
255 data1_start[proc_id]++; in MatIncreaseOverlap_MPISBAIJ_Once()
257 if (proc_id == rank) { in MatIncreaseOverlap_MPISBAIJ_Once()
258 *data1_start[proc_id] = n[j]; in MatIncreaseOverlap_MPISBAIJ_Once()
260 *data1_start[proc_id] = 0; in MatIncreaseOverlap_MPISBAIJ_Once()
262 data1_start[proc_id]++; in MatIncreaseOverlap_MPISBAIJ_Once()
273 …for (PetscMPIInt proc_id = 0; proc_id <= proc_end; proc_id++) { /* for others to proces… in MatIncreaseOverlap_MPISBAIJ_Once() local
274 …if (proc_id == rank) continue; /* done before this loop … in MatIncreaseOverlap_MPISBAIJ_Once()
275 …if (proc_id < proc_end && !PetscBTLookup(table[proc_id], idx)) continue; /* no need for sending id… in MatIncreaseOverlap_MPISBAIJ_Once()
276 *data1_start[proc_id] = idx; in MatIncreaseOverlap_MPISBAIJ_Once()
277 data1_start[proc_id]++; in MatIncreaseOverlap_MPISBAIJ_Once()
278 len_s[proc_id]++; in MatIncreaseOverlap_MPISBAIJ_Once()
282 for (PetscMPIInt proc_id = 0; proc_id < size; proc_id++) { in MatIncreaseOverlap_MPISBAIJ_Once() local
283 if (proc_id == rank) continue; in MatIncreaseOverlap_MPISBAIJ_Once()
284 *(data1 + proc_id * len + 1 + i) = len_s[proc_id] - iwork[proc_id]; in MatIncreaseOverlap_MPISBAIJ_Once()
285 iwork[proc_id] = len_s[proc_id]; in MatIncreaseOverlap_MPISBAIJ_Once()
311 for (PetscMPIInt proc_id = 0; proc_id < size; proc_id++) { /* send data1 to processor [proc_id] */ in MatIncreaseOverlap_MPISBAIJ_Once() local
312 if (len_s[proc_id]) { in MatIncreaseOverlap_MPISBAIJ_Once()
313 …PetscCallMPI(MPIU_Isend(data1_start[proc_id], len_s[proc_id], MPIU_INT, proc_id, tag1, comm, s_wai… in MatIncreaseOverlap_MPISBAIJ_Once()
326 for (PetscMPIInt proc_id = 0; proc_id < size; proc_id++) len_s[proc_id] = iwork[proc_id] = 0; in MatIncreaseOverlap_MPISBAIJ_Once() local
348 PetscMPIInt proc_id; in MatIncreaseOverlap_MPISBAIJ_Once() local
351 proc_id = r_status.MPI_SOURCE; in MatIncreaseOverlap_MPISBAIJ_Once()
352 PetscCallMPI(MPIU_Irecv(odata1, ilen, MPIU_INT, proc_id, r_status.MPI_TAG, comm, &r_req)); in MatIncreaseOverlap_MPISBAIJ_Once()
368 PetscCallMPI(MPIU_Isend(odata2, len, MPIU_INT, proc_id, tag2, comm, s_waits2 + k)); in MatIncreaseOverlap_MPISBAIJ_Once()
372 … PetscCall(PetscMPIIntCast(len, &len_s[proc_id])); /* length of message sending back to proc_id */ in MatIncreaseOverlap_MPISBAIJ_Once()
402 PetscMPIInt proc_id, ilen; in MatIncreaseOverlap_MPISBAIJ_Once() local
404 proc_id = r_status.MPI_SOURCE; in MatIncreaseOverlap_MPISBAIJ_Once()
405 PetscCallMPI(MPIU_Irecv(data2, ilen, MPIU_INT, proc_id, r_status.MPI_TAG, comm, &r_req)); in MatIncreaseOverlap_MPISBAIJ_Once()