xref: /petsc/src/mat/impls/baij/mpi/baijov.c (revision 2235c4e21094f57f73a5851b0c05b8ecf6dffc6d)
1 /*
2    Routines to compute overlapping regions of a parallel MPI matrix
3   and to find submatrices that were shared across processors.
4 */
5 #include <../src/mat/impls/baij/mpi/mpibaij.h>
6 #include <petscbt.h>
7 
8 static PetscErrorCode MatIncreaseOverlap_MPIBAIJ_Local(Mat, PetscInt, PetscBT *, PetscInt *, PetscInt **);
9 static PetscErrorCode MatIncreaseOverlap_MPIBAIJ_Receive(Mat, PetscInt, PetscInt **, PetscInt **, PetscInt *);
10 
MatIncreaseOverlap_MPIBAIJ(Mat C,PetscInt imax,IS is[],PetscInt ov)11 PetscErrorCode MatIncreaseOverlap_MPIBAIJ(Mat C, PetscInt imax, IS is[], PetscInt ov)
12 {
13   PetscInt        i, N = C->cmap->N, bs = C->rmap->bs, n;
14   const PetscInt *idx;
15   IS             *is_new;
16 
17   PetscFunctionBegin;
18   PetscCall(PetscMalloc1(imax, &is_new));
19   /* Convert the indices into block format */
20   PetscCall(ISCompressIndicesGeneral(N, C->rmap->n, bs, imax, is, is_new));
21   PetscCheck(ov >= 0, PETSC_COMM_SELF, PETSC_ERR_ARG_OUTOFRANGE, "Negative overlap specified");
22   for (i = 0; i < ov; ++i) PetscCall(MatIncreaseOverlap_MPIBAIJ_Once(C, imax, is_new));
23   for (i = 0; i < imax; i++) {
24     PetscCall(ISDestroy(&is[i]));
25     PetscCall(ISGetLocalSize(is_new[i], &n));
26     PetscCall(ISGetIndices(is_new[i], &idx));
27     PetscCall(ISCreateBlock(PetscObjectComm((PetscObject)is_new[i]), bs, n, idx, PETSC_COPY_VALUES, &is[i]));
28     PetscCall(ISDestroy(&is_new[i]));
29   }
30   PetscCall(PetscFree(is_new));
31   PetscFunctionReturn(PETSC_SUCCESS);
32 }
33 
34 /*
35   Sample message format:
36   If a processor A wants processor B to process some elements corresponding
37   to index sets is[1], is[5]
38   mesg [0] = 2   (no of index sets in the mesg)
39   -----------
40   mesg [1] = 1 => is[1]
41   mesg [2] = sizeof(is[1]);
42   -----------
43   mesg [5] = 5  => is[5]
44   mesg [6] = sizeof(is[5]);
45   -----------
46   mesg [7]
47   mesg [n]  data(is[1])
48   -----------
49   mesg[n+1]
50   mesg[m]  data(is[5])
51   -----------
52 
53   Notes:
54   nrqs - no of requests sent (or to be sent out)
55   nrqr - no of requests received (which have to be or which have been processed)
56 */
MatIncreaseOverlap_MPIBAIJ_Once(Mat C,PetscInt imax,IS is[])57 PetscErrorCode MatIncreaseOverlap_MPIBAIJ_Once(Mat C, PetscInt imax, IS is[])
58 {
59   Mat_MPIBAIJ     *c = (Mat_MPIBAIJ *)C->data;
60   const PetscInt **idx, *idx_i;
61   PetscInt        *n, *w3, *w4, **data, len;
62   PetscMPIInt      size, rank, tag1, tag2, *w2, *w1, nrqs, nrqr, *pa;
63   PetscInt         Mbs, **rbuf, row, msz, **outdat, **ptr;
64   PetscInt        *ctr, *tmp, *isz, *isz1, **xdata, **rbuf2, *d_p;
65   PetscMPIInt     *onodes1, *olengths1, *onodes2, *olengths2, proc = -1;
66   PetscBT         *table;
67   MPI_Comm         comm, *iscomms;
68   MPI_Request     *s_waits1, *r_waits1, *s_waits2, *r_waits2;
69   PetscByte       *t_p;
70 
71   PetscFunctionBegin;
72   PetscCall(PetscObjectGetComm((PetscObject)C, &comm));
73   size = c->size;
74   rank = c->rank;
75   Mbs  = c->Mbs;
76 
77   PetscCall(PetscObjectGetNewTag((PetscObject)C, &tag1));
78   PetscCall(PetscObjectGetNewTag((PetscObject)C, &tag2));
79 
80   PetscCall(PetscMalloc2(imax, (PetscInt ***)&idx, imax, &n));
81 
82   for (PetscInt i = 0; i < imax; i++) {
83     PetscCall(ISGetIndices(is[i], &idx[i]));
84     PetscCall(ISGetLocalSize(is[i], &n[i]));
85   }
86 
87   /* evaluate communication - mesg to who,length of mesg, and buffer space
88      required. Based on this, buffers are allocated, and data copied into them*/
89   PetscCall(PetscCalloc4(size, &w1, size, &w2, size, &w3, size, &w4));
90   for (PetscInt i = 0; i < imax; i++) {
91     PetscCall(PetscArrayzero(w4, size)); /* initialise work vector*/
92     idx_i = idx[i];
93     len   = n[i];
94     for (PetscInt j = 0; j < len; j++) {
95       row = idx_i[j];
96       PetscCheck(row >= 0, PETSC_COMM_SELF, PETSC_ERR_ARG_OUTOFRANGE, "Index set cannot have negative entries");
97       PetscCall(PetscLayoutFindOwner(C->rmap, row * C->rmap->bs, &proc));
98       w4[proc]++;
99     }
100     for (PetscMPIInt j = 0; j < size; j++) {
101       if (w4[j]) {
102         w1[j] += w4[j];
103         w3[j]++;
104       }
105     }
106   }
107 
108   nrqs     = 0; /* no of outgoing messages */
109   msz      = 0; /* total mesg length (for all proc */
110   w1[rank] = 0; /* no mesg sent to itself */
111   w3[rank] = 0;
112   for (PetscMPIInt i = 0; i < size; i++) {
113     if (w1[i]) {
114       w2[i] = 1;
115       nrqs++;
116     } /* there exists a message to proc i */
117   }
118   /* pa - is list of processors to communicate with */
119   PetscCall(PetscMalloc1(nrqs, &pa));
120   for (PetscMPIInt i = 0, j = 0; i < size; i++) {
121     if (w1[i]) {
122       pa[j] = i;
123       j++;
124     }
125   }
126 
127   /* Each message would have a header = 1 + 2*(no of IS) + data */
128   for (PetscMPIInt i = 0; i < nrqs; i++) {
129     PetscMPIInt j = pa[i];
130     w1[j] += w2[j] + 2 * w3[j];
131     msz += w1[j];
132   }
133 
134   /* Determine the number of messages to expect, their lengths, from from-ids */
135   PetscCall(PetscGatherNumberOfMessages(comm, w2, w1, &nrqr));
136   PetscCall(PetscGatherMessageLengths(comm, nrqs, nrqr, w1, &onodes1, &olengths1));
137 
138   /* Now post the Irecvs corresponding to these messages */
139   PetscCall(PetscPostIrecvInt(comm, tag1, nrqr, onodes1, olengths1, &rbuf, &r_waits1));
140 
141   /* Allocate Memory for outgoing messages */
142   PetscCall(PetscMalloc4(size, &outdat, size, &ptr, msz, &tmp, size, &ctr));
143   PetscCall(PetscArrayzero(outdat, size));
144   PetscCall(PetscArrayzero(ptr, size));
145   {
146     PetscInt *iptr = tmp, ict = 0;
147     for (PetscMPIInt i = 0; i < nrqs; i++) {
148       PetscMPIInt j = pa[i];
149       iptr += ict;
150       outdat[j] = iptr;
151       ict       = w1[j];
152     }
153   }
154 
155   /* Form the outgoing messages */
156   /*plug in the headers*/
157   for (PetscMPIInt i = 0; i < nrqs; i++) {
158     PetscMPIInt j = pa[i];
159     outdat[j][0]  = 0;
160     PetscCall(PetscArrayzero(outdat[j] + 1, 2 * w3[j]));
161     ptr[j] = outdat[j] + 2 * w3[j] + 1;
162   }
163 
164   /* Memory for doing local proc's work*/
165   {
166     PetscCall(PetscCalloc5(imax, &table, imax, &data, imax, &isz, Mbs * imax, &d_p, (Mbs / PETSC_BITS_PER_BYTE + 1) * imax, &t_p));
167 
168     for (PetscInt i = 0; i < imax; i++) {
169       table[i] = t_p + (Mbs / PETSC_BITS_PER_BYTE + 1) * i;
170       data[i]  = d_p + (Mbs)*i;
171     }
172   }
173 
174   /* Parse the IS and update local tables and the outgoing buf with the data*/
175   {
176     PetscInt n_i, *data_i, isz_i, *outdat_j, ctr_j, k;
177     PetscBT  table_i;
178 
179     for (PetscInt i = 0; i < imax; i++) {
180       PetscCall(PetscArrayzero(ctr, size));
181       n_i     = n[i];
182       table_i = table[i];
183       idx_i   = idx[i];
184       data_i  = data[i];
185       isz_i   = isz[i];
186       for (PetscInt j = 0; j < n_i; j++) { /* parse the indices of each IS */
187         row = idx_i[j];
188         PetscCall(PetscLayoutFindOwner(C->rmap, row * C->rmap->bs, &proc));
189         if (proc != rank) { /* copy to the outgoing buffer */
190           ctr[proc]++;
191           *ptr[proc] = row;
192           ptr[proc]++;
193         } else { /* Update the local table */
194           if (!PetscBTLookupSet(table_i, row)) data_i[isz_i++] = row;
195         }
196       }
197       /* Update the headers for the current IS */
198       for (PetscMPIInt j = 0; j < size; j++) { /* Can Optimise this loop by using pa[] */
199         if ((ctr_j = ctr[j])) {
200           outdat_j            = outdat[j];
201           k                   = ++outdat_j[0];
202           outdat_j[2 * k]     = ctr_j;
203           outdat_j[2 * k - 1] = i;
204         }
205       }
206       isz[i] = isz_i;
207     }
208   }
209 
210   /*  Now  post the sends */
211   PetscCall(PetscMalloc1(nrqs, &s_waits1));
212   for (PetscMPIInt i = 0; i < nrqs; ++i) {
213     PetscMPIInt j = pa[i];
214     PetscCallMPI(MPIU_Isend(outdat[j], w1[j], MPIU_INT, j, tag1, comm, s_waits1 + i));
215   }
216 
217   /* No longer need the original indices*/
218   for (PetscInt i = 0; i < imax; ++i) PetscCall(ISRestoreIndices(is[i], idx + i));
219   PetscCall(PetscFree2(*(PetscInt ***)&idx, n));
220 
221   PetscCall(PetscMalloc1(imax, &iscomms));
222   for (PetscInt i = 0; i < imax; ++i) {
223     PetscCall(PetscCommDuplicate(PetscObjectComm((PetscObject)is[i]), &iscomms[i], NULL));
224     PetscCall(ISDestroy(&is[i]));
225   }
226 
227   /* Do Local work*/
228   PetscCall(MatIncreaseOverlap_MPIBAIJ_Local(C, imax, table, isz, data));
229 
230   /* Receive messages*/
231   PetscCallMPI(MPI_Waitall(nrqr, r_waits1, MPI_STATUSES_IGNORE));
232   PetscCallMPI(MPI_Waitall(nrqs, s_waits1, MPI_STATUSES_IGNORE));
233 
234   /* Phase 1 sends are complete - deallocate buffers */
235   PetscCall(PetscFree4(outdat, ptr, tmp, ctr));
236   PetscCall(PetscFree4(w1, w2, w3, w4));
237 
238   PetscCall(PetscMalloc1(nrqr, &xdata));
239   PetscCall(PetscMalloc1(nrqr, &isz1));
240   PetscCall(MatIncreaseOverlap_MPIBAIJ_Receive(C, nrqr, rbuf, xdata, isz1));
241   if (rbuf) {
242     PetscCall(PetscFree(rbuf[0]));
243     PetscCall(PetscFree(rbuf));
244   }
245 
246   /* Send the data back*/
247   /* Do a global reduction to know the buffer space req for incoming messages*/
248   {
249     PetscMPIInt *rw1;
250 
251     PetscCall(PetscCalloc1(size, &rw1));
252     for (PetscMPIInt i = 0; i < nrqr; ++i) {
253       proc = onodes1[i];
254       PetscCall(PetscMPIIntCast(isz1[i], &rw1[proc]));
255     }
256 
257     /* Determine the number of messages to expect, their lengths, from from-ids */
258     PetscCall(PetscGatherMessageLengths(comm, nrqr, nrqs, rw1, &onodes2, &olengths2));
259     PetscCall(PetscFree(rw1));
260   }
261   /* Now post the Irecvs corresponding to these messages */
262   PetscCall(PetscPostIrecvInt(comm, tag2, nrqs, onodes2, olengths2, &rbuf2, &r_waits2));
263 
264   /*  Now  post the sends */
265   PetscCall(PetscMalloc1(nrqr, &s_waits2));
266   for (PetscMPIInt i = 0; i < nrqr; ++i) {
267     PetscMPIInt j = onodes1[i];
268     PetscCallMPI(MPIU_Isend(xdata[i], isz1[i], MPIU_INT, j, tag2, comm, s_waits2 + i));
269   }
270 
271   PetscCall(PetscFree(onodes1));
272   PetscCall(PetscFree(olengths1));
273 
274   /* receive work done on other processors*/
275   {
276     PetscMPIInt idex;
277     PetscInt    is_no, ct1, max, *rbuf2_i, isz_i, *data_i, jmax;
278     PetscBT     table_i;
279 
280     for (PetscMPIInt i = 0; i < nrqs; ++i) {
281       PetscCallMPI(MPI_Waitany(nrqs, r_waits2, &idex, MPI_STATUS_IGNORE));
282       /* Process the message*/
283       rbuf2_i = rbuf2[idex];
284       ct1     = 2 * rbuf2_i[0] + 1;
285       jmax    = rbuf2[idex][0];
286       for (PetscInt j = 1; j <= jmax; j++) {
287         max     = rbuf2_i[2 * j];
288         is_no   = rbuf2_i[2 * j - 1];
289         isz_i   = isz[is_no];
290         data_i  = data[is_no];
291         table_i = table[is_no];
292         for (PetscInt k = 0; k < max; k++, ct1++) {
293           row = rbuf2_i[ct1];
294           if (!PetscBTLookupSet(table_i, row)) data_i[isz_i++] = row;
295         }
296         isz[is_no] = isz_i;
297       }
298     }
299     PetscCallMPI(MPI_Waitall(nrqr, s_waits2, MPI_STATUSES_IGNORE));
300   }
301 
302   for (PetscInt i = 0; i < imax; ++i) {
303     PetscCall(ISCreateGeneral(iscomms[i], isz[i], data[i], PETSC_COPY_VALUES, is + i));
304     PetscCall(PetscCommDestroy(&iscomms[i]));
305   }
306 
307   PetscCall(PetscFree(iscomms));
308   PetscCall(PetscFree(onodes2));
309   PetscCall(PetscFree(olengths2));
310 
311   PetscCall(PetscFree(pa));
312   if (rbuf2) {
313     PetscCall(PetscFree(rbuf2[0]));
314     PetscCall(PetscFree(rbuf2));
315   }
316   PetscCall(PetscFree(s_waits1));
317   PetscCall(PetscFree(r_waits1));
318   PetscCall(PetscFree(s_waits2));
319   PetscCall(PetscFree(r_waits2));
320   PetscCall(PetscFree5(table, data, isz, d_p, t_p));
321   if (xdata) {
322     PetscCall(PetscFree(xdata[0]));
323     PetscCall(PetscFree(xdata));
324   }
325   PetscCall(PetscFree(isz1));
326   PetscFunctionReturn(PETSC_SUCCESS);
327 }
328 
329 /*
330    MatIncreaseOverlap_MPIBAIJ_Local - Called by MatincreaseOverlap, to do
331        the work on the local processor.
332 
333      Inputs:
334       C      - MAT_MPIBAIJ;
335       imax - total no of index sets processed at a time;
336       table  - an array of char - size = Mbs bits.
337 
338      Output:
339       isz    - array containing the count of the solution elements corresponding
340                to each index set;
341       data   - pointer to the solutions
342 */
MatIncreaseOverlap_MPIBAIJ_Local(Mat C,PetscInt imax,PetscBT * table,PetscInt * isz,PetscInt ** data)343 static PetscErrorCode MatIncreaseOverlap_MPIBAIJ_Local(Mat C, PetscInt imax, PetscBT *table, PetscInt *isz, PetscInt **data)
344 {
345   Mat_MPIBAIJ *c = (Mat_MPIBAIJ *)C->data;
346   Mat          A = c->A, B = c->B;
347   Mat_SeqBAIJ *a = (Mat_SeqBAIJ *)A->data, *b = (Mat_SeqBAIJ *)B->data;
348   PetscInt     start, end, val, max, rstart, cstart, *ai, *aj;
349   PetscInt    *bi, *bj, *garray, i, j, k, row, *data_i, isz_i;
350   PetscBT      table_i;
351 
352   PetscFunctionBegin;
353   rstart = c->rstartbs;
354   cstart = c->cstartbs;
355   ai     = a->i;
356   aj     = a->j;
357   bi     = b->i;
358   bj     = b->j;
359   garray = c->garray;
360 
361   for (i = 0; i < imax; i++) {
362     data_i  = data[i];
363     table_i = table[i];
364     isz_i   = isz[i];
365     for (j = 0, max = isz[i]; j < max; j++) {
366       row   = data_i[j] - rstart;
367       start = ai[row];
368       end   = ai[row + 1];
369       for (k = start; k < end; k++) { /* Amat */
370         val = aj[k] + cstart;
371         if (!PetscBTLookupSet(table_i, val)) data_i[isz_i++] = val;
372       }
373       start = bi[row];
374       end   = bi[row + 1];
375       for (k = start; k < end; k++) { /* Bmat */
376         val = garray[bj[k]];
377         if (!PetscBTLookupSet(table_i, val)) data_i[isz_i++] = val;
378       }
379     }
380     isz[i] = isz_i;
381   }
382   PetscFunctionReturn(PETSC_SUCCESS);
383 }
384 
385 /*
386       MatIncreaseOverlap_MPIBAIJ_Receive - Process the received messages,
387          and return the output
388 
389          Input:
390            C    - the matrix
391            nrqr - no of messages being processed.
392            rbuf - an array of pointers to the received requests
393 
394          Output:
395            xdata - array of messages to be sent back
396            isz1  - size of each message
397 
398   For better efficiency perhaps we should malloc separately each xdata[i],
399 then if a remalloc is required we need only copy the data for that one row
400 rather than all previous rows as it is now where a single large chunk of
401 memory is used.
402 
403 */
MatIncreaseOverlap_MPIBAIJ_Receive(Mat C,PetscInt nrqr,PetscInt ** rbuf,PetscInt ** xdata,PetscInt * isz1)404 static PetscErrorCode MatIncreaseOverlap_MPIBAIJ_Receive(Mat C, PetscInt nrqr, PetscInt **rbuf, PetscInt **xdata, PetscInt *isz1)
405 {
406   Mat_MPIBAIJ *c = (Mat_MPIBAIJ *)C->data;
407   Mat          A = c->A, B = c->B;
408   Mat_SeqBAIJ *a = (Mat_SeqBAIJ *)A->data, *b = (Mat_SeqBAIJ *)B->data;
409   PetscInt     rstart, cstart, *ai, *aj, *bi, *bj, *garray, i, j, k;
410   PetscInt     row, total_sz, ct, ct1, ct2, ct3, mem_estimate, oct2, l, start, end;
411   PetscInt     val, max1, max2, Mbs, no_malloc = 0, *tmp, new_estimate, ctr;
412   PetscInt    *rbuf_i, kmax, rbuf_0;
413   PetscBT      xtable;
414 
415   PetscFunctionBegin;
416   Mbs    = c->Mbs;
417   rstart = c->rstartbs;
418   cstart = c->cstartbs;
419   ai     = a->i;
420   aj     = a->j;
421   bi     = b->i;
422   bj     = b->j;
423   garray = c->garray;
424 
425   for (i = 0, ct = 0, total_sz = 0; i < nrqr; ++i) {
426     rbuf_i = rbuf[i];
427     rbuf_0 = rbuf_i[0];
428     ct += rbuf_0;
429     for (j = 1; j <= rbuf_0; j++) total_sz += rbuf_i[2 * j];
430   }
431 
432   if (c->Mbs) max1 = ct * (a->nz + b->nz) / c->Mbs;
433   else max1 = 1;
434   mem_estimate = 3 * ((total_sz > max1 ? total_sz : max1) + 1);
435   if (nrqr) {
436     PetscCall(PetscMalloc1(mem_estimate, &xdata[0]));
437     ++no_malloc;
438   }
439   PetscCall(PetscBTCreate(Mbs, &xtable));
440   PetscCall(PetscArrayzero(isz1, nrqr));
441 
442   ct3 = 0;
443   for (i = 0; i < nrqr; i++) { /* for easch mesg from proc i */
444     rbuf_i = rbuf[i];
445     rbuf_0 = rbuf_i[0];
446     ct1    = 2 * rbuf_0 + 1;
447     ct2    = ct1;
448     ct3 += ct1;
449     for (j = 1; j <= rbuf_0; j++) { /* for each IS from proc i*/
450       PetscCall(PetscBTMemzero(Mbs, xtable));
451       oct2 = ct2;
452       kmax = rbuf_i[2 * j];
453       for (k = 0; k < kmax; k++, ct1++) {
454         row = rbuf_i[ct1];
455         if (!PetscBTLookupSet(xtable, row)) {
456           if (!(ct3 < mem_estimate)) {
457             new_estimate = (PetscInt)(1.5 * mem_estimate) + 1;
458             PetscCall(PetscMalloc1(new_estimate, &tmp));
459             PetscCall(PetscArraycpy(tmp, xdata[0], mem_estimate));
460             PetscCall(PetscFree(xdata[0]));
461             xdata[0]     = tmp;
462             mem_estimate = new_estimate;
463             ++no_malloc;
464             for (ctr = 1; ctr <= i; ctr++) xdata[ctr] = xdata[ctr - 1] + isz1[ctr - 1];
465           }
466           xdata[i][ct2++] = row;
467           ct3++;
468         }
469       }
470       for (k = oct2, max2 = ct2; k < max2; k++) {
471         row   = xdata[i][k] - rstart;
472         start = ai[row];
473         end   = ai[row + 1];
474         for (l = start; l < end; l++) {
475           val = aj[l] + cstart;
476           if (!PetscBTLookupSet(xtable, val)) {
477             if (!(ct3 < mem_estimate)) {
478               new_estimate = (PetscInt)(1.5 * mem_estimate) + 1;
479               PetscCall(PetscMalloc1(new_estimate, &tmp));
480               PetscCall(PetscArraycpy(tmp, xdata[0], mem_estimate));
481               PetscCall(PetscFree(xdata[0]));
482               xdata[0]     = tmp;
483               mem_estimate = new_estimate;
484               ++no_malloc;
485               for (ctr = 1; ctr <= i; ctr++) xdata[ctr] = xdata[ctr - 1] + isz1[ctr - 1];
486             }
487             xdata[i][ct2++] = val;
488             ct3++;
489           }
490         }
491         start = bi[row];
492         end   = bi[row + 1];
493         for (l = start; l < end; l++) {
494           val = garray[bj[l]];
495           if (!PetscBTLookupSet(xtable, val)) {
496             if (!(ct3 < mem_estimate)) {
497               new_estimate = (PetscInt)(1.5 * mem_estimate) + 1;
498               PetscCall(PetscMalloc1(new_estimate, &tmp));
499               PetscCall(PetscArraycpy(tmp, xdata[0], mem_estimate));
500               PetscCall(PetscFree(xdata[0]));
501               xdata[0]     = tmp;
502               mem_estimate = new_estimate;
503               ++no_malloc;
504               for (ctr = 1; ctr <= i; ctr++) xdata[ctr] = xdata[ctr - 1] + isz1[ctr - 1];
505             }
506             xdata[i][ct2++] = val;
507             ct3++;
508           }
509         }
510       }
511       /* Update the header*/
512       xdata[i][2 * j]     = ct2 - oct2; /* Undo the vector isz1 and use only a var*/
513       xdata[i][2 * j - 1] = rbuf_i[2 * j - 1];
514     }
515     xdata[i][0] = rbuf_0;
516     if (i + 1 < nrqr) xdata[i + 1] = xdata[i] + ct2;
517     isz1[i] = ct2; /* size of each message */
518   }
519   PetscCall(PetscBTDestroy(&xtable));
520   PetscCall(PetscInfo(C, "Allocated %" PetscInt_FMT " bytes, required %" PetscInt_FMT ", no of mallocs = %" PetscInt_FMT "\n", mem_estimate, ct3, no_malloc));
521   PetscFunctionReturn(PETSC_SUCCESS);
522 }
523 
MatCreateSubMatrices_MPIBAIJ(Mat C,PetscInt ismax,const IS isrow[],const IS iscol[],MatReuse scall,Mat * submat[])524 PetscErrorCode MatCreateSubMatrices_MPIBAIJ(Mat C, PetscInt ismax, const IS isrow[], const IS iscol[], MatReuse scall, Mat *submat[])
525 {
526   IS          *isrow_block, *iscol_block;
527   Mat_MPIBAIJ *c = (Mat_MPIBAIJ *)C->data;
528   PetscInt     nmax, nstages, i, pos, max_no, N = C->cmap->N, bs = C->rmap->bs;
529   Mat_SeqBAIJ *subc;
530   Mat_SubSppt *smat;
531   PetscBool    sym = PETSC_FALSE, flg[2];
532 
533   PetscFunctionBegin;
534   PetscCall(PetscObjectTypeCompare((PetscObject)C, MATMPISBAIJ, flg));
535   if (flg[0]) {
536     if (isrow == iscol) sym = PETSC_TRUE;
537     else {
538       flg[0] = flg[1] = PETSC_TRUE;
539       for (i = 0; i < ismax; i++) {
540         if (isrow[i] != iscol[i]) flg[0] = PETSC_FALSE;
541         PetscCall(ISGetLocalSize(iscol[0], &nmax));
542         if (nmax == C->cmap->N && flg[1]) PetscCall(ISIdentity(iscol[0], flg + 1));
543       }
544       sym = (PetscBool)(flg[0] || flg[1]);
545     }
546   }
547   /* The compression and expansion should be avoided. Doesn't point
548      out errors, might change the indices, hence buggey */
549   PetscCall(PetscMalloc2(ismax, &isrow_block, ismax, &iscol_block));
550   PetscCall(ISCompressIndicesGeneral(C->rmap->N, C->rmap->n, bs, ismax, isrow, isrow_block));
551   if (isrow == iscol) {
552     for (i = 0; i < ismax; i++) {
553       iscol_block[i] = isrow_block[i];
554       PetscCall(PetscObjectReference((PetscObject)iscol_block[i]));
555     }
556   } else PetscCall(ISCompressIndicesGeneral(N, C->cmap->n, bs, ismax, iscol, iscol_block));
557 
558   /* Determine the number of stages through which submatrices are done */
559   if (!C->cmap->N) nmax = 20 * 1000000 / sizeof(PetscInt);
560   else nmax = 20 * 1000000 / (c->Nbs * sizeof(PetscInt));
561   if (!nmax) nmax = 1;
562 
563   if (scall == MAT_INITIAL_MATRIX) {
564     nstages = ismax / nmax + ((ismax % nmax) ? 1 : 0); /* local nstages */
565 
566     /* Make sure every processor loops through the nstages */
567     PetscCallMPI(MPIU_Allreduce(MPI_IN_PLACE, &nstages, 1, MPIU_INT, MPI_MAX, PetscObjectComm((PetscObject)C)));
568 
569     /* Allocate memory to hold all the submatrices and dummy submatrices */
570     PetscCall(PetscCalloc1(ismax + nstages, submat));
571   } else { /* MAT_REUSE_MATRIX */
572     if (ismax) {
573       subc = (Mat_SeqBAIJ *)((*submat)[0]->data);
574       smat = subc->submatis1;
575     } else { /* (*submat)[0] is a dummy matrix */
576       smat = (Mat_SubSppt *)(*submat)[0]->data;
577     }
578     PetscCheck(smat, PETSC_COMM_SELF, PETSC_ERR_ARG_NULL, "MatCreateSubMatrices(...,MAT_REUSE_MATRIX,...) requires submat");
579     nstages = smat->nstages;
580   }
581 
582   for (i = 0, pos = 0; i < nstages; i++) {
583     if (pos + nmax <= ismax) max_no = nmax;
584     else if (pos >= ismax) max_no = 0;
585     else max_no = ismax - pos;
586 
587     PetscCall(MatCreateSubMatrices_MPIBAIJ_local(C, max_no, isrow_block + pos, iscol_block + pos, scall, *submat + pos, sym));
588     if (!max_no) {
589       if (scall == MAT_INITIAL_MATRIX) { /* submat[pos] is a dummy matrix */
590         smat          = (Mat_SubSppt *)(*submat)[pos]->data;
591         smat->nstages = nstages;
592       }
593       pos++; /* advance to next dummy matrix if any */
594     } else pos += max_no;
595   }
596 
597   if (scall == MAT_INITIAL_MATRIX && ismax) {
598     /* save nstages for reuse */
599     subc          = (Mat_SeqBAIJ *)((*submat)[0]->data);
600     smat          = subc->submatis1;
601     smat->nstages = nstages;
602   }
603 
604   for (i = 0; i < ismax; i++) {
605     PetscCall(ISDestroy(&isrow_block[i]));
606     PetscCall(ISDestroy(&iscol_block[i]));
607   }
608   PetscCall(PetscFree2(isrow_block, iscol_block));
609   PetscFunctionReturn(PETSC_SUCCESS);
610 }
611 
612 /* This code is used for BAIJ and SBAIJ matrices (unfortunate dependency) */
MatCreateSubMatrices_MPIBAIJ_local(Mat C,PetscInt ismax,const IS isrow[],const IS iscol[],MatReuse scall,Mat * submats,PetscBool sym)613 PetscErrorCode MatCreateSubMatrices_MPIBAIJ_local(Mat C, PetscInt ismax, const IS isrow[], const IS iscol[], MatReuse scall, Mat *submats, PetscBool sym)
614 {
615   Mat_MPIBAIJ     *c = (Mat_MPIBAIJ *)C->data;
616   Mat              A = c->A;
617   Mat_SeqBAIJ     *a = (Mat_SeqBAIJ *)A->data, *b = (Mat_SeqBAIJ *)c->B->data, *subc;
618   const PetscInt **icol, **irow;
619   PetscInt        *nrow, *ncol, start;
620   PetscMPIInt     *pa, **row2proc, *row2proc_i, proc = -1;
621   PetscMPIInt      rank, size, tag0, tag2, tag3, tag4, *w1, *w2, *w3, *w4, nrqr, nrqs = 0, *req_source1 = NULL, *req_source2;
622   PetscInt       **sbuf1, **sbuf2, *sbuf2_i, ct1, ct2, **rbuf1, row;
623   PetscInt         msz, **ptr = NULL, *req_size = NULL, *ctr = NULL, *tmp = NULL, tcol;
624   PetscInt       **rbuf3 = NULL, **sbuf_aj, **rbuf2 = NULL, max1, max2;
625   PetscInt       **lens, is_no, ncols, *cols, mat_i, *mat_j, tmp2, jmax;
626 #if defined(PETSC_USE_CTABLE)
627   PetscHMapI *cmap, cmap_i = NULL, *rmap, rmap_i;
628 #else
629   PetscInt **cmap, *cmap_i = NULL, **rmap, *rmap_i;
630 #endif
631   const PetscInt *irow_i, *icol_i;
632   PetscInt        ctr_j, *sbuf1_j, *sbuf_aj_i, *rbuf1_i, kmax, *lens_i, jcnt;
633   MPI_Request    *s_waits1, *r_waits1, *s_waits2, *r_waits2, *r_waits3;
634   MPI_Request    *r_waits4, *s_waits3, *s_waits4;
635   MPI_Comm        comm;
636   PetscScalar   **rbuf4, *rbuf4_i = NULL, **sbuf_aa, *vals, *mat_a = NULL, *imat_a = NULL, *sbuf_aa_i;
637   PetscMPIInt    *onodes1, *olengths1, end;
638   PetscInt       *imat_ilen, *imat_j, *imat_i;
639   Mat_SubSppt    *smat_i;
640   PetscBool      *issorted, colflag, iscsorted = PETSC_TRUE;
641   PetscInt       *sbuf1_i, *rbuf2_i, *rbuf3_i, ilen;
642   PetscInt        bs = C->rmap->bs, bs2 = c->bs2, rstart = c->rstartbs;
643   PetscBool       ijonly = c->ijonly; /* private flag indicates only matrix data structures are requested */
644   PetscInt        nzA, nzB, *a_i = a->i, *b_i = b->i, *a_j = a->j, *b_j = b->j, ctmp, imark, *cworkA, *cworkB;
645   PetscScalar    *vworkA = NULL, *vworkB = NULL, *a_a = a->a, *b_a = b->a;
646   PetscInt        cstart = c->cstartbs, *bmap = c->garray;
647   PetscBool      *allrows, *allcolumns;
648 
649   PetscFunctionBegin;
650   PetscCall(PetscObjectGetComm((PetscObject)C, &comm));
651   size = c->size;
652   rank = c->rank;
653 
654   PetscCall(PetscMalloc5(ismax, &row2proc, ismax, &cmap, ismax, &rmap, ismax + 1, &allcolumns, ismax, &allrows));
655   PetscCall(PetscMalloc5(ismax, (PetscInt ***)&irow, ismax, (PetscInt ***)&icol, ismax, &nrow, ismax, &ncol, ismax, &issorted));
656 
657   for (PetscInt i = 0; i < ismax; i++) {
658     PetscCall(ISSorted(iscol[i], &issorted[i]));
659     if (!issorted[i]) iscsorted = issorted[i]; /* columns are not sorted! */
660     PetscCall(ISSorted(isrow[i], &issorted[i]));
661 
662     /* Check for special case: allcolumns */
663     PetscCall(ISIdentity(iscol[i], &colflag));
664     PetscCall(ISGetLocalSize(iscol[i], &ncol[i]));
665 
666     if (colflag && ncol[i] == c->Nbs) {
667       allcolumns[i] = PETSC_TRUE;
668       icol[i]       = NULL;
669     } else {
670       allcolumns[i] = PETSC_FALSE;
671       PetscCall(ISGetIndices(iscol[i], &icol[i]));
672     }
673 
674     /* Check for special case: allrows */
675     PetscCall(ISIdentity(isrow[i], &colflag));
676     PetscCall(ISGetLocalSize(isrow[i], &nrow[i]));
677     if (colflag && nrow[i] == c->Mbs) {
678       allrows[i] = PETSC_TRUE;
679       irow[i]    = NULL;
680     } else {
681       allrows[i] = PETSC_FALSE;
682       PetscCall(ISGetIndices(isrow[i], &irow[i]));
683     }
684   }
685 
686   if (scall == MAT_REUSE_MATRIX) {
687     /* Assumes new rows are same length as the old rows */
688     for (PetscInt i = 0; i < ismax; i++) {
689       subc = (Mat_SeqBAIJ *)submats[i]->data;
690 
691       PetscCheck(subc->mbs == nrow[i] && subc->nbs == ncol[i], PETSC_COMM_SELF, PETSC_ERR_ARG_SIZ, "Cannot reuse matrix. wrong size");
692 
693       /* Initial matrix as if empty */
694       PetscCall(PetscArrayzero(subc->ilen, subc->mbs));
695 
696       /* Initial matrix as if empty */
697       submats[i]->factortype = C->factortype;
698 
699       smat_i = subc->submatis1;
700 
701       nrqs        = smat_i->nrqs;
702       nrqr        = smat_i->nrqr;
703       rbuf1       = smat_i->rbuf1;
704       rbuf2       = smat_i->rbuf2;
705       rbuf3       = smat_i->rbuf3;
706       req_source2 = smat_i->req_source2;
707 
708       sbuf1 = smat_i->sbuf1;
709       sbuf2 = smat_i->sbuf2;
710       ptr   = smat_i->ptr;
711       tmp   = smat_i->tmp;
712       ctr   = smat_i->ctr;
713 
714       pa          = smat_i->pa;
715       req_size    = smat_i->req_size;
716       req_source1 = smat_i->req_source1;
717 
718       allcolumns[i] = smat_i->allcolumns;
719       allrows[i]    = smat_i->allrows;
720       row2proc[i]   = smat_i->row2proc;
721       rmap[i]       = smat_i->rmap;
722       cmap[i]       = smat_i->cmap;
723     }
724 
725     if (!ismax) { /* Get dummy submatrices and retrieve struct submatis1 */
726       PetscCheck(submats[0], PETSC_COMM_SELF, PETSC_ERR_ARG_NULL, "submats are null, cannot reuse");
727       smat_i = (Mat_SubSppt *)submats[0]->data;
728 
729       nrqs        = smat_i->nrqs;
730       nrqr        = smat_i->nrqr;
731       rbuf1       = smat_i->rbuf1;
732       rbuf2       = smat_i->rbuf2;
733       rbuf3       = smat_i->rbuf3;
734       req_source2 = smat_i->req_source2;
735 
736       sbuf1 = smat_i->sbuf1;
737       sbuf2 = smat_i->sbuf2;
738       ptr   = smat_i->ptr;
739       tmp   = smat_i->tmp;
740       ctr   = smat_i->ctr;
741 
742       pa          = smat_i->pa;
743       req_size    = smat_i->req_size;
744       req_source1 = smat_i->req_source1;
745 
746       allcolumns[0] = PETSC_FALSE;
747     }
748   } else { /* scall == MAT_INITIAL_MATRIX */
749     /* Get some new tags to keep the communication clean */
750     PetscCall(PetscObjectGetNewTag((PetscObject)C, &tag2));
751     PetscCall(PetscObjectGetNewTag((PetscObject)C, &tag3));
752 
753     /* evaluate communication - mesg to who, length of mesg, and buffer space
754      required. Based on this, buffers are allocated, and data copied into them*/
755     PetscCall(PetscCalloc4(size, &w1, size, &w2, size, &w3, size, &w4)); /* mesg size, initialize work vectors */
756 
757     for (PetscInt i = 0; i < ismax; i++) {
758       jmax   = nrow[i];
759       irow_i = irow[i];
760 
761       PetscCall(PetscMalloc1(jmax, &row2proc_i));
762       row2proc[i] = row2proc_i;
763 
764       if (issorted[i]) proc = 0;
765       for (PetscInt j = 0; j < jmax; j++) {
766         if (!issorted[i]) proc = 0;
767         if (allrows[i]) row = j;
768         else row = irow_i[j];
769 
770         while (row >= c->rangebs[proc + 1]) proc++;
771         w4[proc]++;
772         row2proc_i[j] = proc; /* map row index to proc */
773       }
774       for (PetscMPIInt j = 0; j < size; j++) {
775         if (w4[j]) {
776           w1[j] += w4[j];
777           w3[j]++;
778           w4[j] = 0;
779         }
780       }
781     }
782 
783     nrqs     = 0; /* no of outgoing messages */
784     msz      = 0; /* total mesg length (for all procs) */
785     w1[rank] = 0; /* no mesg sent to self */
786     w3[rank] = 0;
787     for (PetscMPIInt i = 0; i < size; i++) {
788       if (w1[i]) {
789         w2[i] = 1;
790         nrqs++;
791       } /* there exists a message to proc i */
792     }
793     PetscCall(PetscMalloc1(nrqs, &pa)); /*(proc -array)*/
794     for (PetscMPIInt i = 0, j = 0; i < size; i++) {
795       if (w1[i]) pa[j++] = i;
796     }
797 
798     /* Each message would have a header = 1 + 2*(no of IS) + data */
799     for (PetscMPIInt i = 0; i < nrqs; i++) {
800       PetscMPIInt j = pa[i];
801       w1[j] += w2[j] + 2 * w3[j];
802       msz += w1[j];
803     }
804     PetscCall(PetscInfo(0, "Number of outgoing messages %d Total message length %" PetscInt_FMT "\n", nrqs, msz));
805 
806     /* Determine the number of messages to expect, their lengths, from from-ids */
807     PetscCall(PetscGatherNumberOfMessages(comm, w2, w1, &nrqr));
808     PetscCall(PetscGatherMessageLengths(comm, nrqs, nrqr, w1, &onodes1, &olengths1));
809 
810     /* Now post the Irecvs corresponding to these messages */
811     PetscCall(PetscObjectGetNewTag((PetscObject)C, &tag0));
812     PetscCall(PetscPostIrecvInt(comm, tag0, nrqr, onodes1, olengths1, &rbuf1, &r_waits1));
813 
814     /* Allocate Memory for outgoing messages */
815     PetscCall(PetscMalloc4(size, &sbuf1, size, &ptr, 2 * msz, &tmp, size, &ctr));
816     PetscCall(PetscArrayzero(sbuf1, size));
817     PetscCall(PetscArrayzero(ptr, size));
818 
819     {
820       PetscInt   *iptr = tmp;
821       PetscMPIInt k    = 0;
822       for (PetscMPIInt i = 0; i < nrqs; i++) {
823         PetscMPIInt j = pa[i];
824         iptr += k;
825         sbuf1[j] = iptr;
826         k        = w1[j];
827       }
828     }
829 
830     /* Form the outgoing messages. Initialize the header space */
831     for (PetscMPIInt i = 0; i < nrqs; i++) {
832       PetscMPIInt j = pa[i];
833 
834       sbuf1[j][0] = 0;
835       PetscCall(PetscArrayzero(sbuf1[j] + 1, 2 * w3[j]));
836       ptr[j] = sbuf1[j] + 2 * w3[j] + 1;
837     }
838 
839     /* Parse the isrow and copy data into outbuf */
840     for (PetscInt i = 0; i < ismax; i++) {
841       row2proc_i = row2proc[i];
842       PetscCall(PetscArrayzero(ctr, size));
843       irow_i = irow[i];
844       jmax   = nrow[i];
845       for (PetscInt j = 0; j < jmax; j++) { /* parse the indices of each IS */
846         proc = row2proc_i[j];
847         if (allrows[i]) row = j;
848         else row = irow_i[j];
849 
850         if (proc != rank) { /* copy to the outgoing buf*/
851           ctr[proc]++;
852           *ptr[proc] = row;
853           ptr[proc]++;
854         }
855       }
856       /* Update the headers for the current IS */
857       for (PetscMPIInt j = 0; j < size; j++) { /* Can Optimise this loop too */
858         if ((ctr_j = ctr[j])) {
859           PetscInt k;
860 
861           sbuf1_j            = sbuf1[j];
862           k                  = ++sbuf1_j[0];
863           sbuf1_j[2 * k]     = ctr_j;
864           sbuf1_j[2 * k - 1] = i;
865         }
866       }
867     }
868 
869     /*  Now  post the sends */
870     PetscCall(PetscMalloc1(nrqs, &s_waits1));
871     for (PetscMPIInt i = 0; i < nrqs; ++i) {
872       PetscMPIInt j = pa[i];
873       PetscCallMPI(MPIU_Isend(sbuf1[j], w1[j], MPIU_INT, j, tag0, comm, s_waits1 + i));
874     }
875 
876     /* Post Receives to capture the buffer size */
877     PetscCall(PetscMalloc1(nrqs, &r_waits2));
878     PetscCall(PetscMalloc3(nrqs, &req_source2, nrqs, &rbuf2, nrqs, &rbuf3));
879     if (nrqs) rbuf2[0] = tmp + msz;
880     for (PetscMPIInt i = 1; i < nrqs; ++i) rbuf2[i] = rbuf2[i - 1] + w1[pa[i - 1]];
881     for (PetscMPIInt i = 0; i < nrqs; ++i) {
882       PetscMPIInt j = pa[i];
883       PetscCallMPI(MPIU_Irecv(rbuf2[i], w1[j], MPIU_INT, j, tag2, comm, r_waits2 + i));
884     }
885 
886     /* Send to other procs the buf size they should allocate */
887     /* Receive messages*/
888     PetscCall(PetscMalloc1(nrqr, &s_waits2));
889     PetscCall(PetscMalloc3(nrqr, &sbuf2, nrqr, &req_size, nrqr, &req_source1));
890 
891     PetscCallMPI(MPI_Waitall(nrqr, r_waits1, MPI_STATUSES_IGNORE));
892     for (PetscMPIInt i = 0; i < nrqr; ++i) {
893       req_size[i] = 0;
894       rbuf1_i     = rbuf1[i];
895       start       = 2 * rbuf1_i[0] + 1;
896       end         = olengths1[i];
897       PetscCall(PetscMalloc1(end, &sbuf2[i]));
898       sbuf2_i = sbuf2[i];
899       for (PetscInt j = start; j < end; j++) {
900         row        = rbuf1_i[j] - rstart;
901         ncols      = a_i[row + 1] - a_i[row] + b_i[row + 1] - b_i[row];
902         sbuf2_i[j] = ncols;
903         req_size[i] += ncols;
904       }
905       req_source1[i] = onodes1[i];
906       /* form the header */
907       sbuf2_i[0] = req_size[i];
908       for (PetscInt j = 1; j < start; j++) sbuf2_i[j] = rbuf1_i[j];
909 
910       PetscCallMPI(MPIU_Isend(sbuf2_i, end, MPIU_INT, req_source1[i], tag2, comm, s_waits2 + i));
911     }
912     PetscCall(PetscFree(onodes1));
913     PetscCall(PetscFree(olengths1));
914 
915     PetscCall(PetscFree(r_waits1));
916     PetscCall(PetscFree4(w1, w2, w3, w4));
917 
918     /* Receive messages*/
919     PetscCall(PetscMalloc1(nrqs, &r_waits3));
920 
921     PetscCallMPI(MPI_Waitall(nrqs, r_waits2, MPI_STATUSES_IGNORE));
922     for (PetscMPIInt i = 0; i < nrqs; ++i) {
923       PetscCall(PetscMalloc1(rbuf2[i][0], &rbuf3[i]));
924       req_source2[i] = pa[i];
925       PetscCallMPI(MPIU_Irecv(rbuf3[i], rbuf2[i][0], MPIU_INT, req_source2[i], tag3, comm, r_waits3 + i));
926     }
927     PetscCall(PetscFree(r_waits2));
928 
929     /* Wait on sends1 and sends2 */
930     PetscCallMPI(MPI_Waitall(nrqs, s_waits1, MPI_STATUSES_IGNORE));
931     PetscCallMPI(MPI_Waitall(nrqr, s_waits2, MPI_STATUSES_IGNORE));
932     PetscCall(PetscFree(s_waits1));
933     PetscCall(PetscFree(s_waits2));
934 
935     /* Now allocate sending buffers for a->j, and send them off */
936     PetscCall(PetscMalloc1(nrqr, &sbuf_aj));
937     jcnt = 0;
938     for (PetscMPIInt i = 0; i < nrqr; i++) jcnt += req_size[i];
939     if (nrqr) PetscCall(PetscMalloc1(jcnt, &sbuf_aj[0]));
940     for (PetscMPIInt i = 1; i < nrqr; i++) sbuf_aj[i] = sbuf_aj[i - 1] + req_size[i - 1];
941 
942     PetscCall(PetscMalloc1(nrqr, &s_waits3));
943     {
944       for (PetscMPIInt i = 0; i < nrqr; i++) {
945         rbuf1_i   = rbuf1[i];
946         sbuf_aj_i = sbuf_aj[i];
947         ct1       = 2 * rbuf1_i[0] + 1;
948         ct2       = 0;
949         for (PetscInt j = 1, max1 = rbuf1_i[0]; j <= max1; j++) {
950           kmax = rbuf1[i][2 * j];
951           for (PetscInt k = 0; k < kmax; k++, ct1++) {
952             PetscInt l;
953             row    = rbuf1_i[ct1] - rstart;
954             nzA    = a_i[row + 1] - a_i[row];
955             nzB    = b_i[row + 1] - b_i[row];
956             ncols  = nzA + nzB;
957             cworkA = PetscSafePointerPlusOffset(a_j, a_i[row]);
958             cworkB = PetscSafePointerPlusOffset(b_j, b_i[row]);
959 
960             /* load the column indices for this row into cols */
961             cols = sbuf_aj_i + ct2;
962             for (l = 0; l < nzB; l++) {
963               if ((ctmp = bmap[cworkB[l]]) < cstart) cols[l] = ctmp;
964               else break;
965             }
966             imark = l;
967             for (l = 0; l < nzA; l++) cols[imark + l] = cstart + cworkA[l];
968             for (l = imark; l < nzB; l++) cols[nzA + l] = bmap[cworkB[l]];
969             ct2 += ncols;
970           }
971         }
972         PetscCallMPI(MPIU_Isend(sbuf_aj_i, req_size[i], MPIU_INT, req_source1[i], tag3, comm, s_waits3 + i));
973       }
974     }
975 
976     /* create col map: global col of C -> local col of submatrices */
977 #if defined(PETSC_USE_CTABLE)
978     for (PetscInt i = 0; i < ismax; i++) {
979       if (!allcolumns[i]) {
980         PetscCall(PetscHMapICreateWithSize(ncol[i], cmap + i));
981 
982         jmax   = ncol[i];
983         icol_i = icol[i];
984         cmap_i = cmap[i];
985         for (PetscInt j = 0; j < jmax; j++) PetscCall(PetscHMapISet(cmap[i], icol_i[j] + 1, j + 1));
986       } else cmap[i] = NULL;
987     }
988 #else
989     for (PetscInt i = 0; i < ismax; i++) {
990       if (!allcolumns[i]) {
991         PetscCall(PetscCalloc1(c->Nbs, &cmap[i]));
992         jmax   = ncol[i];
993         icol_i = icol[i];
994         cmap_i = cmap[i];
995         for (PetscInt j = 0; j < jmax; j++) cmap_i[icol_i[j]] = j + 1;
996       } else cmap[i] = NULL;
997     }
998 #endif
999 
1000     /* Create lens which is required for MatCreate... */
1001     jcnt = 0;
1002     for (PetscInt i = 0; i < ismax; i++) jcnt += nrow[i];
1003     PetscCall(PetscMalloc1(ismax, &lens));
1004     if (ismax) PetscCall(PetscCalloc1(jcnt, &lens[0]));
1005     for (PetscInt i = 1; i < ismax; i++) lens[i] = PetscSafePointerPlusOffset(lens[i - 1], nrow[i - 1]);
1006 
1007     /* Update lens from local data */
1008     for (PetscInt i = 0; i < ismax; i++) {
1009       row2proc_i = row2proc[i];
1010       jmax       = nrow[i];
1011       if (!allcolumns[i]) cmap_i = cmap[i];
1012       irow_i = irow[i];
1013       lens_i = lens[i];
1014       for (PetscInt j = 0; j < jmax; j++) {
1015         if (allrows[i]) row = j;
1016         else row = irow_i[j]; /* global blocked row of C */
1017 
1018         proc = row2proc_i[j];
1019         if (proc == rank) {
1020           /* Get indices from matA and then from matB */
1021 #if defined(PETSC_USE_CTABLE)
1022           PetscInt tt;
1023 #endif
1024           row    = row - rstart;
1025           nzA    = a_i[row + 1] - a_i[row];
1026           nzB    = b_i[row + 1] - b_i[row];
1027           cworkA = a_j + a_i[row];
1028           cworkB = PetscSafePointerPlusOffset(b_j, b_i[row]);
1029 
1030           if (!allcolumns[i]) {
1031 #if defined(PETSC_USE_CTABLE)
1032             for (PetscInt k = 0; k < nzA; k++) {
1033               PetscCall(PetscHMapIGetWithDefault(cmap_i, cstart + cworkA[k] + 1, 0, &tt));
1034               if (tt) lens_i[j]++;
1035             }
1036             for (PetscInt k = 0; k < nzB; k++) {
1037               PetscCall(PetscHMapIGetWithDefault(cmap_i, bmap[cworkB[k]] + 1, 0, &tt));
1038               if (tt) lens_i[j]++;
1039             }
1040 
1041 #else
1042             for (PetscInt k = 0; k < nzA; k++) {
1043               if (cmap_i[cstart + cworkA[k]]) lens_i[j]++;
1044             }
1045             for (PetscInt k = 0; k < nzB; k++) {
1046               if (cmap_i[bmap[cworkB[k]]]) lens_i[j]++;
1047             }
1048 #endif
1049           } else { /* allcolumns */
1050             lens_i[j] = nzA + nzB;
1051           }
1052         }
1053       }
1054     }
1055 
1056     /* Create row map: global row of C -> local row of submatrices */
1057     for (PetscInt i = 0; i < ismax; i++) {
1058       if (!allrows[i]) {
1059 #if defined(PETSC_USE_CTABLE)
1060         PetscCall(PetscHMapICreateWithSize(nrow[i], rmap + i));
1061         irow_i = irow[i];
1062         jmax   = nrow[i];
1063         for (PetscInt j = 0; j < jmax; j++) {
1064           if (allrows[i]) {
1065             PetscCall(PetscHMapISet(rmap[i], j + 1, j + 1));
1066           } else {
1067             PetscCall(PetscHMapISet(rmap[i], irow_i[j] + 1, j + 1));
1068           }
1069         }
1070 #else
1071         PetscCall(PetscCalloc1(c->Mbs, &rmap[i]));
1072         rmap_i = rmap[i];
1073         irow_i = irow[i];
1074         jmax   = nrow[i];
1075         for (PetscInt j = 0; j < jmax; j++) {
1076           if (allrows[i]) rmap_i[j] = j;
1077           else rmap_i[irow_i[j]] = j;
1078         }
1079 #endif
1080       } else rmap[i] = NULL;
1081     }
1082 
1083     /* Update lens from offproc data */
1084     {
1085       PetscInt *rbuf2_i, *rbuf3_i, *sbuf1_i;
1086 
1087       PetscCallMPI(MPI_Waitall(nrqs, r_waits3, MPI_STATUSES_IGNORE));
1088       for (tmp2 = 0; tmp2 < nrqs; tmp2++) {
1089         sbuf1_i = sbuf1[pa[tmp2]];
1090         jmax    = sbuf1_i[0];
1091         ct1     = 2 * jmax + 1;
1092         ct2     = 0;
1093         rbuf2_i = rbuf2[tmp2];
1094         rbuf3_i = rbuf3[tmp2];
1095         for (PetscInt j = 1; j <= jmax; j++) {
1096           is_no  = sbuf1_i[2 * j - 1];
1097           max1   = sbuf1_i[2 * j];
1098           lens_i = lens[is_no];
1099           if (!allcolumns[is_no]) cmap_i = cmap[is_no];
1100           rmap_i = rmap[is_no];
1101           for (PetscInt k = 0; k < max1; k++, ct1++) {
1102             if (allrows[is_no]) {
1103               row = sbuf1_i[ct1];
1104             } else {
1105 #if defined(PETSC_USE_CTABLE)
1106               PetscCall(PetscHMapIGetWithDefault(rmap_i, sbuf1_i[ct1] + 1, 0, &row));
1107               row--;
1108               PetscCheck(row >= 0, PETSC_COMM_SELF, PETSC_ERR_PLIB, "row not found in table");
1109 #else
1110               row = rmap_i[sbuf1_i[ct1]]; /* the val in the new matrix to be */
1111 #endif
1112             }
1113             max2 = rbuf2_i[ct1];
1114             for (PetscInt l = 0; l < max2; l++, ct2++) {
1115               if (!allcolumns[is_no]) {
1116 #if defined(PETSC_USE_CTABLE)
1117                 PetscCall(PetscHMapIGetWithDefault(cmap_i, rbuf3_i[ct2] + 1, 0, &tcol));
1118 #else
1119                 tcol = cmap_i[rbuf3_i[ct2]];
1120 #endif
1121                 if (tcol) lens_i[row]++;
1122               } else {         /* allcolumns */
1123                 lens_i[row]++; /* lens_i[row] += max2 ? */
1124               }
1125             }
1126           }
1127         }
1128       }
1129     }
1130     PetscCall(PetscFree(r_waits3));
1131     PetscCallMPI(MPI_Waitall(nrqr, s_waits3, MPI_STATUSES_IGNORE));
1132     PetscCall(PetscFree(s_waits3));
1133 
1134     /* Create the submatrices */
1135     for (PetscInt i = 0; i < ismax; i++) {
1136       PetscInt bs_tmp;
1137       if (ijonly) bs_tmp = 1;
1138       else bs_tmp = bs;
1139 
1140       PetscCall(MatCreate(PETSC_COMM_SELF, submats + i));
1141       PetscCall(MatSetSizes(submats[i], nrow[i] * bs_tmp, ncol[i] * bs_tmp, PETSC_DETERMINE, PETSC_DETERMINE));
1142 
1143       PetscCall(MatSetType(submats[i], sym ? ((PetscObject)A)->type_name : MATSEQBAIJ));
1144       PetscCall(MatSeqBAIJSetPreallocation(submats[i], bs_tmp, 0, lens[i]));
1145       PetscCall(MatSeqSBAIJSetPreallocation(submats[i], bs_tmp, 0, lens[i])); /* this subroutine is used by SBAIJ routines */
1146 
1147       /* create struct Mat_SubSppt and attached it to submat */
1148       PetscCall(PetscNew(&smat_i));
1149       subc            = (Mat_SeqBAIJ *)submats[i]->data;
1150       subc->submatis1 = smat_i;
1151 
1152       smat_i->destroy          = submats[i]->ops->destroy;
1153       submats[i]->ops->destroy = MatDestroySubMatrix_SeqBAIJ;
1154       submats[i]->factortype   = C->factortype;
1155 
1156       smat_i->id          = i;
1157       smat_i->nrqs        = nrqs;
1158       smat_i->nrqr        = nrqr;
1159       smat_i->rbuf1       = rbuf1;
1160       smat_i->rbuf2       = rbuf2;
1161       smat_i->rbuf3       = rbuf3;
1162       smat_i->sbuf2       = sbuf2;
1163       smat_i->req_source2 = req_source2;
1164 
1165       smat_i->sbuf1 = sbuf1;
1166       smat_i->ptr   = ptr;
1167       smat_i->tmp   = tmp;
1168       smat_i->ctr   = ctr;
1169 
1170       smat_i->pa          = pa;
1171       smat_i->req_size    = req_size;
1172       smat_i->req_source1 = req_source1;
1173 
1174       smat_i->allcolumns = allcolumns[i];
1175       smat_i->allrows    = allrows[i];
1176       smat_i->singleis   = PETSC_FALSE;
1177       smat_i->row2proc   = row2proc[i];
1178       smat_i->rmap       = rmap[i];
1179       smat_i->cmap       = cmap[i];
1180     }
1181 
1182     if (!ismax) { /* Create dummy submats[0] for reuse struct subc */
1183       PetscCall(MatCreate(PETSC_COMM_SELF, &submats[0]));
1184       PetscCall(MatSetSizes(submats[0], 0, 0, PETSC_DETERMINE, PETSC_DETERMINE));
1185       PetscCall(MatSetType(submats[0], MATDUMMY));
1186 
1187       /* create struct Mat_SubSppt and attached it to submat */
1188       PetscCall(PetscNew(&smat_i));
1189       submats[0]->data = (void *)smat_i;
1190 
1191       smat_i->destroy          = submats[0]->ops->destroy;
1192       submats[0]->ops->destroy = MatDestroySubMatrix_Dummy;
1193       submats[0]->factortype   = C->factortype;
1194 
1195       smat_i->id          = 0;
1196       smat_i->nrqs        = nrqs;
1197       smat_i->nrqr        = nrqr;
1198       smat_i->rbuf1       = rbuf1;
1199       smat_i->rbuf2       = rbuf2;
1200       smat_i->rbuf3       = rbuf3;
1201       smat_i->sbuf2       = sbuf2;
1202       smat_i->req_source2 = req_source2;
1203 
1204       smat_i->sbuf1 = sbuf1;
1205       smat_i->ptr   = ptr;
1206       smat_i->tmp   = tmp;
1207       smat_i->ctr   = ctr;
1208 
1209       smat_i->pa          = pa;
1210       smat_i->req_size    = req_size;
1211       smat_i->req_source1 = req_source1;
1212 
1213       smat_i->allcolumns = PETSC_FALSE;
1214       smat_i->singleis   = PETSC_FALSE;
1215       smat_i->row2proc   = NULL;
1216       smat_i->rmap       = NULL;
1217       smat_i->cmap       = NULL;
1218     }
1219 
1220     if (ismax) PetscCall(PetscFree(lens[0]));
1221     PetscCall(PetscFree(lens));
1222     if (sbuf_aj) {
1223       PetscCall(PetscFree(sbuf_aj[0]));
1224       PetscCall(PetscFree(sbuf_aj));
1225     }
1226 
1227   } /* endof scall == MAT_INITIAL_MATRIX */
1228 
1229   /* Post recv matrix values */
1230   if (!ijonly) {
1231     PetscCall(PetscObjectGetNewTag((PetscObject)C, &tag4));
1232     PetscCall(PetscMalloc1(nrqs, &rbuf4));
1233     PetscCall(PetscMalloc1(nrqs, &r_waits4));
1234     for (PetscMPIInt i = 0; i < nrqs; ++i) {
1235       PetscCall(PetscMalloc1(rbuf2[i][0] * bs2, &rbuf4[i]));
1236       PetscCallMPI(MPIU_Irecv(rbuf4[i], rbuf2[i][0] * bs2, MPIU_SCALAR, req_source2[i], tag4, comm, r_waits4 + i));
1237     }
1238 
1239     /* Allocate sending buffers for a->a, and send them off */
1240     PetscCall(PetscMalloc1(nrqr, &sbuf_aa));
1241     jcnt = 0;
1242     for (PetscMPIInt i = 0; i < nrqr; i++) jcnt += req_size[i];
1243     if (nrqr) PetscCall(PetscMalloc1(jcnt * bs2, &sbuf_aa[0]));
1244     for (PetscMPIInt i = 1; i < nrqr; i++) sbuf_aa[i] = sbuf_aa[i - 1] + req_size[i - 1] * bs2;
1245 
1246     PetscCall(PetscMalloc1(nrqr, &s_waits4));
1247 
1248     for (PetscMPIInt i = 0; i < nrqr; i++) {
1249       rbuf1_i   = rbuf1[i];
1250       sbuf_aa_i = sbuf_aa[i];
1251       ct1       = 2 * rbuf1_i[0] + 1;
1252       ct2       = 0;
1253       for (PetscInt j = 1, max1 = rbuf1_i[0]; j <= max1; j++) {
1254         kmax = rbuf1_i[2 * j];
1255         for (PetscInt k = 0; k < kmax; k++, ct1++) {
1256           PetscInt l;
1257 
1258           row    = rbuf1_i[ct1] - rstart;
1259           nzA    = a_i[row + 1] - a_i[row];
1260           nzB    = b_i[row + 1] - b_i[row];
1261           ncols  = nzA + nzB;
1262           cworkB = PetscSafePointerPlusOffset(b_j, b_i[row]);
1263           vworkA = PetscSafePointerPlusOffset(a_a, a_i[row] * bs2);
1264           vworkB = PetscSafePointerPlusOffset(b_a, b_i[row] * bs2);
1265 
1266           /* load the column values for this row into vals*/
1267           vals = sbuf_aa_i + ct2 * bs2;
1268           for (l = 0; l < nzB; l++) {
1269             if ((bmap[cworkB[l]]) < cstart) PetscCall(PetscArraycpy(vals + l * bs2, vworkB + l * bs2, bs2));
1270             else break;
1271           }
1272           imark = l;
1273           for (l = 0; l < nzA; l++) PetscCall(PetscArraycpy(vals + (imark + l) * bs2, vworkA + l * bs2, bs2));
1274           for (l = imark; l < nzB; l++) PetscCall(PetscArraycpy(vals + (nzA + l) * bs2, vworkB + l * bs2, bs2));
1275 
1276           ct2 += ncols;
1277         }
1278       }
1279       PetscCallMPI(MPIU_Isend(sbuf_aa_i, req_size[i] * bs2, MPIU_SCALAR, req_source1[i], tag4, comm, s_waits4 + i));
1280     }
1281   }
1282 
1283   /* Assemble the matrices */
1284   /* First assemble the local rows */
1285   for (PetscInt i = 0; i < ismax; i++) {
1286     row2proc_i = row2proc[i];
1287     subc       = (Mat_SeqBAIJ *)submats[i]->data;
1288     imat_ilen  = subc->ilen;
1289     imat_j     = subc->j;
1290     imat_i     = subc->i;
1291     imat_a     = subc->a;
1292 
1293     if (!allcolumns[i]) cmap_i = cmap[i];
1294     rmap_i = rmap[i];
1295     irow_i = irow[i];
1296     jmax   = nrow[i];
1297     for (PetscInt j = 0; j < jmax; j++) {
1298       if (allrows[i]) row = j;
1299       else row = irow_i[j];
1300       proc = row2proc_i[j];
1301 
1302       if (proc == rank) {
1303         row    = row - rstart;
1304         nzA    = a_i[row + 1] - a_i[row];
1305         nzB    = b_i[row + 1] - b_i[row];
1306         cworkA = a_j + a_i[row];
1307         cworkB = PetscSafePointerPlusOffset(b_j, b_i[row]);
1308         if (!ijonly) {
1309           vworkA = a_a + a_i[row] * bs2;
1310           vworkB = PetscSafePointerPlusOffset(b_a, b_i[row] * bs2);
1311         }
1312 
1313         if (allrows[i]) {
1314           row = row + rstart;
1315         } else {
1316 #if defined(PETSC_USE_CTABLE)
1317           PetscCall(PetscHMapIGetWithDefault(rmap_i, row + rstart + 1, 0, &row));
1318           row--;
1319 
1320           PetscCheck(row >= 0, PETSC_COMM_SELF, PETSC_ERR_PLIB, "row not found in table");
1321 #else
1322           row = rmap_i[row + rstart];
1323 #endif
1324         }
1325         mat_i = imat_i[row];
1326         if (!ijonly) mat_a = PetscSafePointerPlusOffset(imat_a, mat_i * bs2);
1327         mat_j = PetscSafePointerPlusOffset(imat_j, mat_i);
1328         ilen  = imat_ilen[row];
1329 
1330         /* load the column indices for this row into cols*/
1331         if (!allcolumns[i]) {
1332           PetscInt l;
1333 
1334           for (l = 0; l < nzB; l++) {
1335             if ((ctmp = bmap[cworkB[l]]) < cstart) {
1336 #if defined(PETSC_USE_CTABLE)
1337               PetscCall(PetscHMapIGetWithDefault(cmap_i, ctmp + 1, 0, &tcol));
1338               if (tcol) {
1339 #else
1340               if ((tcol = cmap_i[ctmp])) {
1341 #endif
1342                 *mat_j++ = tcol - 1;
1343                 PetscCall(PetscArraycpy(mat_a, vworkB + l * bs2, bs2));
1344                 mat_a += bs2;
1345                 ilen++;
1346               }
1347             } else break;
1348           }
1349           imark = l;
1350           for (PetscInt l = 0; l < nzA; l++) {
1351 #if defined(PETSC_USE_CTABLE)
1352             PetscCall(PetscHMapIGetWithDefault(cmap_i, cstart + cworkA[l] + 1, 0, &tcol));
1353             if (tcol) {
1354 #else
1355             if ((tcol = cmap_i[cstart + cworkA[l]])) {
1356 #endif
1357               *mat_j++ = tcol - 1;
1358               if (!ijonly) {
1359                 PetscCall(PetscArraycpy(mat_a, vworkA + l * bs2, bs2));
1360                 mat_a += bs2;
1361               }
1362               ilen++;
1363             }
1364           }
1365           for (l = imark; l < nzB; l++) {
1366 #if defined(PETSC_USE_CTABLE)
1367             PetscCall(PetscHMapIGetWithDefault(cmap_i, bmap[cworkB[l]] + 1, 0, &tcol));
1368             if (tcol) {
1369 #else
1370             if ((tcol = cmap_i[bmap[cworkB[l]]])) {
1371 #endif
1372               *mat_j++ = tcol - 1;
1373               if (!ijonly) {
1374                 PetscCall(PetscArraycpy(mat_a, vworkB + l * bs2, bs2));
1375                 mat_a += bs2;
1376               }
1377               ilen++;
1378             }
1379           }
1380         } else { /* allcolumns */
1381           PetscInt l;
1382           for (l = 0; l < nzB; l++) {
1383             if ((ctmp = bmap[cworkB[l]]) < cstart) {
1384               *mat_j++ = ctmp;
1385               PetscCall(PetscArraycpy(mat_a, vworkB + l * bs2, bs2));
1386               mat_a += bs2;
1387               ilen++;
1388             } else break;
1389           }
1390           imark = l;
1391           for (l = 0; l < nzA; l++) {
1392             *mat_j++ = cstart + cworkA[l];
1393             if (!ijonly) {
1394               PetscCall(PetscArraycpy(mat_a, vworkA + l * bs2, bs2));
1395               mat_a += bs2;
1396             }
1397             ilen++;
1398           }
1399           for (l = imark; l < nzB; l++) {
1400             *mat_j++ = bmap[cworkB[l]];
1401             if (!ijonly) {
1402               PetscCall(PetscArraycpy(mat_a, vworkB + l * bs2, bs2));
1403               mat_a += bs2;
1404             }
1405             ilen++;
1406           }
1407         }
1408         imat_ilen[row] = ilen;
1409       }
1410     }
1411   }
1412 
1413   /* Now assemble the off proc rows */
1414   if (!ijonly) PetscCallMPI(MPI_Waitall(nrqs, r_waits4, MPI_STATUSES_IGNORE));
1415   for (tmp2 = 0; tmp2 < nrqs; tmp2++) {
1416     sbuf1_i = sbuf1[pa[tmp2]];
1417     jmax    = sbuf1_i[0];
1418     ct1     = 2 * jmax + 1;
1419     ct2     = 0;
1420     rbuf2_i = rbuf2[tmp2];
1421     rbuf3_i = rbuf3[tmp2];
1422     if (!ijonly) rbuf4_i = rbuf4[tmp2];
1423     for (PetscInt j = 1; j <= jmax; j++) {
1424       is_no  = sbuf1_i[2 * j - 1];
1425       rmap_i = rmap[is_no];
1426       if (!allcolumns[is_no]) cmap_i = cmap[is_no];
1427       subc      = (Mat_SeqBAIJ *)submats[is_no]->data;
1428       imat_ilen = subc->ilen;
1429       imat_j    = subc->j;
1430       imat_i    = subc->i;
1431       if (!ijonly) imat_a = subc->a;
1432       max1 = sbuf1_i[2 * j];
1433       for (PetscInt k = 0; k < max1; k++, ct1++) { /* for each recved block row */
1434         row = sbuf1_i[ct1];
1435 
1436         if (allrows[is_no]) {
1437           row = sbuf1_i[ct1];
1438         } else {
1439 #if defined(PETSC_USE_CTABLE)
1440           PetscCall(PetscHMapIGetWithDefault(rmap_i, row + 1, 0, &row));
1441           row--;
1442           PetscCheck(row >= 0, PETSC_COMM_SELF, PETSC_ERR_PLIB, "row not found in table");
1443 #else
1444           row = rmap_i[row];
1445 #endif
1446         }
1447         ilen  = imat_ilen[row];
1448         mat_i = imat_i[row];
1449         if (!ijonly) mat_a = imat_a + mat_i * bs2;
1450         mat_j = imat_j + mat_i;
1451         max2  = rbuf2_i[ct1];
1452         if (!allcolumns[is_no]) {
1453           for (PetscInt l = 0; l < max2; l++, ct2++) {
1454 #if defined(PETSC_USE_CTABLE)
1455             PetscCall(PetscHMapIGetWithDefault(cmap_i, rbuf3_i[ct2] + 1, 0, &tcol));
1456 #else
1457             tcol = cmap_i[rbuf3_i[ct2]];
1458 #endif
1459             if (tcol) {
1460               *mat_j++ = tcol - 1;
1461               if (!ijonly) {
1462                 PetscCall(PetscArraycpy(mat_a, rbuf4_i + ct2 * bs2, bs2));
1463                 mat_a += bs2;
1464               }
1465               ilen++;
1466             }
1467           }
1468         } else { /* allcolumns */
1469           for (PetscInt l = 0; l < max2; l++, ct2++) {
1470             *mat_j++ = rbuf3_i[ct2]; /* same global column index of C */
1471             if (!ijonly) {
1472               PetscCall(PetscArraycpy(mat_a, rbuf4_i + ct2 * bs2, bs2));
1473               mat_a += bs2;
1474             }
1475             ilen++;
1476           }
1477         }
1478         imat_ilen[row] = ilen;
1479       }
1480     }
1481   }
1482 
1483   if (!iscsorted) { /* sort column indices of the rows */
1484     MatScalar *work;
1485 
1486     PetscCall(PetscMalloc1(bs2, &work));
1487     for (PetscInt i = 0; i < ismax; i++) {
1488       subc      = (Mat_SeqBAIJ *)submats[i]->data;
1489       imat_ilen = subc->ilen;
1490       imat_j    = subc->j;
1491       imat_i    = subc->i;
1492       if (!ijonly) imat_a = subc->a;
1493       if (allcolumns[i]) continue;
1494 
1495       jmax = nrow[i];
1496       for (PetscInt j = 0; j < jmax; j++) {
1497         mat_i = imat_i[j];
1498         mat_j = imat_j + mat_i;
1499         ilen  = imat_ilen[j];
1500         if (ijonly) {
1501           PetscCall(PetscSortInt(ilen, mat_j));
1502         } else {
1503           mat_a = imat_a + mat_i * bs2;
1504           PetscCall(PetscSortIntWithDataArray(ilen, mat_j, mat_a, bs2 * sizeof(MatScalar), work));
1505         }
1506       }
1507     }
1508     PetscCall(PetscFree(work));
1509   }
1510 
1511   if (!ijonly) {
1512     PetscCall(PetscFree(r_waits4));
1513     PetscCallMPI(MPI_Waitall(nrqr, s_waits4, MPI_STATUSES_IGNORE));
1514     PetscCall(PetscFree(s_waits4));
1515   }
1516 
1517   /* Restore the indices */
1518   for (PetscInt i = 0; i < ismax; i++) {
1519     if (!allrows[i]) PetscCall(ISRestoreIndices(isrow[i], irow + i));
1520     if (!allcolumns[i]) PetscCall(ISRestoreIndices(iscol[i], icol + i));
1521   }
1522 
1523   for (PetscInt i = 0; i < ismax; i++) {
1524     PetscCall(MatAssemblyBegin(submats[i], MAT_FINAL_ASSEMBLY));
1525     PetscCall(MatAssemblyEnd(submats[i], MAT_FINAL_ASSEMBLY));
1526   }
1527 
1528   PetscCall(PetscFree5(*(PetscInt ***)&irow, *(PetscInt ***)&icol, nrow, ncol, issorted));
1529   PetscCall(PetscFree5(row2proc, cmap, rmap, allcolumns, allrows));
1530 
1531   if (!ijonly) {
1532     if (sbuf_aa) {
1533       PetscCall(PetscFree(sbuf_aa[0]));
1534       PetscCall(PetscFree(sbuf_aa));
1535     }
1536 
1537     for (PetscMPIInt i = 0; i < nrqs; ++i) PetscCall(PetscFree(rbuf4[i]));
1538     PetscCall(PetscFree(rbuf4));
1539   }
1540   c->ijonly = PETSC_FALSE; /* set back to the default */
1541   PetscFunctionReturn(PETSC_SUCCESS);
1542 }
1543