xref: /petsc/src/mat/impls/adj/mpi/mpiadj.c (revision c3d102fe2dfc19c1832704c537829f04ceb136a2)
1 /*
2     Defines the basic matrix operations for the ADJ adjacency list matrix data-structure.
3 */
4 #include "src/mat/impls/adj/mpi/mpiadj.h"
5 #include "petscsys.h"
6 
7 #undef __FUNCT__
8 #define __FUNCT__ "MatView_MPIAdj_ASCII"
9 PetscErrorCode MatView_MPIAdj_ASCII(Mat A,PetscViewer viewer)
10 {
11   Mat_MPIAdj        *a = (Mat_MPIAdj*)A->data;
12   PetscErrorCode    ierr;
13   PetscInt          i,j,m = A->m;
14   char              *name;
15   PetscViewerFormat format;
16 
17   PetscFunctionBegin;
18   ierr = PetscObjectGetName((PetscObject)A,&name);CHKERRQ(ierr);
19   ierr = PetscViewerGetFormat(viewer,&format);CHKERRQ(ierr);
20   if (format == PETSC_VIEWER_ASCII_INFO) {
21     PetscFunctionReturn(0);
22   } else if (format == PETSC_VIEWER_ASCII_MATLAB) {
23     SETERRQ(PETSC_ERR_SUP,"Matlab format not supported");
24   } else {
25     ierr = PetscViewerASCIIUseTabs(viewer,PETSC_NO);CHKERRQ(ierr);
26     for (i=0; i<m; i++) {
27       ierr = PetscViewerASCIISynchronizedPrintf(viewer,"row %D:",i+a->rstart);CHKERRQ(ierr);
28       for (j=a->i[i]; j<a->i[i+1]; j++) {
29         ierr = PetscViewerASCIISynchronizedPrintf(viewer," %D ",a->j[j]);CHKERRQ(ierr);
30       }
31       ierr = PetscViewerASCIISynchronizedPrintf(viewer,"\n");CHKERRQ(ierr);
32     }
33     ierr = PetscViewerASCIIUseTabs(viewer,PETSC_YES);CHKERRQ(ierr);
34   }
35   ierr = PetscViewerFlush(viewer);CHKERRQ(ierr);
36   PetscFunctionReturn(0);
37 }
38 
39 #undef __FUNCT__
40 #define __FUNCT__ "MatView_MPIAdj"
41 PetscErrorCode MatView_MPIAdj(Mat A,PetscViewer viewer)
42 {
43   PetscErrorCode ierr;
44   PetscTruth     iascii;
45 
46   PetscFunctionBegin;
47   ierr = PetscTypeCompare((PetscObject)viewer,PETSC_VIEWER_ASCII,&iascii);CHKERRQ(ierr);
48   if (iascii) {
49     ierr = MatView_MPIAdj_ASCII(A,viewer);CHKERRQ(ierr);
50   } else {
51     SETERRQ1(PETSC_ERR_SUP,"Viewer type %s not supported by MPIAdj",((PetscObject)viewer)->type_name);
52   }
53   PetscFunctionReturn(0);
54 }
55 
56 #undef __FUNCT__
57 #define __FUNCT__ "MatDestroy_MPIAdj"
58 PetscErrorCode MatDestroy_MPIAdj(Mat mat)
59 {
60   Mat_MPIAdj     *a = (Mat_MPIAdj*)mat->data;
61   PetscErrorCode ierr;
62 
63   PetscFunctionBegin;
64 #if defined(PETSC_USE_LOG)
65   PetscLogObjectState((PetscObject)mat,"Rows=%D, Cols=%D, NZ=%D",mat->m,mat->n,a->nz);
66 #endif
67   if (a->diag) {ierr = PetscFree(a->diag);CHKERRQ(ierr);}
68   if (a->freeaij) {
69     ierr = PetscFree(a->i);CHKERRQ(ierr);
70     ierr = PetscFree(a->j);CHKERRQ(ierr);
71     if (a->values) {ierr = PetscFree(a->values);CHKERRQ(ierr);}
72   }
73   ierr = PetscFree(a->rowners);CHKERRQ(ierr);
74   ierr = PetscFree(a);CHKERRQ(ierr);
75 
76   ierr = PetscObjectComposeFunction((PetscObject)mat,"MatMPIAdjSetPreallocation_C","",PETSC_NULL);CHKERRQ(ierr);
77   PetscFunctionReturn(0);
78 }
79 
80 #undef __FUNCT__
81 #define __FUNCT__ "MatSetOption_MPIAdj"
82 PetscErrorCode MatSetOption_MPIAdj(Mat A,MatOption op)
83 {
84   Mat_MPIAdj *a = (Mat_MPIAdj*)A->data;
85 
86   PetscFunctionBegin;
87   switch (op) {
88   case MAT_SYMMETRIC:
89   case MAT_STRUCTURALLY_SYMMETRIC:
90   case MAT_HERMITIAN:
91     a->symmetric = PETSC_TRUE;
92     break;
93   case MAT_NOT_SYMMETRIC:
94   case MAT_NOT_STRUCTURALLY_SYMMETRIC:
95   case MAT_NOT_HERMITIAN:
96     a->symmetric = PETSC_FALSE;
97     break;
98   case MAT_SYMMETRY_ETERNAL:
99   case MAT_NOT_SYMMETRY_ETERNAL:
100     break;
101   default:
102     PetscLogInfo(A,"MatSetOption_MPIAdj:Option ignored\n");
103     break;
104   }
105   PetscFunctionReturn(0);
106 }
107 
108 
109 /*
110      Adds diagonal pointers to sparse matrix structure.
111 */
112 
113 #undef __FUNCT__
114 #define __FUNCT__ "MatMarkDiagonal_MPIAdj"
115 PetscErrorCode MatMarkDiagonal_MPIAdj(Mat A)
116 {
117   Mat_MPIAdj     *a = (Mat_MPIAdj*)A->data;
118   PetscErrorCode ierr;
119   PetscInt       i,j,*diag,m = A->m;
120 
121   PetscFunctionBegin;
122   ierr = PetscMalloc((m+1)*sizeof(PetscInt),&diag);CHKERRQ(ierr);
123   ierr = PetscLogObjectMemory(A,(m+1)*sizeof(PetscInt));CHKERRQ(ierr);
124   for (i=0; i<A->m; i++) {
125     for (j=a->i[i]; j<a->i[i+1]; j++) {
126       if (a->j[j] == i) {
127         diag[i] = j;
128         break;
129       }
130     }
131   }
132   a->diag = diag;
133   PetscFunctionReturn(0);
134 }
135 
136 #undef __FUNCT__
137 #define __FUNCT__ "MatGetRow_MPIAdj"
138 PetscErrorCode MatGetRow_MPIAdj(Mat A,PetscInt row,PetscInt *nz,PetscInt **idx,PetscScalar **v)
139 {
140   Mat_MPIAdj *a = (Mat_MPIAdj*)A->data;
141   PetscInt   *itmp;
142 
143   PetscFunctionBegin;
144   row -= a->rstart;
145 
146   if (row < 0 || row >= A->m) SETERRQ(PETSC_ERR_ARG_OUTOFRANGE,"Row out of range");
147 
148   *nz = a->i[row+1] - a->i[row];
149   if (v) *v = PETSC_NULL;
150   if (idx) {
151     itmp = a->j + a->i[row];
152     if (*nz) {
153       *idx = itmp;
154     }
155     else *idx = 0;
156   }
157   PetscFunctionReturn(0);
158 }
159 
160 #undef __FUNCT__
161 #define __FUNCT__ "MatRestoreRow_MPIAdj"
162 PetscErrorCode MatRestoreRow_MPIAdj(Mat A,PetscInt row,PetscInt *nz,PetscInt **idx,PetscScalar **v)
163 {
164   PetscFunctionBegin;
165   PetscFunctionReturn(0);
166 }
167 
168 #undef __FUNCT__
169 #define __FUNCT__ "MatEqual_MPIAdj"
170 PetscErrorCode MatEqual_MPIAdj(Mat A,Mat B,PetscTruth* flg)
171 {
172   Mat_MPIAdj     *a = (Mat_MPIAdj *)A->data,*b = (Mat_MPIAdj *)B->data;
173   PetscErrorCode ierr;
174   PetscTruth     flag;
175 
176   PetscFunctionBegin;
177   /* If the  matrix dimensions are not equal,or no of nonzeros */
178   if ((A->m != B->m) ||(a->nz != b->nz)) {
179     flag = PETSC_FALSE;
180   }
181 
182   /* if the a->i are the same */
183   ierr = PetscMemcmp(a->i,b->i,(A->m+1)*sizeof(PetscInt),&flag);CHKERRQ(ierr);
184 
185   /* if a->j are the same */
186   ierr = PetscMemcmp(a->j,b->j,(a->nz)*sizeof(PetscInt),&flag);CHKERRQ(ierr);
187 
188   ierr = MPI_Allreduce(&flag,flg,1,MPI_INT,MPI_LAND,A->comm);CHKERRQ(ierr);
189   PetscFunctionReturn(0);
190 }
191 
192 #undef __FUNCT__
193 #define __FUNCT__ "MatGetRowIJ_MPIAdj"
194 PetscErrorCode MatGetRowIJ_MPIAdj(Mat A,PetscInt oshift,PetscTruth symmetric,PetscInt *m,PetscInt *ia[],PetscInt *ja[],PetscTruth *done)
195 {
196   PetscErrorCode ierr;
197   PetscMPIInt    size;
198   PetscInt       i;
199   Mat_MPIAdj     *a = (Mat_MPIAdj *)A->data;
200 
201   PetscFunctionBegin;
202   ierr = MPI_Comm_size(A->comm,&size);CHKERRQ(ierr);
203   if (size > 1) {*done = PETSC_FALSE; PetscFunctionReturn(0);}
204   *m    = A->m;
205   *ia   = a->i;
206   *ja   = a->j;
207   *done = PETSC_TRUE;
208   if (oshift) {
209     for (i=0; i<(*ia)[*m]; i++) {
210       (*ja)[i]++;
211     }
212     for (i=0; i<=(*m); i++) (*ia)[i]++;
213   }
214   PetscFunctionReturn(0);
215 }
216 
217 #undef __FUNCT__
218 #define __FUNCT__ "MatRestoreRowIJ_MPIAdj"
219 PetscErrorCode MatRestoreRowIJ_MPIAdj(Mat A,PetscInt oshift,PetscTruth symmetric,PetscInt *m,PetscInt *ia[],PetscInt *ja[],PetscTruth *done)
220 {
221   PetscInt   i;
222   Mat_MPIAdj *a = (Mat_MPIAdj *)A->data;
223 
224   PetscFunctionBegin;
225   if (ia && a->i != *ia) SETERRQ(PETSC_ERR_ARG_WRONGSTATE,"ia passed back is not one obtained with MatGetRowIJ()");
226   if (ja && a->j != *ja) SETERRQ(PETSC_ERR_ARG_WRONGSTATE,"ja passed back is not one obtained with MatGetRowIJ()");
227   if (oshift) {
228     for (i=0; i<=(*m); i++) (*ia)[i]--;
229     for (i=0; i<(*ia)[*m]; i++) {
230       (*ja)[i]--;
231     }
232   }
233   PetscFunctionReturn(0);
234 }
235 
236 /* -------------------------------------------------------------------*/
237 static struct _MatOps MatOps_Values = {0,
238        MatGetRow_MPIAdj,
239        MatRestoreRow_MPIAdj,
240        0,
241 /* 4*/ 0,
242        0,
243        0,
244        0,
245        0,
246        0,
247 /*10*/ 0,
248        0,
249        0,
250        0,
251        0,
252 /*15*/ 0,
253        MatEqual_MPIAdj,
254        0,
255        0,
256        0,
257 /*20*/ 0,
258        0,
259        0,
260        MatSetOption_MPIAdj,
261        0,
262 /*25*/ 0,
263        0,
264        0,
265        0,
266        0,
267 /*30*/ 0,
268        0,
269        0,
270        0,
271        0,
272 /*35*/ 0,
273        0,
274        0,
275        0,
276        0,
277 /*40*/ 0,
278        0,
279        0,
280        0,
281        0,
282 /*45*/ 0,
283        0,
284        0,
285        0,
286        0,
287 /*50*/ 0,
288        MatGetRowIJ_MPIAdj,
289        MatRestoreRowIJ_MPIAdj,
290        0,
291        0,
292 /*55*/ 0,
293        0,
294        0,
295        0,
296        0,
297 /*60*/ 0,
298        MatDestroy_MPIAdj,
299        MatView_MPIAdj,
300        MatGetPetscMaps_Petsc,
301        0,
302 /*65*/ 0,
303        0,
304        0,
305        0,
306        0,
307 /*70*/ 0,
308        0,
309        0,
310        0,
311        0,
312 /*75*/ 0,
313        0,
314        0,
315        0,
316        0,
317 /*80*/ 0,
318        0,
319        0,
320        0,
321        0,
322 /*85*/ 0,
323        0,
324        0,
325        0,
326        0,
327 /*90*/ 0,
328        0,
329        0,
330        0,
331        0,
332 /*95*/ 0,
333        0,
334        0,
335        0};
336 
337 EXTERN_C_BEGIN
338 #undef __FUNCT__
339 #define __FUNCT__ "MatMPIAdjSetPreallocation_MPIAdj"
340 PetscErrorCode MatMPIAdjSetPreallocation_MPIAdj(Mat B,PetscInt *i,PetscInt *j,PetscInt *values)
341 {
342   Mat_MPIAdj     *b = (Mat_MPIAdj *)B->data;
343   PetscErrorCode ierr;
344 #if defined(PETSC_USE_DEBUG)
345   PetscInt       ii;
346 #endif
347 
348   PetscFunctionBegin;
349   B->preallocated = PETSC_TRUE;
350 #if defined(PETSC_USE_DEBUG)
351   if (i[0] != 0) SETERRQ1(PETSC_ERR_ARG_OUTOFRANGE,"First i[] index must be zero, instead it is %D\n",i[0]);
352   for (ii=1; ii<B->m; ii++) {
353     if (i[ii] < 0 || i[ii] < i[ii-1]) {
354       SETERRQ4(PETSC_ERR_ARG_OUTOFRANGE,"i[%D]=%D index is out of range: i[%D]=%D",ii,i[ii],ii-1,i[ii-1]);
355     }
356   }
357   for (ii=0; ii<i[B->m]; ii++) {
358     if (j[ii] < 0 || j[ii] >= B->N) {
359       SETERRQ2(PETSC_ERR_ARG_OUTOFRANGE,"Column index %D out of range %D\n",ii,j[ii]);
360     }
361   }
362 #endif
363 
364   b->j      = j;
365   b->i      = i;
366   b->values = values;
367 
368   b->nz               = i[B->m];
369   b->diag             = 0;
370   b->symmetric        = PETSC_FALSE;
371   b->freeaij          = PETSC_TRUE;
372 
373   ierr = MatAssemblyBegin(B,MAT_FINAL_ASSEMBLY);CHKERRQ(ierr);
374   ierr = MatAssemblyEnd(B,MAT_FINAL_ASSEMBLY);CHKERRQ(ierr);
375   PetscFunctionReturn(0);
376 }
377 EXTERN_C_END
378 
379 /*MC
380    MATMPIADJ - MATMPIADJ = "mpiadj" - A matrix type to be used for distributed adjacency matrices,
381    intended for use constructing orderings and partitionings.
382 
383   Level: beginner
384 
385 .seealso: MatCreateMPIAdj
386 M*/
387 
388 EXTERN_C_BEGIN
389 #undef __FUNCT__
390 #define __FUNCT__ "MatCreate_MPIAdj"
391 PetscErrorCode MatCreate_MPIAdj(Mat B)
392 {
393   Mat_MPIAdj     *b;
394   PetscErrorCode ierr;
395   PetscInt       ii;
396   PetscMPIInt    size,rank;
397 
398   PetscFunctionBegin;
399   ierr = MPI_Comm_size(B->comm,&size);CHKERRQ(ierr);
400   ierr = MPI_Comm_rank(B->comm,&rank);CHKERRQ(ierr);
401 
402   ierr                = PetscNew(Mat_MPIAdj,&b);CHKERRQ(ierr);
403   B->data             = (void*)b;
404   ierr                = PetscMemcpy(B->ops,&MatOps_Values,sizeof(struct _MatOps));CHKERRQ(ierr);
405   B->factor           = 0;
406   B->lupivotthreshold = 1.0;
407   B->mapping          = 0;
408   B->assembled        = PETSC_FALSE;
409 
410   ierr = PetscSplitOwnership(B->comm,&B->m,&B->M);CHKERRQ(ierr);
411   B->n = B->N = PetscMax(B->N,B->n);CHKERRQ(ierr);
412 
413   /* the information in the maps duplicates the information computed below, eventually
414      we should remove the duplicate information that is not contained in the maps */
415   ierr = PetscMapCreateMPI(B->comm,B->m,B->M,&B->rmap);CHKERRQ(ierr);
416   /* we don't know the "local columns" so just use the row information :-(*/
417   ierr = PetscMapCreateMPI(B->comm,B->m,B->M,&B->cmap);CHKERRQ(ierr);
418 
419   ierr = PetscMalloc((size+1)*sizeof(PetscInt),&b->rowners);CHKERRQ(ierr);
420   ierr = PetscLogObjectMemory(B,(size+2)*sizeof(PetscInt)+sizeof(struct _p_Mat)+sizeof(Mat_MPIAdj));CHKERRQ(ierr);
421   ierr = MPI_Allgather(&B->m,1,MPI_INT,b->rowners+1,1,MPI_INT,B->comm);CHKERRQ(ierr);
422   b->rowners[0] = 0;
423   for (ii=2; ii<=size; ii++) {
424     b->rowners[ii] += b->rowners[ii-1];
425   }
426   b->rstart = b->rowners[rank];
427   b->rend   = b->rowners[rank+1];
428   ierr = PetscObjectComposeFunctionDynamic((PetscObject)B,"MatMPIAdjSetPreallocation_C",
429                                     "MatMPIAdjSetPreallocation_MPIAdj",
430                                      MatMPIAdjSetPreallocation_MPIAdj);CHKERRQ(ierr);
431   PetscFunctionReturn(0);
432 }
433 EXTERN_C_END
434 
435 #undef __FUNCT__
436 #define __FUNCT__ "MatMPIAdjSetPreallocation"
437 /*@C
438    MatMPIAdjSetPreallocation - Sets the array used for storing the matrix elements
439 
440    Collective on MPI_Comm
441 
442    Input Parameters:
443 +  A - the matrix
444 .  i - the indices into j for the start of each row
445 .  j - the column indices for each row (sorted for each row).
446        The indices in i and j start with zero (NOT with one).
447 -  values - [optional] edge weights
448 
449    Level: intermediate
450 
451 .seealso: MatCreate(), MatCreateMPIAdj(), MatSetValues()
452 @*/
453 PetscErrorCode MatMPIAdjSetPreallocation(Mat B,PetscInt *i,PetscInt *j,PetscInt *values)
454 {
455   PetscErrorCode ierr,(*f)(Mat,PetscInt*,PetscInt*,PetscInt*);
456 
457   PetscFunctionBegin;
458   ierr = PetscObjectQueryFunction((PetscObject)B,"MatMPIAdjSetPreallocation_C",(void (**)(void))&f);CHKERRQ(ierr);
459   if (f) {
460     ierr = (*f)(B,i,j,values);CHKERRQ(ierr);
461   }
462   PetscFunctionReturn(0);
463 }
464 
465 #undef __FUNCT__
466 #define __FUNCT__ "MatCreateMPIAdj"
467 /*@C
468    MatCreateMPIAdj - Creates a sparse matrix representing an adjacency list.
469    The matrix does not have numerical values associated with it, but is
470    intended for ordering (to reduce bandwidth etc) and partitioning.
471 
472    Collective on MPI_Comm
473 
474    Input Parameters:
475 +  comm - MPI communicator
476 .  m - number of local rows
477 .  n - number of columns
478 .  i - the indices into j for the start of each row
479 .  j - the column indices for each row (sorted for each row).
480        The indices in i and j start with zero (NOT with one).
481 -  values -[optional] edge weights
482 
483    Output Parameter:
484 .  A - the matrix
485 
486    Level: intermediate
487 
488    Notes: This matrix object does not support most matrix operations, include
489    MatSetValues().
490    You must NOT free the ii, values and jj arrays yourself. PETSc will free them
491    when the matrix is destroyed; you must allocate them with PetscMalloc(). If you
492     call from Fortran you need not create the arrays with PetscMalloc().
493    Should not include the matrix diagonals.
494 
495    If you already have a matrix, you can create its adjacency matrix by a call
496    to MatConvert, specifying a type of MATMPIADJ.
497 
498    Possible values for MatSetOption() - MAT_STRUCTURALLY_SYMMETRIC
499 
500 .seealso: MatCreate(), MatConvert(), MatGetOrdering()
501 @*/
502 PetscErrorCode MatCreateMPIAdj(MPI_Comm comm,PetscInt m,PetscInt n,PetscInt *i,PetscInt *j,PetscInt *values,Mat *A)
503 {
504   PetscErrorCode ierr;
505 
506   PetscFunctionBegin;
507   ierr = MatCreate(comm,m,n,PETSC_DETERMINE,n,A);CHKERRQ(ierr);
508   ierr = MatSetType(*A,MATMPIADJ);CHKERRQ(ierr);
509   ierr = MatMPIAdjSetPreallocation(*A,i,j,values);CHKERRQ(ierr);
510   PetscFunctionReturn(0);
511 }
512 
513 EXTERN_C_BEGIN
514 #undef __FUNCT__
515 #define __FUNCT__ "MatConvertTo_MPIAdj"
516 PetscErrorCode MatConvertTo_MPIAdj(Mat A,MatType type,MatReuse reuse,Mat *newmat)
517 {
518   Mat               B;
519   PetscErrorCode    ierr;
520   PetscInt          i,m,N,nzeros = 0,*ia,*ja,len,rstart,cnt,j,*a;
521   const PetscInt    *rj;
522   const PetscScalar *ra;
523   MPI_Comm          comm;
524 
525   PetscFunctionBegin;
526   ierr = MatGetSize(A,PETSC_NULL,&N);CHKERRQ(ierr);
527   ierr = MatGetLocalSize(A,&m,PETSC_NULL);CHKERRQ(ierr);
528   ierr = MatGetOwnershipRange(A,&rstart,PETSC_NULL);CHKERRQ(ierr);
529 
530   /* count the number of nonzeros per row */
531   for (i=0; i<m; i++) {
532     ierr   = MatGetRow(A,i+rstart,&len,&rj,PETSC_NULL);CHKERRQ(ierr);
533     for (j=0; j<len; j++) {
534       if (rj[j] == i+rstart) {len--; break;}    /* don't count diagonal */
535     }
536     ierr   = MatRestoreRow(A,i+rstart,&len,&rj,PETSC_NULL);CHKERRQ(ierr);
537     nzeros += len;
538   }
539 
540   /* malloc space for nonzeros */
541   ierr = PetscMalloc((nzeros+1)*sizeof(PetscInt),&a);CHKERRQ(ierr);
542   ierr = PetscMalloc((N+1)*sizeof(PetscInt),&ia);CHKERRQ(ierr);
543   ierr = PetscMalloc((nzeros+1)*sizeof(PetscInt),&ja);CHKERRQ(ierr);
544 
545   nzeros = 0;
546   ia[0]  = 0;
547   for (i=0; i<m; i++) {
548     ierr    = MatGetRow(A,i+rstart,&len,&rj,&ra);CHKERRQ(ierr);
549     cnt     = 0;
550     for (j=0; j<len; j++) {
551       if (rj[j] != i+rstart) { /* if not diagonal */
552         a[nzeros+cnt]    = (PetscInt) PetscAbsScalar(ra[j]);
553         ja[nzeros+cnt++] = rj[j];
554       }
555     }
556     ierr    = MatRestoreRow(A,i+rstart,&len,&rj,&ra);CHKERRQ(ierr);
557     nzeros += cnt;
558     ia[i+1] = nzeros;
559   }
560 
561   ierr = PetscObjectGetComm((PetscObject)A,&comm);CHKERRQ(ierr);
562   ierr = MatCreate(comm,m,N,PETSC_DETERMINE,N,&B);CHKERRQ(ierr);
563   ierr = MatSetType(B,type);CHKERRQ(ierr);
564   ierr = MatMPIAdjSetPreallocation(B,ia,ja,a);CHKERRQ(ierr);
565 
566   if (reuse == MAT_REUSE_MATRIX) {
567     ierr = MatHeaderCopy(A,B);CHKERRQ(ierr);
568   } else {
569     *newmat = B;
570   }
571   PetscFunctionReturn(0);
572 }
573 EXTERN_C_END
574 
575 
576 
577 
578 
579