xref: /petsc/src/mat/utils/overlapsplit.c (revision bdeb5f4eacf5dccc3fa35d5507b443449c4fb7e7)
12452736bSFande Kong #include <petscsf.h>
22452736bSFande Kong #include <petsc/private/matimpl.h>
32452736bSFande Kong 
42452736bSFande Kong 
5ca2fc57aSFande Kong #undef __FUNCT__
6ca2fc57aSFande Kong #define __FUNCT__ "MatIncreaseOverlapSplit_Single"
7ca2fc57aSFande Kong 
82452736bSFande Kong /*
92452736bSFande Kong  * Increase overlap for the sub-matrix across sub communicator
102452736bSFande Kong  * sub-matrix could be a graph or numerical matrix
112452736bSFande Kong  * */
122452736bSFande Kong PetscErrorCode  MatIncreaseOverlapSplit_Single(Mat mat,IS *is,PetscInt ov)
132452736bSFande Kong {
142452736bSFande Kong   PetscInt         i,nindx,*indices_sc,*indices_ov,localsize,*localsizes_sc,localsize_tmp;
152452736bSFande Kong   PetscInt         *indices_ov_rd,nroots,nleaves,*localoffsets,*indices_recv,*sources_sc,*sources_sc_rd;
162452736bSFande Kong   const PetscInt   *indices;
17a69400e0SFande Kong   PetscMPIInt      srank,ssize,issamecomm,k,grank;
18a69400e0SFande Kong   IS               is_sc,allis_sc,allis_sc_tmp,partitioning;
19a69400e0SFande Kong   MPI_Comm         gcomm,dcomm,scomm;
202452736bSFande Kong   PetscSF          sf;
212452736bSFande Kong   PetscSFNode      *remote;
222452736bSFande Kong   Mat              *smat;
232452736bSFande Kong   MatPartitioning  part;
242452736bSFande Kong   PetscErrorCode   ierr;
252452736bSFande Kong 
262452736bSFande Kong   PetscFunctionBegin;
272452736bSFande Kong   /* get a sub communicator before call individual MatIncreaseOverlap
282452736bSFande Kong    * since the sub communicator may be changed.
292452736bSFande Kong    * */
30a69400e0SFande Kong   ierr = PetscObjectGetComm((PetscObject)(*is),&dcomm);CHKERRQ(ierr);
31a69400e0SFande Kong   /*make a copy before the original one is deleted*/
32a69400e0SFande Kong   ierr = PetscCommDuplicate(dcomm,&scomm,NULL);CHKERRQ(ierr);
33ca2fc57aSFande Kong   /*get a global communicator, where mat should be a global matrix  */
34ca2fc57aSFande Kong   ierr = PetscObjectGetComm((PetscObject)mat,&gcomm);CHKERRQ(ierr);
35ca2fc57aSFande Kong #if 1
36ca2fc57aSFande Kong   ierr = PetscPrintf(gcomm,"before mat->ops->increaseoverlap\n");CHKERRQ(ierr);
37ca2fc57aSFande Kong #endif
382452736bSFande Kong   /*increase overlap on each individual subdomain*/
392452736bSFande Kong   ierr = (*mat->ops->increaseoverlap)(mat,1,is,ov);CHKERRQ(ierr);
40ca2fc57aSFande Kong #if 1
41ca2fc57aSFande Kong   ierr = PetscPrintf(gcomm,"after mat->ops->increaseoverlap \n");CHKERRQ(ierr);
42a69400e0SFande Kong   ierr = ISView(*is,PETSC_VIEWER_STDOUT_SELF);CHKERRQ(ierr);
43ca2fc57aSFande Kong #endif
442452736bSFande Kong   /*compare communicators */
452452736bSFande Kong   ierr = MPI_Comm_compare(gcomm,scomm,&issamecomm);CHKERRQ(ierr);
462452736bSFande Kong   /* if the sub-communicator is the same as the global communicator,
472452736bSFande Kong    * user does not want to use a sub-communicator
482452736bSFande Kong    * */
49a69400e0SFande Kong   if(issamecomm == MPI_IDENT || issamecomm == MPI_CONGRUENT) PetscFunctionReturn(0);
502452736bSFande Kong   /* if the sub-communicator is petsc_comm_self,
512452736bSFande Kong    * user also does not care the sub-communicator
522452736bSFande Kong    * */
532452736bSFande Kong   ierr = MPI_Comm_compare(scomm,PETSC_COMM_SELF,&issamecomm);CHKERRQ(ierr);
54a69400e0SFande Kong   if(issamecomm == MPI_IDENT || issamecomm == MPI_CONGRUENT){PetscFunctionReturn(0);}
55ca2fc57aSFande Kong   /*local rank, size in a sub-communicator  */
562452736bSFande Kong   ierr = MPI_Comm_rank(scomm,&srank);CHKERRQ(ierr);
572452736bSFande Kong   ierr = MPI_Comm_size(scomm,&ssize);CHKERRQ(ierr);
58a69400e0SFande Kong   ierr = MPI_Comm_rank(gcomm,&grank);CHKERRQ(ierr);
59ca2fc57aSFande Kong   /*create a new IS based on sub-communicator
60ca2fc57aSFande Kong    * since the old IS is often based on petsc_comm_self
612452736bSFande Kong    * */
622452736bSFande Kong   ierr = ISGetLocalSize(*is,&nindx);CHKERRQ(ierr);
632452736bSFande Kong   ierr = PetscCalloc1(nindx,&indices_sc);CHKERRQ(ierr);
642452736bSFande Kong   ierr = ISGetIndices(*is,&indices);CHKERRQ(ierr);
652452736bSFande Kong   ierr = PetscMemcpy(indices_sc,indices,sizeof(PetscInt)*nindx);CHKERRQ(ierr);
662452736bSFande Kong   ierr = ISRestoreIndices(*is,&indices);CHKERRQ(ierr);
672452736bSFande Kong   /*we do not need any more*/
682452736bSFande Kong   ierr = ISDestroy(is);CHKERRQ(ierr);
692452736bSFande Kong   /*create a index set based on the sub communicator  */
702452736bSFande Kong   ierr = ISCreateGeneral(scomm,nindx,indices_sc,PETSC_OWN_POINTER,&is_sc);CHKERRQ(ierr);
712452736bSFande Kong   /*gather all indices within  the sub communicator*/
722452736bSFande Kong   ierr = ISAllGather(is_sc,&allis_sc);CHKERRQ(ierr);
73a69400e0SFande Kong #if 1
74a69400e0SFande Kong   ierr = ISView(allis_sc,PETSC_VIEWER_STDOUT_SELF);CHKERRQ(ierr);
75a69400e0SFande Kong   ierr = MPI_Barrier(gcomm);CHKERRQ(ierr);
76a69400e0SFande Kong #endif
772452736bSFande Kong   ierr = ISDestroy(&is_sc);CHKERRQ(ierr);
782452736bSFande Kong   /* gather local sizes */
792452736bSFande Kong   ierr = PetscMalloc1(ssize,&localsizes_sc);CHKERRQ(ierr);
80ca2fc57aSFande Kong   /*get individual local sizes for all index sets*/
812452736bSFande Kong   ierr = MPI_Gather(&nindx,1,MPIU_INT,localsizes_sc,1,MPIU_INT,0,scomm);CHKERRQ(ierr);
82a69400e0SFande Kong #if 1
83a69400e0SFande Kong   if(!srank){
84a69400e0SFande Kong 	for(i=0; i<ssize; i++){
85a69400e0SFande Kong 	  ierr = PetscPrintf(PETSC_COMM_SELF," localsize[%d]: %d \n",i,localsizes_sc[i]);CHKERRQ(ierr);
86a69400e0SFande Kong 	}
87a69400e0SFande Kong   }
88a69400e0SFande Kong   ierr = MPI_Barrier(gcomm);CHKERRQ(ierr);
89a69400e0SFande Kong #endif
90ca2fc57aSFande Kong   /*only root does these computations */
912452736bSFande Kong   if(!srank){
922452736bSFande Kong    /*get local size for the big index set*/
932452736bSFande Kong    ierr = ISGetLocalSize(allis_sc,&localsize);CHKERRQ(ierr);
942452736bSFande Kong    ierr = PetscCalloc2(localsize,&indices_ov,localsize,&sources_sc);CHKERRQ(ierr);
952452736bSFande Kong    ierr = PetscCalloc2(localsize,&indices_ov_rd,localsize,&sources_sc_rd);CHKERRQ(ierr);
962452736bSFande Kong    ierr = ISGetIndices(allis_sc,&indices);CHKERRQ(ierr);
972452736bSFande Kong    ierr = PetscMemcpy(indices_ov,indices,sizeof(PetscInt)*localsize);CHKERRQ(ierr);
982452736bSFande Kong    ierr = ISRestoreIndices(allis_sc,&indices);CHKERRQ(ierr);
99ca2fc57aSFande Kong    /*we do not need it any more */
1002452736bSFande Kong    ierr = ISDestroy(&allis_sc);CHKERRQ(ierr);
1012452736bSFande Kong    /*assign corresponding sources */
1022452736bSFande Kong    localsize_tmp = 0;
1032452736bSFande Kong    for(k=0; k<ssize; k++){
1042452736bSFande Kong      for(i=0; i<localsizes_sc[k]; i++){
1052452736bSFande Kong        sources_sc[localsize_tmp++] = k;
1062452736bSFande Kong      }
1072452736bSFande Kong    }
1082452736bSFande Kong    /*record where indices come from */
1092452736bSFande Kong    ierr = PetscSortIntWithArray(localsize,indices_ov,sources_sc);CHKERRQ(ierr);
110a69400e0SFande Kong #if 0
111a69400e0SFande Kong    ierr = PetscIntView(localsize,indices_ov,PETSC_VIEWER_STDOUT_SELF);CHKERRQ(ierr);
112a69400e0SFande Kong    ierr = PetscIntView(localsize,sources_sc,PETSC_VIEWER_STDOUT_SELF);CHKERRQ(ierr);
113a69400e0SFande Kong #endif
114ca2fc57aSFande Kong    /*count local sizes for reduced indices */
1152452736bSFande Kong    ierr = PetscMemzero(localsizes_sc,sizeof(PetscInt)*ssize);CHKERRQ(ierr);
116ca2fc57aSFande Kong    /*initialize the first entity*/
1172452736bSFande Kong    if(localsize){
1182452736bSFande Kong 	 indices_ov_rd[0] = indices_ov[0];
1192452736bSFande Kong 	 sources_sc_rd[0] = sources_sc[0];
1202452736bSFande Kong 	 localsizes_sc[sources_sc[0]]++;
1212452736bSFande Kong    }
122ca2fc57aSFande Kong    localsize_tmp = 1;
1232452736bSFande Kong    /*remove duplicate integers */
1242452736bSFande Kong    for(i=1; i<localsize; i++){
1252452736bSFande Kong 	 if(indices_ov[i] != indices_ov[i-1]){
1262452736bSFande Kong 	   indices_ov_rd[localsize_tmp]   = indices_ov[i];
1272452736bSFande Kong 	   sources_sc_rd[localsize_tmp++] = sources_sc[i];
1282452736bSFande Kong 	   localsizes_sc[sources_sc[i]]++;
1292452736bSFande Kong 	 }
1302452736bSFande Kong    }
1312452736bSFande Kong    ierr = PetscFree2(indices_ov,sources_sc);CHKERRQ(ierr);
1322452736bSFande Kong    ierr = PetscCalloc1(ssize+1,&localoffsets);CHKERRQ(ierr);
1332452736bSFande Kong    for(k=0; k<ssize; k++){
134a69400e0SFande Kong 	 localoffsets[k+1] = localoffsets[k] + localsizes_sc[k];
1352452736bSFande Kong    }
136ca2fc57aSFande Kong    /*construct a star forest to send data back */
1372452736bSFande Kong    nleaves = localoffsets[ssize];
1382452736bSFande Kong    ierr = PetscMemzero(localoffsets,(ssize+1)*sizeof(PetscInt));CHKERRQ(ierr);
1392452736bSFande Kong    nroots  = localsizes_sc[srank];
1402452736bSFande Kong    ierr = PetscCalloc1(nleaves,&remote);CHKERRQ(ierr);
1412452736bSFande Kong    for(i=0; i<nleaves; i++){
142ca2fc57aSFande Kong 	 remote[i].rank  = sources_sc_rd[i];
143ca2fc57aSFande Kong 	 remote[i].index = localoffsets[sources_sc_rd[i]]++;
1442452736bSFande Kong    }
145ca2fc57aSFande Kong    ierr = PetscFree(localoffsets);CHKERRQ(ierr);
146a69400e0SFande Kong #if 0
147a69400e0SFande Kong    if(grank==2){
148a69400e0SFande Kong 	 ierr = PetscIntView(localsize_tmp,indices_ov_rd,PETSC_VIEWER_STDOUT_SELF);CHKERRQ(ierr);
149a69400e0SFande Kong    }
150a69400e0SFande Kong #endif
1512452736bSFande Kong   }else{
1522452736bSFande Kong    ierr = ISDestroy(&allis_sc);CHKERRQ(ierr);
1532452736bSFande Kong    nleaves = 0;
1542452736bSFande Kong    indices_ov_rd = 0;
1552452736bSFande Kong    sources_sc_rd = 0;
1562452736bSFande Kong   }
1572452736bSFande Kong   /*scatter sizes to everybody */
1582452736bSFande Kong   ierr = MPI_Scatter(localsizes_sc,1, MPIU_INT,&nroots,1, MPIU_INT,0,scomm);CHKERRQ(ierr);
159ca2fc57aSFande Kong   /*free memory */
160ca2fc57aSFande Kong   ierr = PetscFree(localsizes_sc);CHKERRQ(ierr);
1612452736bSFande Kong   ierr = PetscCalloc1(nroots,&indices_recv);CHKERRQ(ierr);
162a69400e0SFande Kong   /*ierr = MPI_Comm_dup(scomm,&dcomm);CHKERRQ(ierr);*/
1632452736bSFande Kong   /*set data back to every body */
1642452736bSFande Kong   ierr = PetscSFCreate(scomm,&sf);CHKERRQ(ierr);
1652452736bSFande Kong   ierr = PetscSFSetType(sf,PETSCSFBASIC);CHKERRQ(ierr);
1662452736bSFande Kong   ierr = PetscSFSetFromOptions(sf);CHKERRQ(ierr);
1672452736bSFande Kong   ierr = PetscSFSetGraph(sf,nroots,nleaves,PETSC_NULL,PETSC_OWN_POINTER,remote,PETSC_OWN_POINTER);CHKERRQ(ierr);
168a69400e0SFande Kong #if 0
169a69400e0SFande Kong   ierr = PetscSFView(sf,PETSC_NULL);CHKERRQ(ierr);
170a69400e0SFande Kong   ierr = MPI_Barrier(gcomm);CHKERRQ(ierr);
171a69400e0SFande Kong #endif
172a69400e0SFande Kong   ierr = PetscSFReduceBegin(sf,MPIU_INT,indices_ov_rd,indices_recv,MPIU_REPLACE);CHKERRQ(ierr);
173a69400e0SFande Kong   ierr = PetscSFReduceEnd(sf,MPIU_INT,indices_ov_rd,indices_recv,MPIU_REPLACE);CHKERRQ(ierr);
1742452736bSFande Kong   ierr = PetscSFDestroy(&sf);CHKERRQ(ierr);
1752452736bSFande Kong   /* free memory */
1762452736bSFande Kong   ierr = PetscFree2(indices_ov_rd,sources_sc_rd);CHKERRQ(ierr);
1772452736bSFande Kong   /*create a index set*/
1782452736bSFande Kong   ierr = ISCreateGeneral(scomm,nroots,indices_recv,PETSC_OWN_POINTER,&is_sc);CHKERRQ(ierr);
179*bdeb5f4eSFande Kong #if 1
180*bdeb5f4eSFande Kong   ierr = ISView(is_sc,PETSC_NULL);CHKERRQ(ierr);
181a69400e0SFande Kong   ierr = MPI_Barrier(gcomm);CHKERRQ(ierr);
182a69400e0SFande Kong #endif
1832452736bSFande Kong   /*create a index set for cols */
184a69400e0SFande Kong   ierr = ISAllGather(is_sc,&allis_sc_tmp);CHKERRQ(ierr);
185a69400e0SFande Kong   ierr = ISGetLocalSize(allis_sc_tmp,&localsize_tmp);CHKERRQ(ierr);
186a69400e0SFande Kong   ierr = ISGetIndices(allis_sc_tmp,&indices);CHKERRQ(ierr);
187a69400e0SFande Kong   ierr = ISCreateGeneral(scomm,localsize_tmp,indices,PETSC_COPY_VALUES,&allis_sc);CHKERRQ(ierr);
188a69400e0SFande Kong   ierr = ISRestoreIndices(allis_sc_tmp,&indices);CHKERRQ(ierr);
189a69400e0SFande Kong   ierr = ISDestroy(&allis_sc_tmp);CHKERRQ(ierr);
1902452736bSFande Kong   /*construct a parallel submatrix */
1912452736bSFande Kong   ierr = PetscCalloc1(1,&smat);CHKERRQ(ierr);
192a69400e0SFande Kong #if 0
193a69400e0SFande Kong   ierr = ISView(allis_sc,PETSC_NULL);CHKERRQ(ierr);
194a69400e0SFande Kong   ierr = MPI_Barrier(gcomm);CHKERRQ(ierr);
195a69400e0SFande Kong   //ierr = ISView(is_sc,PETSC_VIEWER_STDOUT_SELF);CHKERRQ(ierr);
196a69400e0SFande Kong   MPI_Comm   comm1, comm2;
197a69400e0SFande Kong   ierr = PetscObjectGetComm((PetscObject)is_sc,&comm1);CHKERRQ(ierr);
198a69400e0SFande Kong   ierr = PetscObjectGetComm((PetscObject)allis_sc,&comm2);CHKERRQ(ierr);
199a69400e0SFande Kong   /*ierr = PetscCommDuplicate(comm1,&comm2,NULL);CHKERRQ(ierr);*/
200a69400e0SFande Kong   /*ierr = MPI_Comm_dup(comm1,&comm2);CHKERRQ(ierr);*/
201a69400e0SFande Kong   ierr = MPI_Comm_compare(comm2,comm1,&issamecomm);CHKERRQ(ierr);
202a69400e0SFande Kong   if(issamecomm == MPI_IDENT){
203a69400e0SFande Kong     ierr=PetscPrintf(gcomm,"the same communicator \n");CHKERRQ(ierr);
204a69400e0SFande Kong   }else{
205a69400e0SFande Kong   	ierr=PetscPrintf(gcomm,"different communicator \n");CHKERRQ(ierr);
206a69400e0SFande Kong   }
207a69400e0SFande Kong #endif
208a69400e0SFande Kong #if 1
209a69400e0SFande Kong   ierr = ISView(allis_sc,PETSC_NULL);CHKERRQ(ierr);
210a69400e0SFande Kong   ierr = ISView(is_sc,PETSC_NULL);CHKERRQ(ierr);
211a69400e0SFande Kong #endif
2122452736bSFande Kong   ierr = MatGetSubMatricesMPI(mat,1,&is_sc,&allis_sc,MAT_INITIAL_MATRIX,&smat);CHKERRQ(ierr);
2132452736bSFande Kong   /* we do not need them any more */
2142452736bSFande Kong   ierr = ISDestroy(&allis_sc);CHKERRQ(ierr);
215a69400e0SFande Kong #if 1
216a69400e0SFande Kong   ierr = MatView(smat[0],PETSC_NULL);CHKERRQ(ierr);
217a69400e0SFande Kong #endif
2182452736bSFande Kong   /*create a partitioner to repartition the sub-matrix*/
2192452736bSFande Kong   ierr = MatPartitioningCreate(scomm,&part);CHKERRQ(ierr);
2202452736bSFande Kong   ierr = MatPartitioningSetAdjacency(part,smat[0]);CHKERRQ(ierr);
2212452736bSFande Kong #if PETSC_HAVE_PARMETIS
2222452736bSFande Kong   /* if there exists a ParMETIS installation, we try to use ParMETIS
2232452736bSFande Kong    * because a repartition routine possibly work better
2242452736bSFande Kong    * */
2252452736bSFande Kong   ierr = MatPartitioningSetType(part,MATPARTITIONINGPARMETIS);CHKERRQ(ierr);
2262452736bSFande Kong   /*try to use reparition function, instead of partition function */
2272452736bSFande Kong   ierr = MatPartitioningParmetisSetRepartition(part);CHKERRQ(ierr);
2282452736bSFande Kong #else
2292452736bSFande Kong   /*we at least provide a default partitioner to rebalance the computation  */
2302452736bSFande Kong   ierr = MatPartitioningSetType(part,MATPARTITIONINGAVERAGE);CHKERRQ(ierr);
2312452736bSFande Kong #endif
2322452736bSFande Kong   /*user can pick up any partitioner by using an option*/
2332452736bSFande Kong   ierr = MatPartitioningSetFromOptions(part);CHKERRQ(ierr);
2342452736bSFande Kong   /* apply partition */
2352452736bSFande Kong   ierr = MatPartitioningApply(part,&partitioning);CHKERRQ(ierr);
2362452736bSFande Kong   ierr = MatPartitioningDestroy(&part);CHKERRQ(ierr);
237ca2fc57aSFande Kong   ierr = MatDestroy(&(smat[0]));CHKERRQ(ierr);
2382452736bSFande Kong   ierr = PetscFree(smat);CHKERRQ(ierr);
239*bdeb5f4eSFande Kong #if 1
240*bdeb5f4eSFande Kong   ierr = ISView(partitioning,PETSC_NULL);CHKERRQ(ierr);
241*bdeb5f4eSFande Kong #endif
2422452736bSFande Kong   /* get local rows including  overlap */
243*bdeb5f4eSFande Kong   ierr = ISBuildTwoSided(partitioning,is_sc,is);CHKERRQ(ierr);
244*bdeb5f4eSFande Kong   /* destroy */
245*bdeb5f4eSFande Kong   ierr = ISDestroy(&is_sc);CHKERRQ(ierr);
2462452736bSFande Kong   PetscFunctionReturn(0);
2472452736bSFande Kong }
2482452736bSFande Kong 
2492452736bSFande Kong 
250