1 2 static char help[] = "Tests MPI parallel matrix creation. Test MatCreateRedundantMatrix() \n\n"; 3 4 #include <petscmat.h> 5 6 int main(int argc,char **args) 7 { 8 Mat C,Credundant; 9 MatInfo info; 10 PetscMPIInt rank,size,subsize; 11 PetscInt i,j,m = 3,n = 2,low,high,iglobal; 12 PetscInt Ii,J,ldim,nsubcomms; 13 PetscErrorCode ierr; 14 PetscBool flg_info,flg_mat; 15 PetscScalar v,one = 1.0; 16 Vec x,y; 17 MPI_Comm subcomm; 18 19 ierr = PetscInitialize(&argc,&args,(char*)0,help);if (ierr) return ierr; 20 ierr = PetscOptionsGetInt(NULL,NULL,"-m",&m,NULL);CHKERRQ(ierr); 21 ierr = MPI_Comm_rank(PETSC_COMM_WORLD,&rank);CHKERRQ(ierr); 22 ierr = MPI_Comm_size(PETSC_COMM_WORLD,&size);CHKERRQ(ierr); 23 n = 2*size; 24 25 ierr = MatCreate(PETSC_COMM_WORLD,&C);CHKERRQ(ierr); 26 ierr = MatSetSizes(C,PETSC_DECIDE,PETSC_DECIDE,m*n,m*n);CHKERRQ(ierr); 27 ierr = MatSetFromOptions(C);CHKERRQ(ierr); 28 ierr = MatSetUp(C);CHKERRQ(ierr); 29 30 /* Create the matrix for the five point stencil, YET AGAIN */ 31 for (i=0; i<m; i++) { 32 for (j=2*rank; j<2*rank+2; j++) { 33 v = -1.0; Ii = j + n*i; 34 if (i>0) {J = Ii - n; ierr = MatSetValues(C,1,&Ii,1,&J,&v,INSERT_VALUES);CHKERRQ(ierr);} 35 if (i<m-1) {J = Ii + n; ierr = MatSetValues(C,1,&Ii,1,&J,&v,INSERT_VALUES);CHKERRQ(ierr);} 36 if (j>0) {J = Ii - 1; ierr = MatSetValues(C,1,&Ii,1,&J,&v,INSERT_VALUES);CHKERRQ(ierr);} 37 if (j<n-1) {J = Ii + 1; ierr = MatSetValues(C,1,&Ii,1,&J,&v,INSERT_VALUES);CHKERRQ(ierr);} 38 v = 4.0; ierr = MatSetValues(C,1,&Ii,1,&Ii,&v,INSERT_VALUES);CHKERRQ(ierr); 39 } 40 } 41 42 /* Add extra elements (to illustrate variants of MatGetInfo) */ 43 Ii = n; J = n-2; v = 100.0; 44 ierr = MatSetValues(C,1,&Ii,1,&J,&v,INSERT_VALUES);CHKERRQ(ierr); 45 Ii = n-2; J = n; v = 100.0; 46 ierr = MatSetValues(C,1,&Ii,1,&J,&v,INSERT_VALUES);CHKERRQ(ierr); 47 48 ierr = MatAssemblyBegin(C,MAT_FINAL_ASSEMBLY);CHKERRQ(ierr); 49 ierr = MatAssemblyEnd(C,MAT_FINAL_ASSEMBLY);CHKERRQ(ierr); 50 51 /* Form vectors */ 52 ierr = MatCreateVecs(C,&x,&y);CHKERRQ(ierr); 53 ierr = VecGetLocalSize(x,&ldim);CHKERRQ(ierr); 54 ierr = VecGetOwnershipRange(x,&low,&high);CHKERRQ(ierr); 55 for (i=0; i<ldim; i++) { 56 iglobal = i + low; 57 v = one*((PetscReal)i) + 100.0*rank; 58 ierr = VecSetValues(x,1,&iglobal,&v,INSERT_VALUES);CHKERRQ(ierr); 59 } 60 ierr = VecAssemblyBegin(x);CHKERRQ(ierr); 61 ierr = VecAssemblyEnd(x);CHKERRQ(ierr); 62 63 ierr = MatMult(C,x,y);CHKERRQ(ierr); 64 65 ierr = PetscOptionsHasName(NULL,NULL,"-view_info",&flg_info);CHKERRQ(ierr); 66 if (flg_info) { 67 ierr = PetscViewerPushFormat(PETSC_VIEWER_STDOUT_WORLD,PETSC_VIEWER_ASCII_INFO);CHKERRQ(ierr); 68 ierr = MatView(C,PETSC_VIEWER_STDOUT_WORLD);CHKERRQ(ierr); 69 70 ierr = MatGetInfo(C,MAT_GLOBAL_SUM,&info);CHKERRQ(ierr); 71 ierr = PetscViewerASCIIPrintf(PETSC_VIEWER_STDOUT_WORLD,"matrix information (global sums):\nnonzeros = %D, allocated nonzeros = %D\n",(PetscInt)info.nz_used,(PetscInt)info.nz_allocated);CHKERRQ(ierr); 72 ierr = MatGetInfo (C,MAT_GLOBAL_MAX,&info);CHKERRQ(ierr); 73 ierr = PetscViewerASCIIPrintf(PETSC_VIEWER_STDOUT_WORLD,"matrix information (global max):\nnonzeros = %D, allocated nonzeros = %D\n",(PetscInt)info.nz_used,(PetscInt)info.nz_allocated);CHKERRQ(ierr); 74 } 75 76 ierr = PetscOptionsHasName(NULL,NULL,"-view_mat",&flg_mat);CHKERRQ(ierr); 77 if (flg_mat) { 78 ierr = MatView(C,PETSC_VIEWER_STDOUT_WORLD);CHKERRQ(ierr); 79 } 80 81 /* Test MatCreateRedundantMatrix() */ 82 nsubcomms = size; 83 ierr = PetscOptionsGetInt(NULL,NULL,"-nsubcomms",&nsubcomms,NULL);CHKERRQ(ierr); 84 ierr = MatCreateRedundantMatrix(C,nsubcomms,MPI_COMM_NULL,MAT_INITIAL_MATRIX,&Credundant);CHKERRQ(ierr); 85 ierr = MatCreateRedundantMatrix(C,nsubcomms,MPI_COMM_NULL,MAT_REUSE_MATRIX,&Credundant);CHKERRQ(ierr); 86 87 ierr = PetscObjectGetComm((PetscObject)Credundant,&subcomm);CHKERRQ(ierr); 88 ierr = MPI_Comm_size(subcomm,&subsize);CHKERRQ(ierr); 89 90 if (subsize==2 && flg_mat) { 91 ierr = PetscViewerASCIIPrintf(PETSC_VIEWER_STDOUT_(subcomm),"\n[%d] Credundant:\n",rank);CHKERRQ(ierr); 92 ierr = MatView(Credundant,PETSC_VIEWER_STDOUT_(subcomm));CHKERRQ(ierr); 93 } 94 ierr = MatDestroy(&Credundant);CHKERRQ(ierr); 95 96 /* Test MatCreateRedundantMatrix() with user-provided subcomm */ 97 { 98 PetscSubcomm psubcomm; 99 100 ierr = PetscSubcommCreate(PETSC_COMM_WORLD,&psubcomm);CHKERRQ(ierr); 101 ierr = PetscSubcommSetNumber(psubcomm,nsubcomms);CHKERRQ(ierr); 102 ierr = PetscSubcommSetType(psubcomm,PETSC_SUBCOMM_CONTIGUOUS);CHKERRQ(ierr); 103 /* enable runtime switch of psubcomm type, e.g., '-psubcomm_type interlaced */ 104 ierr = PetscSubcommSetFromOptions(psubcomm);CHKERRQ(ierr); 105 106 ierr = MatCreateRedundantMatrix(C,nsubcomms,PetscSubcommChild(psubcomm),MAT_INITIAL_MATRIX,&Credundant);CHKERRQ(ierr); 107 ierr = MatCreateRedundantMatrix(C,nsubcomms,PetscSubcommChild(psubcomm),MAT_REUSE_MATRIX,&Credundant);CHKERRQ(ierr); 108 109 ierr = PetscSubcommDestroy(&psubcomm);CHKERRQ(ierr); 110 ierr = MatDestroy(&Credundant);CHKERRQ(ierr); 111 } 112 113 ierr = VecDestroy(&x);CHKERRQ(ierr); 114 ierr = VecDestroy(&y);CHKERRQ(ierr); 115 ierr = MatDestroy(&C);CHKERRQ(ierr); 116 ierr = PetscFinalize(); 117 return ierr; 118 } 119 120 121 /*TEST 122 123 test: 124 nsize: 3 125 args: -view_info 126 127 test: 128 suffix: 2 129 nsize: 3 130 args: -nsubcomms 2 -view_mat -psubcomm_type interlaced 131 132 test: 133 suffix: 3 134 nsize: 3 135 args: -nsubcomms 2 -view_mat -psubcomm_type contiguous 136 137 test: 138 suffix: 3_baij 139 nsize: 3 140 args: -mat_type baij -nsubcomms 2 -view_mat 141 142 test: 143 suffix: 3_sbaij 144 nsize: 3 145 args: -mat_type sbaij -nsubcomms 2 -view_mat 146 147 test: 148 suffix: 3_dense 149 nsize: 3 150 args: -mat_type dense -nsubcomms 2 -view_mat 151 152 test: 153 suffix: 4_baij 154 nsize: 3 155 args: -mat_type baij -nsubcomms 2 -view_mat -psubcomm_type interlaced 156 157 test: 158 suffix: 4_sbaij 159 nsize: 3 160 args: -mat_type sbaij -nsubcomms 2 -view_mat -psubcomm_type interlaced 161 162 test: 163 suffix: 4_dense 164 nsize: 3 165 args: -mat_type dense -nsubcomms 2 -view_mat -psubcomm_type interlaced 166 167 TEST*/ 168