1 static char help[] = "Test that MatPartitioning and PetscPartitioner interfaces are equivalent when using PETSCPARTITIONERMATPARTITIONING\n\n"; 2 static char FILENAME[] = "ex24.c"; 3 4 #include <petscdmplex.h> 5 #include <petscviewerhdf5.h> 6 7 #if defined(PETSC_HAVE_PTSCOTCH) 8 EXTERN_C_BEGIN 9 #include <ptscotch.h> 10 EXTERN_C_END 11 #endif 12 13 typedef struct { 14 PetscBool compare_is; /* Compare ISs and PetscSections */ 15 PetscBool compare_dm; /* Compare DM */ 16 PetscBool tpw; /* Use target partition weights */ 17 char partitioning[64]; 18 char repartitioning[64]; 19 } AppCtx; 20 21 static PetscErrorCode ProcessOptions(MPI_Comm comm, AppCtx *options) 22 { 23 PetscBool repartition = PETSC_TRUE; 24 25 PetscFunctionBegin; 26 options->compare_is = PETSC_FALSE; 27 options->compare_dm = PETSC_FALSE; 28 29 PetscOptionsBegin(comm, "", "Meshing Interpolation Test Options", "DMPLEX"); 30 PetscCall(PetscOptionsBool("-compare_is", "Compare ISs and PetscSections?", FILENAME, options->compare_is, &options->compare_is, NULL)); 31 PetscCall(PetscOptionsBool("-compare_dm", "Compare DMs?", FILENAME, options->compare_dm, &options->compare_dm, NULL)); 32 PetscCall(PetscStrncpy(options->partitioning, MATPARTITIONINGPARMETIS, sizeof(options->partitioning))); 33 PetscCall(PetscOptionsString("-partitioning", "The mat partitioning type to test", "None", options->partitioning, options->partitioning, sizeof(options->partitioning), NULL)); 34 PetscCall(PetscOptionsBool("-repartition", "Partition again after the first partition?", FILENAME, repartition, &repartition, NULL)); 35 if (repartition) { 36 PetscCall(PetscStrncpy(options->repartitioning, MATPARTITIONINGPARMETIS, 64)); 37 PetscCall(PetscOptionsString("-repartitioning", "The mat partitioning type to test (second partitioning)", "None", options->repartitioning, options->repartitioning, sizeof(options->repartitioning), NULL)); 38 } else { 39 options->repartitioning[0] = '\0'; 40 } 41 PetscCall(PetscOptionsBool("-tpweight", "Use target partition weights", FILENAME, options->tpw, &options->tpw, NULL)); 42 PetscOptionsEnd(); 43 PetscFunctionReturn(0); 44 } 45 46 static PetscErrorCode ScotchResetRandomSeed() 47 { 48 PetscFunctionBegin; 49 #if defined(PETSC_HAVE_PTSCOTCH) 50 SCOTCH_randomReset(); 51 #endif 52 PetscFunctionReturn(0); 53 } 54 55 static PetscErrorCode CreateMesh(MPI_Comm comm, AppCtx *user, DM *dm) 56 { 57 PetscFunctionBegin; 58 PetscCall(DMCreate(comm, dm)); 59 PetscCall(DMSetType(*dm, DMPLEX)); 60 PetscCall(DMPlexDistributeSetDefault(*dm, PETSC_FALSE)); 61 PetscCall(DMSetFromOptions(*dm)); 62 PetscCall(DMViewFromOptions(*dm, NULL, "-dm_view")); 63 PetscFunctionReturn(0); 64 } 65 66 int main(int argc, char **argv) 67 { 68 MPI_Comm comm; 69 DM dm1, dm2, dmdist1, dmdist2; 70 DMPlexInterpolatedFlag interp; 71 MatPartitioning mp; 72 PetscPartitioner part1, part2; 73 AppCtx user; 74 IS is1 = NULL, is2 = NULL; 75 IS is1g, is2g; 76 PetscSection s1 = NULL, s2 = NULL, tpws = NULL; 77 PetscInt i; 78 PetscBool flg; 79 PetscMPIInt size; 80 81 PetscFunctionBeginUser; 82 PetscCall(PetscInitialize(&argc, &argv, NULL, help)); 83 comm = PETSC_COMM_WORLD; 84 PetscCallMPI(MPI_Comm_size(comm, &size)); 85 PetscCall(ProcessOptions(comm, &user)); 86 PetscCall(CreateMesh(comm, &user, &dm1)); 87 PetscCall(CreateMesh(comm, &user, &dm2)); 88 89 if (user.tpw) { 90 PetscCall(PetscSectionCreate(comm, &tpws)); 91 PetscCall(PetscSectionSetChart(tpws, 0, size)); 92 for (i = 0; i < size; i++) { 93 PetscInt tdof = i % 2 ? 2 * i - 1 : i + 2; 94 PetscCall(PetscSectionSetDof(tpws, i, tdof)); 95 } 96 if (size > 1) { /* test zero tpw entry */ 97 PetscCall(PetscSectionSetDof(tpws, 0, 0)); 98 } 99 PetscCall(PetscSectionSetUp(tpws)); 100 } 101 102 /* partition dm1 using PETSCPARTITIONERPARMETIS */ 103 PetscCall(ScotchResetRandomSeed()); 104 PetscCall(DMPlexGetPartitioner(dm1, &part1)); 105 PetscCall(PetscObjectSetOptionsPrefix((PetscObject)part1, "p1_")); 106 PetscCall(PetscPartitionerSetType(part1, user.partitioning)); 107 PetscCall(PetscPartitionerSetFromOptions(part1)); 108 PetscCall(PetscSectionCreate(comm, &s1)); 109 PetscCall(PetscPartitionerDMPlexPartition(part1, dm1, tpws, s1, &is1)); 110 111 /* partition dm2 using PETSCPARTITIONERMATPARTITIONING with MATPARTITIONINGPARMETIS */ 112 PetscCall(ScotchResetRandomSeed()); 113 PetscCall(DMPlexGetPartitioner(dm2, &part2)); 114 PetscCall(PetscObjectSetOptionsPrefix((PetscObject)part2, "p2_")); 115 PetscCall(PetscPartitionerSetType(part2, PETSCPARTITIONERMATPARTITIONING)); 116 PetscCall(PetscPartitionerMatPartitioningGetMatPartitioning(part2, &mp)); 117 PetscCall(MatPartitioningSetType(mp, user.partitioning)); 118 PetscCall(PetscPartitionerSetFromOptions(part2)); 119 PetscCall(PetscSectionCreate(comm, &s2)); 120 PetscCall(PetscPartitionerDMPlexPartition(part2, dm2, tpws, s2, &is2)); 121 122 PetscCall(ISOnComm(is1, comm, PETSC_USE_POINTER, &is1g)); 123 PetscCall(ISOnComm(is2, comm, PETSC_USE_POINTER, &is2g)); 124 PetscCall(ISViewFromOptions(is1g, NULL, "-seq_is1_view")); 125 PetscCall(ISViewFromOptions(is2g, NULL, "-seq_is2_view")); 126 /* compare the two ISs */ 127 if (user.compare_is) { 128 PetscCall(ISEqualUnsorted(is1g, is2g, &flg)); 129 if (!flg) PetscCall(PetscPrintf(comm, "ISs are not equal with type %s with size %d.\n", user.partitioning, size)); 130 } 131 PetscCall(ISDestroy(&is1g)); 132 PetscCall(ISDestroy(&is2g)); 133 134 /* compare the two PetscSections */ 135 PetscCall(PetscSectionViewFromOptions(s1, NULL, "-seq_s1_view")); 136 PetscCall(PetscSectionViewFromOptions(s2, NULL, "-seq_s2_view")); 137 if (user.compare_is) { 138 PetscCall(PetscSectionCompare(s1, s2, &flg)); 139 if (!flg) PetscCall(PetscPrintf(comm, "PetscSections are not equal with %s with size %d.\n", user.partitioning, size)); 140 } 141 142 /* distribute both DMs */ 143 PetscCall(ScotchResetRandomSeed()); 144 PetscCall(DMPlexDistribute(dm1, 0, NULL, &dmdist1)); 145 PetscCall(ScotchResetRandomSeed()); 146 PetscCall(DMPlexDistribute(dm2, 0, NULL, &dmdist2)); 147 148 /* cleanup */ 149 PetscCall(PetscSectionDestroy(&tpws)); 150 PetscCall(PetscSectionDestroy(&s1)); 151 PetscCall(PetscSectionDestroy(&s2)); 152 PetscCall(ISDestroy(&is1)); 153 PetscCall(ISDestroy(&is2)); 154 PetscCall(DMDestroy(&dm1)); 155 PetscCall(DMDestroy(&dm2)); 156 157 /* if distributed DMs are NULL (sequential case), then quit */ 158 if (!dmdist1 && !dmdist2) return 0; 159 160 PetscCall(DMViewFromOptions(dmdist1, NULL, "-dm_dist1_view")); 161 PetscCall(DMViewFromOptions(dmdist2, NULL, "-dm_dist2_view")); 162 163 /* compare the two distributed DMs */ 164 if (user.compare_dm) { 165 PetscCall(DMPlexEqual(dmdist1, dmdist2, &flg)); 166 if (!flg) PetscCall(PetscPrintf(comm, "Distributed DMs are not equal %s with size %d.\n", user.partitioning, size)); 167 } 168 169 /* if repartitioning is disabled, then quit */ 170 if (user.repartitioning[0] == '\0') return 0; 171 172 if (user.tpw) { 173 PetscCall(PetscSectionCreate(comm, &tpws)); 174 PetscCall(PetscSectionSetChart(tpws, 0, size)); 175 for (i = 0; i < size; i++) { 176 PetscInt tdof = i % 2 ? i + 1 : size - i; 177 PetscCall(PetscSectionSetDof(tpws, i, tdof)); 178 } 179 PetscCall(PetscSectionSetUp(tpws)); 180 } 181 182 /* repartition distributed DM dmdist1 */ 183 PetscCall(ScotchResetRandomSeed()); 184 PetscCall(DMPlexGetPartitioner(dmdist1, &part1)); 185 PetscCall(PetscObjectSetOptionsPrefix((PetscObject)part1, "dp1_")); 186 PetscCall(PetscPartitionerSetType(part1, user.repartitioning)); 187 PetscCall(PetscPartitionerSetFromOptions(part1)); 188 PetscCall(PetscSectionCreate(comm, &s1)); 189 PetscCall(PetscPartitionerDMPlexPartition(part1, dmdist1, tpws, s1, &is1)); 190 191 /* repartition distributed DM dmdist2 */ 192 PetscCall(ScotchResetRandomSeed()); 193 PetscCall(DMPlexGetPartitioner(dmdist2, &part2)); 194 PetscCall(PetscObjectSetOptionsPrefix((PetscObject)part2, "dp2_")); 195 PetscCall(PetscPartitionerSetType(part2, PETSCPARTITIONERMATPARTITIONING)); 196 PetscCall(PetscPartitionerMatPartitioningGetMatPartitioning(part2, &mp)); 197 PetscCall(MatPartitioningSetType(mp, user.repartitioning)); 198 PetscCall(PetscPartitionerSetFromOptions(part2)); 199 PetscCall(PetscSectionCreate(comm, &s2)); 200 PetscCall(PetscPartitionerDMPlexPartition(part2, dmdist2, tpws, s2, &is2)); 201 202 /* compare the two ISs */ 203 PetscCall(ISOnComm(is1, comm, PETSC_USE_POINTER, &is1g)); 204 PetscCall(ISOnComm(is2, comm, PETSC_USE_POINTER, &is2g)); 205 PetscCall(ISViewFromOptions(is1g, NULL, "-dist_is1_view")); 206 PetscCall(ISViewFromOptions(is2g, NULL, "-dist_is2_view")); 207 if (user.compare_is) { 208 PetscCall(ISEqualUnsorted(is1g, is2g, &flg)); 209 if (!flg) PetscCall(PetscPrintf(comm, "Distributed ISs are not equal, with %s with size %d.\n", user.repartitioning, size)); 210 } 211 PetscCall(ISDestroy(&is1g)); 212 PetscCall(ISDestroy(&is2g)); 213 214 /* compare the two PetscSections */ 215 PetscCall(PetscSectionViewFromOptions(s1, NULL, "-dist_s1_view")); 216 PetscCall(PetscSectionViewFromOptions(s2, NULL, "-dist_s2_view")); 217 if (user.compare_is) { 218 PetscCall(PetscSectionCompare(s1, s2, &flg)); 219 if (!flg) PetscCall(PetscPrintf(comm, "Distributed PetscSections are not equal, with %s with size %d.\n", user.repartitioning, size)); 220 } 221 222 /* redistribute both distributed DMs */ 223 PetscCall(ScotchResetRandomSeed()); 224 PetscCall(DMPlexDistribute(dmdist1, 0, NULL, &dm1)); 225 PetscCall(ScotchResetRandomSeed()); 226 PetscCall(DMPlexDistribute(dmdist2, 0, NULL, &dm2)); 227 228 /* compare the two distributed DMs */ 229 PetscCall(DMPlexIsInterpolated(dm1, &interp)); 230 if (interp == DMPLEX_INTERPOLATED_NONE) { 231 PetscCall(DMPlexEqual(dm1, dm2, &flg)); 232 if (!flg) PetscCall(PetscPrintf(comm, "Redistributed DMs are not equal, with %s with size %d.\n", user.repartitioning, size)); 233 } 234 235 /* cleanup */ 236 PetscCall(PetscSectionDestroy(&tpws)); 237 PetscCall(PetscSectionDestroy(&s1)); 238 PetscCall(PetscSectionDestroy(&s2)); 239 PetscCall(ISDestroy(&is1)); 240 PetscCall(ISDestroy(&is2)); 241 PetscCall(DMDestroy(&dm1)); 242 PetscCall(DMDestroy(&dm2)); 243 PetscCall(DMDestroy(&dmdist1)); 244 PetscCall(DMDestroy(&dmdist2)); 245 PetscCall(PetscFinalize()); 246 return 0; 247 } 248 249 /*TEST 250 251 test: 252 # partition sequential mesh loaded from Exodus file 253 suffix: 0 254 nsize: {{1 2 3 4 8}} 255 requires: chaco parmetis ptscotch exodusii 256 args: -dm_plex_filename ${PETSC_DIR}/share/petsc/datafiles/meshes/blockcylinder-50.exo 257 args: -partitioning {{chaco parmetis ptscotch}} -repartitioning {{parmetis ptscotch}} -tpweight {{0 1}} 258 test: 259 # repartition mesh already partitioned naively by MED loader 260 suffix: 1 261 nsize: {{1 2 3 4 8}} 262 TODO: MED 263 requires: parmetis ptscotch med 264 args: -dm_plex_filename ${PETSC_DIR}/share/petsc/datafiles/meshes/cylinder.med 265 args: -repartition 0 -partitioning {{parmetis ptscotch}} 266 test: 267 # partition mesh generated by ctetgen using scotch, then repartition with scotch, diff view 268 suffix: 3 269 nsize: 4 270 requires: ptscotch ctetgen 271 args: -dm_plex_dim 3 -dm_plex_box_faces 2,3,2 -partitioning ptscotch -repartitioning ptscotch 272 args: -p1_petscpartitioner_view -p2_petscpartitioner_view -dp1_petscpartitioner_view -dp2_petscpartitioner_view -tpweight {{0 1}} 273 test: 274 # partition mesh generated by ctetgen using partitioners supported both by MatPartitioning and PetscPartitioner 275 suffix: 4 276 nsize: {{1 2 3 4 8}} 277 requires: chaco parmetis ptscotch ctetgen 278 args: -dm_plex_dim 3 -dm_plex_box_faces {{2,3,4 5,4,3 7,11,5}} -partitioning {{chaco parmetis ptscotch}} -repartitioning {{parmetis ptscotch}} -tpweight {{0 1}} 279 280 TEST*/ 281