1 static char help[] = "3D, tensor hexahedra (Q1-K), displacement finite element formulation\n\ 2 of linear elasticity. E=1.0, nu=1/3.\n\ 3 Unit cube domain with Dirichlet boundary\n\n"; 4 5 #include <petscdmplex.h> 6 #include <petscsnes.h> 7 #include <petscds.h> 8 #include <petscdmforest.h> 9 10 static PetscReal s_soft_alpha=1.e-3; 11 static PetscReal s_mu=0.4; 12 static PetscReal s_lambda=0.4; 13 14 static void f0_bd_u_3d(PetscInt dim, PetscInt Nf, PetscInt NfAux, 15 const PetscInt uOff[], const PetscInt uOff_x[], const PetscScalar u[], const PetscScalar u_t[], const PetscScalar u_x[], 16 const PetscInt aOff[], const PetscInt aOff_x[], const PetscScalar a[], const PetscScalar a_t[], const PetscScalar a_x[], 17 PetscReal t, const PetscReal x[], const PetscReal n[], PetscInt numConstants, const PetscScalar constants[], PetscScalar f0[]) 18 { 19 f0[0] = 1; /* x direction pull */ 20 f0[1] = -x[2]; /* add a twist around x-axis */ 21 f0[2] = x[1]; 22 } 23 24 static void f1_bd_u(PetscInt dim, PetscInt Nf, PetscInt NfAux, 25 const PetscInt uOff[], const PetscInt uOff_x[], const PetscScalar u[], const PetscScalar u_t[], const PetscScalar u_x[], 26 const PetscInt aOff[], const PetscInt aOff_x[], const PetscScalar a[], const PetscScalar a_t[], const PetscScalar a_x[], 27 PetscReal t, const PetscReal x[], const PetscReal n[], PetscInt numConstants, const PetscScalar constants[], PetscScalar f1[]) 28 { 29 const PetscInt Ncomp = dim; 30 PetscInt comp, d; 31 for (comp = 0; comp < Ncomp; ++comp) { 32 for (d = 0; d < dim; ++d) { 33 f1[comp*dim+d] = 0.0; 34 } 35 } 36 } 37 38 /* gradU[comp*dim+d] = {u_x, u_y} or {u_x, u_y, u_z} */ 39 static void f1_u_3d_alpha(PetscInt dim, PetscInt Nf, PetscInt NfAux, 40 const PetscInt uOff[], const PetscInt uOff_x[], const PetscScalar u[], const PetscScalar u_t[], const PetscScalar u_x[], 41 const PetscInt aOff[], const PetscInt aOff_x[], const PetscScalar a[], const PetscScalar a_t[], const PetscScalar a_x[], 42 PetscReal t, const PetscReal x[], PetscInt numConstants, const PetscScalar constants[], PetscScalar f1[]) 43 { 44 PetscReal trace,mu=s_mu,lambda=s_lambda,rad; 45 PetscInt i,j; 46 for (i=0,rad=0.;i<dim;i++) { 47 PetscReal t=x[i]; 48 rad += t*t; 49 } 50 rad = PetscSqrtReal(rad); 51 if (rad>0.25) { 52 mu *= s_soft_alpha; 53 lambda *= s_soft_alpha; /* we could keep the bulk the same like rubberish */ 54 } 55 for (i=0,trace=0; i < dim; ++i) { 56 trace += PetscRealPart(u_x[i*dim+i]); 57 } 58 for (i=0; i < dim; ++i) { 59 for (j=0; j < dim; ++j) { 60 f1[i*dim+j] = mu*(u_x[i*dim+j]+u_x[j*dim+i]); 61 } 62 f1[i*dim+i] += lambda*trace; 63 } 64 } 65 66 /* gradU[comp*dim+d] = {u_x, u_y} or {u_x, u_y, u_z} */ 67 static void f1_u_3d(PetscInt dim, PetscInt Nf, PetscInt NfAux, 68 const PetscInt uOff[], const PetscInt uOff_x[], const PetscScalar u[], const PetscScalar u_t[], const PetscScalar u_x[], 69 const PetscInt aOff[], const PetscInt aOff_x[], const PetscScalar a[], const PetscScalar a_t[], const PetscScalar a_x[], 70 PetscReal t, const PetscReal x[], PetscInt numConstants, const PetscScalar constants[], PetscScalar f1[]) 71 { 72 PetscReal trace,mu=s_mu,lambda=s_lambda; 73 PetscInt i,j; 74 for (i=0,trace=0; i < dim; ++i) { 75 trace += PetscRealPart(u_x[i*dim+i]); 76 } 77 for (i=0; i < dim; ++i) { 78 for (j=0; j < dim; ++j) { 79 f1[i*dim+j] = mu*(u_x[i*dim+j]+u_x[j*dim+i]); 80 } 81 f1[i*dim+i] += lambda*trace; 82 } 83 } 84 85 /* 3D elasticity */ 86 #define IDX(ii,jj,kk,ll) (27*ii+9*jj+3*kk+ll) 87 88 void g3_uu_3d_private( PetscScalar g3[], const PetscReal mu, const PetscReal lambda) 89 { 90 if (1) { 91 g3[0] += lambda; 92 g3[0] += mu; 93 g3[0] += mu; 94 g3[4] += lambda; 95 g3[8] += lambda; 96 g3[10] += mu; 97 g3[12] += mu; 98 g3[20] += mu; 99 g3[24] += mu; 100 g3[28] += mu; 101 g3[30] += mu; 102 g3[36] += lambda; 103 g3[40] += lambda; 104 g3[40] += mu; 105 g3[40] += mu; 106 g3[44] += lambda; 107 g3[50] += mu; 108 g3[52] += mu; 109 g3[56] += mu; 110 g3[60] += mu; 111 g3[68] += mu; 112 g3[70] += mu; 113 g3[72] += lambda; 114 g3[76] += lambda; 115 g3[80] += lambda; 116 g3[80] += mu; 117 g3[80] += mu; 118 } else { 119 int i,j,k,l; 120 static int cc=-1; 121 cc++; 122 for (i = 0; i < 3; ++i) { 123 for (j = 0; j < 3; ++j) { 124 for (k = 0; k < 3; ++k) { 125 for (l = 0; l < 3; ++l) { 126 if (k==l && i==j) g3[IDX(i,j,k,l)] += lambda; 127 if (i==k && j==l) g3[IDX(i,j,k,l)] += mu; 128 if (i==l && j==k) g3[IDX(i,j,k,l)] += mu; 129 if (k==l && i==j && !cc) (void) PetscPrintf(PETSC_COMM_WORLD,"g3[%d] += lambda;\n",IDX(i,j,k,l)); 130 if (i==k && j==l && !cc) (void) PetscPrintf(PETSC_COMM_WORLD,"g3[%d] += mu;\n",IDX(i,j,k,l)); 131 if (i==l && j==k && !cc) (void) PetscPrintf(PETSC_COMM_WORLD,"g3[%d] += mu;\n",IDX(i,j,k,l)); 132 } 133 } 134 } 135 } 136 } 137 } 138 139 static void g3_uu_3d_alpha(PetscInt dim, PetscInt Nf, PetscInt NfAux, 140 const PetscInt uOff[], const PetscInt uOff_x[], const PetscScalar u[], const PetscScalar u_t[], const PetscScalar u_x[], 141 const PetscInt aOff[], const PetscInt aOff_x[], const PetscScalar a[], const PetscScalar a_t[], const PetscScalar a_x[], 142 PetscReal t, PetscReal u_tShift, const PetscReal x[], PetscInt numConstants, const PetscScalar constants[], PetscScalar g3[]) 143 { 144 PetscReal mu=s_mu, lambda=s_lambda,rad; 145 PetscInt i; 146 for (i=0,rad=0.;i<dim;i++) { 147 PetscReal t=x[i]; 148 rad += t*t; 149 } 150 rad = PetscSqrtReal(rad); 151 if (rad>0.25) { 152 mu *= s_soft_alpha; 153 lambda *= s_soft_alpha; /* we could keep the bulk the same like rubberish */ 154 } 155 g3_uu_3d_private(g3,mu,lambda); 156 } 157 158 static void g3_uu_3d(PetscInt dim, PetscInt Nf, PetscInt NfAux, 159 const PetscInt uOff[], const PetscInt uOff_x[], const PetscScalar u[], const PetscScalar u_t[], const PetscScalar u_x[], 160 const PetscInt aOff[], const PetscInt aOff_x[], const PetscScalar a[], const PetscScalar a_t[], const PetscScalar a_x[], 161 PetscReal t, PetscReal u_tShift, const PetscReal x[], PetscInt numConstants, const PetscScalar constants[], PetscScalar g3[]) 162 { 163 g3_uu_3d_private(g3,s_mu,s_lambda); 164 } 165 166 static void f0_u(PetscInt dim, PetscInt Nf, PetscInt NfAux, 167 const PetscInt uOff[], const PetscInt uOff_x[], const PetscScalar u[], const PetscScalar u_t[], const PetscScalar u_x[], 168 const PetscInt aOff[], const PetscInt aOff_x[], const PetscScalar a[], const PetscScalar a_t[], const PetscScalar a_x[], 169 PetscReal t, const PetscReal x[], PetscInt numConstants, const PetscScalar constants[], PetscScalar f0[]) 170 { 171 const PetscInt Ncomp = dim; 172 PetscInt comp; 173 174 for (comp = 0; comp < Ncomp; ++comp) f0[comp] = 0.0; 175 } 176 177 /* PI_i (x_i^4 - x_i^2) */ 178 static void f0_u_x4(PetscInt dim, PetscInt Nf, PetscInt NfAux, 179 const PetscInt uOff[], const PetscInt uOff_x[], const PetscScalar u[], const PetscScalar u_t[], const PetscScalar u_x[], 180 const PetscInt aOff[], const PetscInt aOff_x[], const PetscScalar a[], const PetscScalar a_t[], const PetscScalar a_x[], 181 PetscReal t, const PetscReal x[], PetscInt numConstants, const PetscScalar constants[], PetscScalar f0[]) 182 { 183 const PetscInt Ncomp = dim; 184 PetscInt comp,i; 185 186 for (comp = 0; comp < Ncomp; ++comp) { 187 f0[comp] = 1e5; 188 for (i = 0; i < Ncomp; ++i) { 189 f0[comp] *= /* (comp+1)* */(x[i]*x[i]*x[i]*x[i] - x[i]*x[i]); /* assumes (0,1]^D domain */ 190 } 191 } 192 } 193 194 PetscErrorCode zero(PetscInt dim, PetscReal time, const PetscReal x[], PetscInt Nf, PetscScalar *u, void *ctx) 195 { 196 const PetscInt Ncomp = dim; 197 PetscInt comp; 198 199 for (comp = 0; comp < Ncomp; ++comp) u[comp] = 0; 200 return 0; 201 } 202 203 int main(int argc,char **args) 204 { 205 Mat Amat; 206 SNES snes; 207 KSP ksp; 208 MPI_Comm comm; 209 PetscMPIInt rank; 210 #if defined(PETSC_USE_LOG) 211 PetscLogStage stage[17]; 212 #endif 213 PetscBool test_nonzero_cols = PETSC_FALSE,use_nearnullspace = PETSC_TRUE,attach_nearnullspace = PETSC_FALSE; 214 Vec xx,bb; 215 PetscInt iter,i,N,dim = 3,cells[3] = {1,1,1},max_conv_its,local_sizes[7],run_type = 1; 216 DM dm,distdm,basedm; 217 PetscBool flg; 218 char convType[256]; 219 PetscReal Lx,mdisp[10],err[10]; 220 const char * const options[10] = {"-ex56_dm_refine 0", 221 "-ex56_dm_refine 1", 222 "-ex56_dm_refine 2", 223 "-ex56_dm_refine 3", 224 "-ex56_dm_refine 4", 225 "-ex56_dm_refine 5", 226 "-ex56_dm_refine 6", 227 "-ex56_dm_refine 7", 228 "-ex56_dm_refine 8", 229 "-ex56_dm_refine 9"}; 230 PetscFunctionBeginUser; 231 PetscFunctionBeginUser; 232 PetscCall(PetscInitialize(&argc,&args,(char*)0,help)); 233 comm = PETSC_COMM_WORLD; 234 PetscCallMPI(MPI_Comm_rank(comm, &rank)); 235 /* options */ 236 PetscOptionsBegin(comm,NULL,"3D bilinear Q1 elasticity options",""); 237 { 238 i = 3; 239 PetscCall(PetscOptionsIntArray("-cells", "Number of (flux tube) processor in each dimension", "ex56.c", cells, &i, NULL)); 240 241 Lx = 1.; /* or ne for rod */ 242 max_conv_its = 3; 243 PetscCall(PetscOptionsInt("-max_conv_its","Number of iterations in convergence study","",max_conv_its,&max_conv_its,NULL)); 244 PetscCheck(max_conv_its > 0 && max_conv_its < 7,PETSC_COMM_WORLD, PETSC_ERR_USER, "Bad number of iterations for convergence test (%" PetscInt_FMT ")",max_conv_its); 245 PetscCall(PetscOptionsReal("-lx","Length of domain","",Lx,&Lx,NULL)); 246 PetscCall(PetscOptionsReal("-alpha","material coefficient inside circle","",s_soft_alpha,&s_soft_alpha,NULL)); 247 PetscCall(PetscOptionsBool("-test_nonzero_cols","nonzero test","",test_nonzero_cols,&test_nonzero_cols,NULL)); 248 PetscCall(PetscOptionsBool("-use_mat_nearnullspace","MatNearNullSpace API test","",use_nearnullspace,&use_nearnullspace,NULL)); 249 PetscCall(PetscOptionsBool("-attach_mat_nearnullspace","MatNearNullSpace API test (via MatSetNearNullSpace)","",attach_nearnullspace,&attach_nearnullspace,NULL)); 250 PetscCall(PetscOptionsInt("-run_type","0: twisting load on cantalever, 1: 3rd order accurate convergence test","",run_type,&run_type,NULL)); 251 } 252 PetscOptionsEnd(); 253 PetscCall(PetscLogStageRegister("Mesh Setup", &stage[16])); 254 for (iter=0 ; iter<max_conv_its ; iter++) { 255 char str[] = "Solve 0"; 256 str[6] += iter; 257 PetscCall(PetscLogStageRegister(str, &stage[iter])); 258 } 259 /* create DM, Plex calls DMSetup */ 260 PetscCall(PetscLogStagePush(stage[16])); 261 PetscCall(DMPlexCreateBoxMesh(comm, dim, PETSC_FALSE, cells, NULL, NULL, NULL, PETSC_TRUE, &dm)); 262 { 263 DMLabel label; 264 IS is; 265 PetscCall(DMCreateLabel(dm, "boundary")); 266 PetscCall(DMGetLabel(dm, "boundary", &label)); 267 PetscCall(DMPlexMarkBoundaryFaces(dm, 1, label)); 268 if (run_type == 0) { 269 PetscCall(DMGetStratumIS(dm, "boundary", 1, &is)); 270 PetscCall(DMCreateLabel(dm,"Faces")); 271 if (is) { 272 PetscInt d, f, Nf; 273 const PetscInt *faces; 274 PetscInt csize; 275 PetscSection cs; 276 Vec coordinates ; 277 DM cdm; 278 PetscCall(ISGetLocalSize(is, &Nf)); 279 PetscCall(ISGetIndices(is, &faces)); 280 PetscCall(DMGetCoordinatesLocal(dm, &coordinates)); 281 PetscCall(DMGetCoordinateDM(dm, &cdm)); 282 PetscCall(DMGetLocalSection(cdm, &cs)); 283 /* Check for each boundary face if any component of its centroid is either 0.0 or 1.0 */ 284 for (f = 0; f < Nf; ++f) { 285 PetscReal faceCoord; 286 PetscInt b,v; 287 PetscScalar *coords = NULL; 288 PetscInt Nv; 289 PetscCall(DMPlexVecGetClosure(cdm, cs, coordinates, faces[f], &csize, &coords)); 290 Nv = csize/dim; /* Calculate mean coordinate vector */ 291 for (d = 0; d < dim; ++d) { 292 faceCoord = 0.0; 293 for (v = 0; v < Nv; ++v) faceCoord += PetscRealPart(coords[v*dim+d]); 294 faceCoord /= Nv; 295 for (b = 0; b < 2; ++b) { 296 if (PetscAbs(faceCoord - b) < PETSC_SMALL) { /* domain have not been set yet, still [0,1]^3 */ 297 PetscCall(DMSetLabelValue(dm, "Faces", faces[f], d*2+b+1)); 298 } 299 } 300 } 301 PetscCall(DMPlexVecRestoreClosure(cdm, cs, coordinates, faces[f], &csize, &coords)); 302 } 303 PetscCall(ISRestoreIndices(is, &faces)); 304 } 305 PetscCall(ISDestroy(&is)); 306 PetscCall(DMGetLabel(dm, "Faces", &label)); 307 PetscCall(DMPlexLabelComplete(dm, label)); 308 } 309 } 310 { 311 PetscInt dimEmbed, i; 312 PetscInt nCoords; 313 PetscScalar *coords,bounds[] = {0,1,-.5,.5,-.5,.5,}; /* x_min,x_max,y_min,y_max */ 314 Vec coordinates; 315 bounds[1] = Lx; 316 if (run_type==1) { 317 for (i = 0; i < 2*dim; i++) bounds[i] = (i%2) ? 1 : 0; 318 } 319 PetscCall(DMGetCoordinatesLocal(dm,&coordinates)); 320 PetscCall(DMGetCoordinateDim(dm,&dimEmbed)); 321 PetscCheck(dimEmbed == dim,PETSC_COMM_SELF,PETSC_ERR_ARG_SIZ,"dimEmbed != dim %" PetscInt_FMT,dimEmbed); 322 PetscCall(VecGetLocalSize(coordinates,&nCoords)); 323 PetscCheck((nCoords % dimEmbed) == 0,PETSC_COMM_SELF,PETSC_ERR_ARG_SIZ,"Coordinate vector the wrong size"); 324 PetscCall(VecGetArray(coordinates,&coords)); 325 for (i = 0; i < nCoords; i += dimEmbed) { 326 PetscInt j; 327 PetscScalar *coord = &coords[i]; 328 for (j = 0; j < dimEmbed; j++) { 329 coord[j] = bounds[2 * j] + coord[j] * (bounds[2 * j + 1] - bounds[2 * j]); 330 } 331 } 332 PetscCall(VecRestoreArray(coordinates,&coords)); 333 PetscCall(DMSetCoordinatesLocal(dm,coordinates)); 334 } 335 336 /* convert to p4est, and distribute */ 337 PetscOptionsBegin(comm, "", "Mesh conversion options", "DMPLEX"); 338 PetscCall(PetscOptionsFList("-dm_type","Convert DMPlex to another format (should not be Plex!)","ex56.c",DMList,DMPLEX,convType,256,&flg)); 339 PetscOptionsEnd(); 340 if (flg) { 341 DM newdm; 342 PetscCall(DMConvert(dm,convType,&newdm)); 343 if (newdm) { 344 const char *prefix; 345 PetscBool isForest; 346 PetscCall(PetscObjectGetOptionsPrefix((PetscObject)dm,&prefix)); 347 PetscCall(PetscObjectSetOptionsPrefix((PetscObject)newdm,prefix)); 348 PetscCall(DMIsForest(newdm,&isForest)); 349 PetscCheck(isForest,PETSC_COMM_WORLD, PETSC_ERR_USER, "Converted to non Forest?"); 350 PetscCall(DMDestroy(&dm)); 351 dm = newdm; 352 } else SETERRQ(PETSC_COMM_WORLD, PETSC_ERR_USER, "Convert failed?"); 353 } else { 354 PetscPartitioner part; 355 /* Plex Distribute mesh over processes */ 356 PetscCall(DMPlexGetPartitioner(dm,&part)); 357 PetscCall(PetscPartitionerSetFromOptions(part)); 358 PetscCall(DMPlexDistribute(dm, 0, NULL, &distdm)); 359 if (distdm) { 360 const char *prefix; 361 PetscCall(PetscObjectGetOptionsPrefix((PetscObject)dm,&prefix)); 362 PetscCall(PetscObjectSetOptionsPrefix((PetscObject)distdm,prefix)); 363 PetscCall(DMDestroy(&dm)); 364 dm = distdm; 365 } 366 } 367 PetscCall(PetscLogStagePop()); 368 basedm = dm; dm = NULL; 369 370 for (iter=0 ; iter<max_conv_its ; iter++) { 371 PetscCall(PetscLogStagePush(stage[16])); 372 /* make new DM */ 373 PetscCall(DMClone(basedm, &dm)); 374 PetscCall(PetscObjectSetOptionsPrefix((PetscObject) dm, "ex56_")); 375 PetscCall(PetscObjectSetName( (PetscObject)dm,"Mesh")); 376 if (max_conv_its > 1) { 377 /* If max_conv_its == 1, then we are not doing a convergence study. */ 378 PetscCall(PetscOptionsInsertString(NULL,options[iter])); 379 } 380 PetscCall(DMSetFromOptions(dm)); /* refinement done here in Plex, p4est */ 381 /* snes */ 382 PetscCall(SNESCreate(comm, &snes)); 383 PetscCall(SNESSetDM(snes, dm)); 384 /* fem */ 385 { 386 const PetscInt Ncomp = dim; 387 const PetscInt components[] = {0,1,2}; 388 const PetscInt Nfid = 1, Npid = 1; 389 const PetscInt fid[] = {1}; /* The fixed faces (x=0) */ 390 const PetscInt pid[] = {2}; /* The faces with loading (x=L_x) */ 391 PetscFE fe; 392 PetscDS prob; 393 DMLabel label; 394 DM cdm = dm; 395 396 PetscCall(PetscFECreateDefault(PetscObjectComm((PetscObject) dm), dim, dim, PETSC_FALSE, NULL, PETSC_DECIDE, &fe)); /* elasticity */ 397 PetscCall(PetscObjectSetName((PetscObject) fe, "deformation")); 398 /* FEM prob */ 399 PetscCall(DMSetField(dm, 0, NULL, (PetscObject) fe)); 400 PetscCall(DMCreateDS(dm)); 401 PetscCall(DMGetDS(dm, &prob)); 402 /* setup problem */ 403 if (run_type==1) { 404 PetscCall(PetscDSSetJacobian(prob, 0, 0, NULL, NULL, NULL, g3_uu_3d)); 405 PetscCall(PetscDSSetResidual(prob, 0, f0_u_x4, f1_u_3d)); 406 } else { 407 PetscWeakForm wf; 408 PetscInt bd, i; 409 410 PetscCall(PetscDSSetJacobian(prob, 0, 0, NULL, NULL, NULL, g3_uu_3d_alpha)); 411 PetscCall(PetscDSSetResidual(prob, 0, f0_u, f1_u_3d_alpha)); 412 413 PetscCall(DMGetLabel(dm, "Faces", &label)); 414 PetscCall(DMAddBoundary(dm, DM_BC_NATURAL, "traction", label, Npid, pid, 0, Ncomp, components, NULL, NULL, NULL, &bd)); 415 PetscCall(PetscDSGetBoundary(prob, bd, &wf, NULL, NULL, NULL, NULL, NULL, NULL, NULL, NULL, NULL, NULL, NULL)); 416 for (i = 0; i < Npid; ++i) PetscCall(PetscWeakFormSetIndexBdResidual(wf, label, pid[i], 0, 0, 0, f0_bd_u_3d, 0, f1_bd_u)); 417 } 418 /* bcs */ 419 if (run_type==1) { 420 PetscInt id = 1; 421 PetscCall(DMGetLabel(dm, "boundary", &label)); 422 PetscCall(DMAddBoundary(dm, DM_BC_ESSENTIAL, "wall", label, 1, &id, 0, 0, NULL, (void (*)(void)) zero, NULL, NULL, NULL)); 423 } else { 424 PetscCall(DMGetLabel(dm, "Faces", &label)); 425 PetscCall(DMAddBoundary(dm, DM_BC_ESSENTIAL, "fixed", label, Nfid, fid, 0, Ncomp, components, (void (*)(void)) zero, NULL, NULL, NULL)); 426 } 427 while (cdm) { 428 PetscCall(DMCopyDisc(dm, cdm)); 429 PetscCall(DMGetCoarseDM(cdm, &cdm)); 430 } 431 PetscCall(PetscFEDestroy(&fe)); 432 } 433 /* vecs & mat */ 434 PetscCall(DMCreateGlobalVector(dm,&xx)); 435 PetscCall(VecDuplicate(xx, &bb)); 436 PetscCall(PetscObjectSetName((PetscObject) bb, "b")); 437 PetscCall(PetscObjectSetName((PetscObject) xx, "u")); 438 PetscCall(DMCreateMatrix(dm, &Amat)); 439 PetscCall(MatSetOption(Amat,MAT_SYMMETRIC,PETSC_TRUE)); /* Some matrix kernels can take advantage of symmetry if we set this. */ 440 PetscCall(MatSetOption(Amat,MAT_SYMMETRY_ETERNAL,PETSC_TRUE)); /* Inform PETSc that Amat is always symmetric, so info set above isn't lost. */ 441 PetscCall(MatSetBlockSize(Amat,3)); 442 PetscCall(MatSetOption(Amat,MAT_SPD,PETSC_TRUE)); 443 PetscCall(MatSetOption(Amat,MAT_SPD_ETERNAL,PETSC_TRUE)); 444 PetscCall(VecGetSize(bb,&N)); 445 local_sizes[iter] = N; 446 PetscCall(PetscInfo(snes,"%" PetscInt_FMT " global equations, %" PetscInt_FMT " vertices\n",N,N/dim)); 447 if ((use_nearnullspace || attach_nearnullspace) && N/dim > 1) { 448 /* Set up the near null space (a.k.a. rigid body modes) that will be used by the multigrid preconditioner */ 449 DM subdm; 450 MatNullSpace nearNullSpace; 451 PetscInt fields = 0; 452 PetscObject deformation; 453 PetscCall(DMCreateSubDM(dm, 1, &fields, NULL, &subdm)); 454 PetscCall(DMPlexCreateRigidBody(subdm, 0, &nearNullSpace)); 455 PetscCall(DMGetField(dm, 0, NULL, &deformation)); 456 PetscCall(PetscObjectCompose(deformation, "nearnullspace", (PetscObject) nearNullSpace)); 457 PetscCall(DMDestroy(&subdm)); 458 if (attach_nearnullspace) PetscCall(MatSetNearNullSpace(Amat,nearNullSpace)); 459 PetscCall(MatNullSpaceDestroy(&nearNullSpace)); /* created by DM and destroyed by Mat */ 460 } 461 PetscCall(DMPlexSetSNESLocalFEM(dm,NULL,NULL,NULL)); 462 PetscCall(SNESSetJacobian(snes, Amat, Amat, NULL, NULL)); 463 PetscCall(SNESSetFromOptions(snes)); 464 PetscCall(DMSetUp(dm)); 465 PetscCall(PetscLogStagePop()); 466 PetscCall(PetscLogStagePush(stage[16])); 467 /* ksp */ 468 PetscCall(SNESGetKSP(snes, &ksp)); 469 PetscCall(KSPSetComputeSingularValues(ksp,PETSC_TRUE)); 470 /* test BCs */ 471 PetscCall(VecZeroEntries(xx)); 472 if (test_nonzero_cols) { 473 if (rank == 0) { 474 PetscCall(VecSetValue(xx,0,1.0,INSERT_VALUES)); 475 } 476 PetscCall(VecAssemblyBegin(xx)); 477 PetscCall(VecAssemblyEnd(xx)); 478 } 479 PetscCall(VecZeroEntries(bb)); 480 PetscCall(VecGetSize(bb,&i)); 481 local_sizes[iter] = i; 482 PetscCall(PetscInfo(snes,"%" PetscInt_FMT " equations in vector, %" PetscInt_FMT " vertices\n",i,i/dim)); 483 PetscCall(PetscLogStagePop()); 484 /* solve */ 485 PetscCall(PetscLogStagePush(stage[iter])); 486 PetscCall(SNESSolve(snes, bb, xx)); 487 PetscCall(PetscLogStagePop()); 488 PetscCall(VecNorm(xx,NORM_INFINITY,&mdisp[iter])); 489 PetscCall(DMViewFromOptions(dm, NULL, "-dm_view")); 490 { 491 PetscViewer viewer = NULL; 492 PetscViewerFormat fmt; 493 PetscCall(PetscOptionsGetViewer(comm,NULL,"ex56_","-vec_view",&viewer,&fmt,&flg)); 494 if (flg) { 495 PetscCall(PetscViewerPushFormat(viewer,fmt)); 496 PetscCall(VecView(xx,viewer)); 497 PetscCall(VecView(bb,viewer)); 498 PetscCall(PetscViewerPopFormat(viewer)); 499 } 500 PetscCall(PetscViewerDestroy(&viewer)); 501 } 502 /* Free work space */ 503 PetscCall(DMDestroy(&dm)); 504 PetscCall(SNESDestroy(&snes)); 505 PetscCall(VecDestroy(&xx)); 506 PetscCall(VecDestroy(&bb)); 507 PetscCall(MatDestroy(&Amat)); 508 } 509 PetscCall(DMDestroy(&basedm)); 510 if (run_type==1) err[0] = 59.975208 - mdisp[0]; /* error with what I think is the exact solution */ 511 else err[0] = 171.038 - mdisp[0]; 512 for (iter=1 ; iter<max_conv_its ; iter++) { 513 if (run_type==1) err[iter] = 59.975208 - mdisp[iter]; 514 else err[iter] = 171.038 - mdisp[iter]; 515 PetscCall(PetscPrintf(PETSC_COMM_WORLD,"[%d] %" PetscInt_FMT ") N=%12" PetscInt_FMT ", max displ=%9.7e, disp diff=%9.2e, error=%4.3e, rate=%3.2g\n",rank,iter,local_sizes[iter],(double)mdisp[iter], 516 (double)(mdisp[iter]-mdisp[iter-1]),(double)err[iter],(double)(PetscLogReal(err[iter-1]/err[iter])/PetscLogReal(2.)))); 517 } 518 519 PetscCall(PetscFinalize()); 520 return 0; 521 } 522 523 /*TEST 524 525 test: 526 suffix: 0 527 nsize: 4 528 requires: !single 529 args: -cells 2,2,1 -max_conv_its 2 -petscspace_degree 3 -snes_max_it 1 -ksp_max_it 100 -ksp_type cg -ksp_rtol 1.e-10 -ksp_norm_type unpreconditioned -pc_type gamg -pc_gamg_coarse_eq_limit 10 -pc_gamg_reuse_interpolation true -pc_gamg_aggressive_coarsening 0 -pc_gamg_threshold 0.001 -ksp_converged_reason -snes_converged_reason -use_mat_nearnullspace true -mg_levels_ksp_max_it 2 -mg_levels_ksp_type chebyshev -mg_levels_ksp_chebyshev_esteig 0,0.2,0,1.1 -mg_levels_pc_type jacobi -petscpartitioner_type simple -ex56_dm_view -snes_lag_jacobian -2 -snes_type ksponly -use_gpu_aware_mpi true 530 timeoutfactor: 2 531 532 # HYPRE PtAP broken with complex numbers 533 test: 534 suffix: hypre 535 requires: hypre !single !complex !defined(PETSC_HAVE_HYPRE_DEVICE) 536 nsize: 4 537 args: -cells 2,2,1 -max_conv_its 2 -lx 1. -alpha .01 -petscspace_degree 2 -ksp_type cg -ksp_monitor_short -ksp_rtol 1.e-8 -pc_type hypre -pc_hypre_type boomeramg -pc_hypre_boomeramg_no_CF true -pc_hypre_boomeramg_agg_nl 1 -pc_hypre_boomeramg_coarsen_type HMIS -pc_hypre_boomeramg_interp_type ext+i -ksp_converged_reason -use_mat_nearnullspace true -petscpartitioner_type simple 538 539 test: 540 suffix: ml 541 requires: ml !single 542 nsize: 4 543 args: -cells 2,2,1 -max_conv_its 2 -lx 1. -alpha .01 -petscspace_degree 2 -ksp_type cg -ksp_monitor_short -ksp_converged_reason -ksp_rtol 1.e-8 -pc_type ml -mg_levels_ksp_type chebyshev -mg_levels_ksp_max_it 3 -mg_levels_ksp_chebyshev_esteig 0,0.05,0,1.05 -mg_levels_pc_type sor -petscpartitioner_type simple -use_mat_nearnullspace 544 545 test: 546 suffix: hpddm 547 requires: hpddm slepc !single defined(PETSC_HAVE_DYNAMIC_LIBRARIES) defined(PETSC_USE_SHARED_LIBRARIES) 548 nsize: 4 549 args: -cells 2,2,1 -max_conv_its 2 -lx 1. -alpha .01 -petscspace_degree 2 -ksp_type fgmres -ksp_monitor_short -ksp_converged_reason -ksp_rtol 1.e-8 -pc_type hpddm -petscpartitioner_type simple -pc_hpddm_levels_1_sub_pc_type lu -pc_hpddm_levels_1_eps_nev 6 -pc_hpddm_coarse_p 1 -pc_hpddm_coarse_pc_type svd 550 551 test: 552 suffix: repart 553 nsize: 4 554 requires: parmetis !single 555 args: -cells 8,2,2 -max_conv_its 1 -petscspace_degree 2 -snes_max_it 4 -ksp_max_it 100 -ksp_type cg -ksp_rtol 1.e-2 -ksp_norm_type unpreconditioned -snes_rtol 1.e-3 -pc_type gamg -pc_gamg_esteig_ksp_max_it 10 -pc_gamg_type agg -pc_gamg_agg_nsmooths 1 -pc_gamg_aggressive_coarsening 1 -pc_gamg_threshold 0.05 -pc_gamg_threshold_scale .0 -use_mat_nearnullspace true -mg_levels_ksp_max_it 2 -mg_levels_ksp_type chebyshev -mg_levels_ksp_chebyshev_esteig 0,0.05,0,1.05 -mg_levels_pc_type jacobi -pc_gamg_mat_partitioning_type parmetis -pc_gamg_repartition true -snes_converged_reason -pc_gamg_process_eq_limit 20 -pc_gamg_coarse_eq_limit 10 -ksp_converged_reason -snes_converged_reason -pc_gamg_reuse_interpolation true 556 557 test: 558 suffix: bddc 559 nsize: 4 560 requires: !single 561 args: -cells 2,2,1 -max_conv_its 2 -lx 1. -alpha .01 -petscspace_degree 2 -ksp_type cg -ksp_monitor_short -ksp_rtol 1.e-8 -ksp_converged_reason -petscpartitioner_type simple -ex56_dm_mat_type is -matis_localmat_type {{sbaij baij aij}} -pc_type bddc 562 563 testset: 564 nsize: 4 565 requires: !single 566 args: -cells 2,2,1 -max_conv_its 2 -lx 1. -alpha .01 -petscspace_degree 2 -ksp_type cg -ksp_monitor_short -ksp_rtol 1.e-10 -ksp_converged_reason -petscpartitioner_type simple -ex56_dm_mat_type is -matis_localmat_type aij -pc_type bddc -attach_mat_nearnullspace {{0 1}separate output} 567 test: 568 suffix: bddc_approx_gamg 569 args: -pc_bddc_switch_static -prefix_push pc_bddc_dirichlet_ -approximate -pc_type gamg -pc_gamg_esteig_ksp_max_it 10 -pc_gamg_type agg -pc_gamg_agg_nsmooths 1 -pc_gamg_reuse_interpolation true -pc_gamg_aggressive_coarsening 1 -pc_gamg_threshold 0.05 -pc_gamg_threshold_scale .0 -mg_levels_ksp_max_it 1 -mg_levels_ksp_type chebyshev -prefix_pop -prefix_push pc_bddc_neumann_ -approximate -pc_type gamg -pc_gamg_esteig_ksp_max_it 10 -pc_gamg_type agg -pc_gamg_agg_nsmooths 1 -pc_gamg_coarse_eq_limit 10 -pc_gamg_reuse_interpolation true -pc_gamg_aggressive_coarsening 1 -pc_gamg_threshold 0.05 -pc_gamg_threshold_scale .0 -mg_levels_ksp_max_it 1 -mg_levels_ksp_type chebyshev -prefix_pop 570 # HYPRE PtAP broken with complex numbers 571 test: 572 requires: hypre !complex !defined(PETSC_HAVE_HYPRE_DEVICE) 573 suffix: bddc_approx_hypre 574 args: -pc_bddc_switch_static -prefix_push pc_bddc_dirichlet_ -pc_type hypre -pc_hypre_boomeramg_no_CF true -pc_hypre_boomeramg_strong_threshold 0.75 -pc_hypre_boomeramg_agg_nl 1 -pc_hypre_boomeramg_coarsen_type HMIS -pc_hypre_boomeramg_interp_type ext+i -prefix_pop -prefix_push pc_bddc_neumann_ -pc_type hypre -pc_hypre_boomeramg_no_CF true -pc_hypre_boomeramg_strong_threshold 0.75 -pc_hypre_boomeramg_agg_nl 1 -pc_hypre_boomeramg_coarsen_type HMIS -pc_hypre_boomeramg_interp_type ext+i -prefix_pop 575 test: 576 requires: ml 577 suffix: bddc_approx_ml 578 args: -pc_bddc_switch_static -prefix_push pc_bddc_dirichlet_ -approximate -pc_type ml -mg_levels_ksp_max_it 1 -mg_levels_ksp_type chebyshev -prefix_pop -prefix_push pc_bddc_neumann_ -approximate -pc_type ml -mg_levels_ksp_max_it 1 -mg_levels_ksp_type chebyshev -prefix_pop 579 580 test: 581 suffix: fetidp 582 nsize: 4 583 requires: !single 584 args: -cells 2,2,1 -max_conv_its 2 -lx 1. -alpha .01 -petscspace_degree 2 -ksp_type fetidp -fetidp_ksp_type cg -ksp_monitor_short -ksp_rtol 1.e-8 -ksp_converged_reason -petscpartitioner_type simple -ex56_dm_mat_type is -matis_localmat_type {{sbaij baij aij}} 585 586 test: 587 suffix: bddc_elast 588 nsize: 4 589 requires: !single 590 args: -cells 2,2,1 -max_conv_its 2 -lx 1. -alpha .01 -petscspace_degree 2 -ksp_type cg -ksp_monitor_short -ksp_rtol 1.e-8 -ksp_converged_reason -petscpartitioner_type simple -ex56_dm_mat_type is -matis_localmat_type sbaij -pc_type bddc -pc_bddc_monolithic -attach_mat_nearnullspace 591 592 test: 593 suffix: fetidp_elast 594 nsize: 4 595 requires: !single 596 args: -cells 2,2,1 -max_conv_its 2 -lx 1. -alpha .01 -petscspace_degree 2 -ksp_type fetidp -fetidp_ksp_type cg -ksp_monitor_short -ksp_rtol 1.e-8 -ksp_converged_reason -petscpartitioner_type simple -ex56_dm_mat_type is -matis_localmat_type sbaij -fetidp_bddc_pc_bddc_monolithic -attach_mat_nearnullspace 597 598 test: 599 suffix: gdsw 600 nsize: 4 601 requires: !single 602 args: -cells 2,2,1 -max_conv_its 2 -lx 1. -alpha .01 -petscspace_degree 2 -ksp_type cg -ksp_monitor_short -ksp_rtol 1.e-8 -ksp_converged_reason -petscpartitioner_type simple -ex56_dm_mat_type is -attach_mat_nearnullspace \ 603 -pc_type mg -pc_mg_galerkin -pc_mg_adapt_interp_coarse_space gdsw -pc_mg_levels 2 -mg_levels_pc_type bjacobi -mg_levels_sub_pc_type icc 604 605 testset: 606 nsize: 4 607 requires: !single 608 args: -cells 2,2,1 -max_conv_its 2 -petscspace_degree 2 -snes_max_it 2 -ksp_max_it 100 -ksp_type cg -ksp_rtol 1.e-10 -ksp_norm_type unpreconditioned -snes_rtol 1.e-10 -pc_type gamg -pc_gamg_esteig_ksp_max_it 10 -pc_gamg_type agg -pc_gamg_agg_nsmooths 1 -pc_gamg_coarse_eq_limit 10 -pc_gamg_reuse_interpolation true -pc_gamg_aggressive_coarsening 1 -pc_gamg_threshold 0.05 -pc_gamg_threshold_scale .0 -use_mat_nearnullspace true -mg_levels_ksp_max_it 2 -mg_levels_ksp_type chebyshev -mg_levels_ksp_chebyshev_esteig 0,0.05,0,1.05 -mg_levels_pc_type jacobi -ksp_monitor_short -ksp_converged_reason -snes_converged_reason -snes_monitor_short -ex56_dm_view -petscpartitioner_type simple -pc_gamg_process_eq_limit 20 609 output_file: output/ex56_cuda.out 610 611 test: 612 suffix: cuda 613 requires: cuda 614 args: -ex56_dm_mat_type aijcusparse -ex56_dm_vec_type cuda 615 616 test: 617 suffix: viennacl 618 requires: viennacl 619 args: -ex56_dm_mat_type aijviennacl -ex56_dm_vec_type viennacl 620 621 test: 622 suffix: kokkos 623 requires: !sycl kokkos_kernels 624 args: -ex56_dm_mat_type aijkokkos -ex56_dm_vec_type kokkos 625 # Don't run AIJMKL caes with complex scalars because of convergence issues. 626 # Note that we need to test both single and multiple MPI rank cases, because these use different sparse MKL routines to implement the PtAP operation. 627 test: 628 suffix: seqaijmkl 629 nsize: 1 630 requires: defined(PETSC_HAVE_MKL_SPARSE_OPTIMIZE) !single !complex 631 args: -cells 2,2,1 -max_conv_its 2 -petscspace_degree 2 -snes_max_it 2 -ksp_max_it 100 -ksp_type cg -ksp_rtol 1.e-11 -ksp_norm_type unpreconditioned -snes_rtol 1.e-10 -pc_type gamg -pc_gamg_type agg -pc_gamg_agg_nsmooths 1 -pc_gamg_coarse_eq_limit 1000 -pc_gamg_reuse_interpolation true -pc_gamg_aggressive_coarsening 1 -pc_gamg_threshold 0.05 -pc_gamg_threshold_scale .0 -ksp_converged_reason -snes_monitor_short -ksp_monitor_short -snes_converged_reason -use_mat_nearnullspace true -mg_levels_ksp_max_it 1 -mg_levels_ksp_type chebyshev -pc_gamg_esteig_ksp_type cg -pc_gamg_esteig_ksp_max_it 10 -mg_levels_ksp_chebyshev_esteig 0,0.05,0,1.1 -mg_levels_pc_type jacobi -petscpartitioner_type simple -mat_block_size 3 -ex56_dm_view -run_type 1 -mat_seqaij_type seqaijmkl 632 timeoutfactor: 2 633 634 test: 635 suffix: mpiaijmkl 636 nsize: 2 637 requires: defined(PETSC_HAVE_MKL_SPARSE_OPTIMIZE) !single !complex 638 args: -cells 2,2,1 -max_conv_its 2 -petscspace_degree 2 -snes_max_it 2 -ksp_max_it 100 -ksp_type cg -ksp_rtol 1.e-11 -ksp_norm_type unpreconditioned -snes_rtol 1.e-10 -pc_type gamg -pc_gamg_type agg -pc_gamg_agg_nsmooths 1 -pc_gamg_coarse_eq_limit 1000 -pc_gamg_reuse_interpolation true -pc_gamg_aggressive_coarsening 1 -pc_gamg_threshold 0.05 -pc_gamg_threshold_scale .0 -ksp_converged_reason -snes_monitor_short -ksp_monitor_short -snes_converged_reason -use_mat_nearnullspace true -mg_levels_ksp_max_it 1 -mg_levels_ksp_type chebyshev -pc_gamg_esteig_ksp_type cg -pc_gamg_esteig_ksp_max_it 10 -mg_levels_ksp_chebyshev_esteig 0,0.05,0,1.1 -mg_levels_pc_type jacobi -petscpartitioner_type simple -mat_block_size 3 -ex56_dm_view -run_type 1 -mat_seqaij_type seqaijmkl 639 timeoutfactor: 2 640 641 TEST*/ 642