#include /*I "petscdmda.h" I*/ #include /*I "petscksp.h" I*/ #include typedef struct { PCExoticType type; Mat P; /* the constructed interpolation matrix */ PetscBool directSolve; /* use direct LU factorization to construct interpolation */ KSP ksp; } PC_Exotic; const char *const PCExoticTypes[] = {"face", "wirebasket", "PCExoticType", "PC_Exotic", NULL}; /* DMDAGetWireBasketInterpolation - Gets the interpolation for a wirebasket based coarse space */ PetscErrorCode DMDAGetWireBasketInterpolation(PC pc, DM da, PC_Exotic *exotic, Mat Aglobal, MatReuse reuse, Mat *P) { PetscInt dim, i, j, k, m, n, p, dof, Nint, Nface, Nwire, Nsurf, *Iint, *Isurf, cint = 0, csurf = 0, istart, jstart, kstart, *II, N, c = 0; PetscInt mwidth, nwidth, pwidth, cnt, mp, np, pp, Ntotal, gl[26], *globals, Ng, *IIint, *IIsurf; Mat Xint, Xsurf, Xint_tmp; IS isint, issurf, is, row, col; ISLocalToGlobalMapping ltg; MPI_Comm comm; Mat A, Aii, Ais, Asi, *Aholder, iAii; MatFactorInfo info; PetscScalar *xsurf, *xint; const PetscScalar *rxint; #if defined(PETSC_USE_DEBUG_foo) PetscScalar tmp; #endif PetscHMapI ht = NULL; PetscFunctionBegin; PetscCall(DMDAGetInfo(da, &dim, NULL, NULL, NULL, &mp, &np, &pp, &dof, NULL, NULL, NULL, NULL, NULL)); PetscCheck(dof == 1, PetscObjectComm((PetscObject)da), PETSC_ERR_SUP, "Only for single field problems"); PetscCheck(dim == 3, PetscObjectComm((PetscObject)da), PETSC_ERR_SUP, "Only coded for 3d problems"); PetscCall(DMDAGetCorners(da, NULL, NULL, NULL, &m, &n, &p)); PetscCall(DMDAGetGhostCorners(da, &istart, &jstart, &kstart, &mwidth, &nwidth, &pwidth)); istart = istart ? -1 : 0; jstart = jstart ? -1 : 0; kstart = kstart ? -1 : 0; /* the columns of P are the interpolation of each coarse grid point (one for each vertex and edge) to all the local degrees of freedom (this includes the vertices, edges and faces). Xint are the subset of the interpolation into the interior Xface are the interpolation onto faces but not into the interior Xsurf are the interpolation onto the vertices and edges (the surfbasket) Xint Symbolically one could write P = (Xface) after interchanging the rows to match the natural ordering on the domain Xsurf */ N = (m - istart) * (n - jstart) * (p - kstart); Nint = (m - 2 - istart) * (n - 2 - jstart) * (p - 2 - kstart); Nface = 2 * ((m - 2 - istart) * (n - 2 - jstart) + (m - 2 - istart) * (p - 2 - kstart) + (n - 2 - jstart) * (p - 2 - kstart)); Nwire = 4 * ((m - 2 - istart) + (n - 2 - jstart) + (p - 2 - kstart)) + 8; Nsurf = Nface + Nwire; PetscCall(MatCreateSeqDense(MPI_COMM_SELF, Nint, 26, NULL, &Xint)); PetscCall(MatCreateSeqDense(MPI_COMM_SELF, Nsurf, 26, NULL, &Xsurf)); PetscCall(MatDenseGetArray(Xsurf, &xsurf)); /* Require that all 12 edges and 6 faces have at least one grid point. Otherwise some of the columns of Xsurf will be all zero (thus making the coarse matrix singular). */ PetscCheck(m - istart >= 3, PETSC_COMM_SELF, PETSC_ERR_SUP, "Number of grid points per process in X direction must be at least 3"); PetscCheck(n - jstart >= 3, PETSC_COMM_SELF, PETSC_ERR_SUP, "Number of grid points per process in Y direction must be at least 3"); PetscCheck(p - kstart >= 3, PETSC_COMM_SELF, PETSC_ERR_SUP, "Number of grid points per process in Z direction must be at least 3"); cnt = 0; xsurf[cnt++] = 1; for (i = 1; i < m - istart - 1; i++) xsurf[cnt++ + Nsurf] = 1; xsurf[cnt++ + 2 * Nsurf] = 1; for (j = 1; j < n - 1 - jstart; j++) { xsurf[cnt++ + 3 * Nsurf] = 1; for (i = 1; i < m - istart - 1; i++) xsurf[cnt++ + 4 * Nsurf] = 1; xsurf[cnt++ + 5 * Nsurf] = 1; } xsurf[cnt++ + 6 * Nsurf] = 1; for (i = 1; i < m - istart - 1; i++) xsurf[cnt++ + 7 * Nsurf] = 1; xsurf[cnt++ + 8 * Nsurf] = 1; for (k = 1; k < p - 1 - kstart; k++) { xsurf[cnt++ + 9 * Nsurf] = 1; for (i = 1; i < m - istart - 1; i++) xsurf[cnt++ + 10 * Nsurf] = 1; xsurf[cnt++ + 11 * Nsurf] = 1; for (j = 1; j < n - 1 - jstart; j++) { xsurf[cnt++ + 12 * Nsurf] = 1; /* these are the interior nodes */ xsurf[cnt++ + 13 * Nsurf] = 1; } xsurf[cnt++ + 14 * Nsurf] = 1; for (i = 1; i < m - istart - 1; i++) xsurf[cnt++ + 15 * Nsurf] = 1; xsurf[cnt++ + 16 * Nsurf] = 1; } xsurf[cnt++ + 17 * Nsurf] = 1; for (i = 1; i < m - istart - 1; i++) xsurf[cnt++ + 18 * Nsurf] = 1; xsurf[cnt++ + 19 * Nsurf] = 1; for (j = 1; j < n - 1 - jstart; j++) { xsurf[cnt++ + 20 * Nsurf] = 1; for (i = 1; i < m - istart - 1; i++) xsurf[cnt++ + 21 * Nsurf] = 1; xsurf[cnt++ + 22 * Nsurf] = 1; } xsurf[cnt++ + 23 * Nsurf] = 1; for (i = 1; i < m - istart - 1; i++) xsurf[cnt++ + 24 * Nsurf] = 1; xsurf[cnt++ + 25 * Nsurf] = 1; /* interpolations only sum to 1 when using direct solver */ #if defined(PETSC_USE_DEBUG_foo) for (i = 0; i < Nsurf; i++) { tmp = 0.0; for (j = 0; j < 26; j++) tmp += xsurf[i + j * Nsurf]; PetscCheck(PetscAbsScalar(tmp - 1.0) <= 1.e-10, PETSC_COMM_SELF, PETSC_ERR_PLIB, "Wrong Xsurf interpolation at i %" PetscInt_FMT " value %g", i, (double)PetscAbsScalar(tmp)); } #endif PetscCall(MatDenseRestoreArray(Xsurf, &xsurf)); /* PetscCall(MatView(Xsurf,PETSC_VIEWER_STDOUT_WORLD));*/ /* I are the indices for all the needed vertices (in global numbering) Iint are the indices for the interior values, I surf for the surface values (This is just for the part of the global matrix obtained with MatCreateSubMatrix(), it is NOT the local DMDA ordering.) IIint and IIsurf are the same as the Iint, Isurf except they are in the global numbering */ #define Endpoint(a, start, b) (a == 0 || a == (b - 1 - start)) PetscCall(PetscMalloc3(N, &II, Nint, &Iint, Nsurf, &Isurf)); PetscCall(PetscMalloc2(Nint, &IIint, Nsurf, &IIsurf)); for (k = 0; k < p - kstart; k++) { for (j = 0; j < n - jstart; j++) { for (i = 0; i < m - istart; i++) { II[c++] = i + j * mwidth + k * mwidth * nwidth; if (!Endpoint(i, istart, m) && !Endpoint(j, jstart, n) && !Endpoint(k, kstart, p)) { IIint[cint] = i + j * mwidth + k * mwidth * nwidth; Iint[cint++] = i + j * (m - istart) + k * (m - istart) * (n - jstart); } else { IIsurf[csurf] = i + j * mwidth + k * mwidth * nwidth; Isurf[csurf++] = i + j * (m - istart) + k * (m - istart) * (n - jstart); } } } } #undef Endpoint PetscCheck(c == N, PETSC_COMM_SELF, PETSC_ERR_PLIB, "c != N"); PetscCheck(cint == Nint, PETSC_COMM_SELF, PETSC_ERR_PLIB, "cint != Nint"); PetscCheck(csurf == Nsurf, PETSC_COMM_SELF, PETSC_ERR_PLIB, "csurf != Nsurf"); PetscCall(DMGetLocalToGlobalMapping(da, <g)); PetscCall(ISLocalToGlobalMappingApply(ltg, N, II, II)); PetscCall(ISLocalToGlobalMappingApply(ltg, Nint, IIint, IIint)); PetscCall(ISLocalToGlobalMappingApply(ltg, Nsurf, IIsurf, IIsurf)); PetscCall(PetscObjectGetComm((PetscObject)da, &comm)); PetscCall(ISCreateGeneral(comm, N, II, PETSC_COPY_VALUES, &is)); PetscCall(ISCreateGeneral(PETSC_COMM_SELF, Nint, Iint, PETSC_COPY_VALUES, &isint)); PetscCall(ISCreateGeneral(PETSC_COMM_SELF, Nsurf, Isurf, PETSC_COPY_VALUES, &issurf)); PetscCall(PetscFree3(II, Iint, Isurf)); PetscCall(MatCreateSubMatrices(Aglobal, 1, &is, &is, MAT_INITIAL_MATRIX, &Aholder)); A = *Aholder; PetscCall(PetscFree(Aholder)); PetscCall(MatCreateSubMatrix(A, isint, isint, MAT_INITIAL_MATRIX, &Aii)); PetscCall(MatCreateSubMatrix(A, isint, issurf, MAT_INITIAL_MATRIX, &Ais)); PetscCall(MatCreateSubMatrix(A, issurf, isint, MAT_INITIAL_MATRIX, &Asi)); /* Solve for the interpolation onto the interior Xint */ PetscCall(MatMatMult(Ais, Xsurf, MAT_INITIAL_MATRIX, PETSC_DETERMINE, &Xint_tmp)); PetscCall(MatScale(Xint_tmp, -1.0)); if (exotic->directSolve) { PetscCall(MatGetFactor(Aii, MATSOLVERPETSC, MAT_FACTOR_LU, &iAii)); PetscCall(MatFactorInfoInitialize(&info)); PetscCall(MatGetOrdering(Aii, MATORDERINGND, &row, &col)); PetscCall(MatLUFactorSymbolic(iAii, Aii, row, col, &info)); PetscCall(ISDestroy(&row)); PetscCall(ISDestroy(&col)); PetscCall(MatLUFactorNumeric(iAii, Aii, &info)); PetscCall(MatMatSolve(iAii, Xint_tmp, Xint)); PetscCall(MatDestroy(&iAii)); } else { Vec b, x; PetscScalar *xint_tmp; PetscCall(MatDenseGetArray(Xint, &xint)); PetscCall(VecCreateSeqWithArray(PETSC_COMM_SELF, 1, Nint, NULL, &x)); PetscCall(MatDenseGetArray(Xint_tmp, &xint_tmp)); PetscCall(VecCreateSeqWithArray(PETSC_COMM_SELF, 1, Nint, NULL, &b)); PetscCall(KSPSetOperators(exotic->ksp, Aii, Aii)); for (i = 0; i < 26; i++) { PetscCall(VecPlaceArray(x, xint + i * Nint)); PetscCall(VecPlaceArray(b, xint_tmp + i * Nint)); PetscCall(KSPSolve(exotic->ksp, b, x)); PetscCall(KSPCheckSolve(exotic->ksp, pc, x)); PetscCall(VecResetArray(x)); PetscCall(VecResetArray(b)); } PetscCall(MatDenseRestoreArray(Xint, &xint)); PetscCall(MatDenseRestoreArray(Xint_tmp, &xint_tmp)); PetscCall(VecDestroy(&x)); PetscCall(VecDestroy(&b)); } PetscCall(MatDestroy(&Xint_tmp)); #if defined(PETSC_USE_DEBUG_foo) PetscCall(MatDenseGetArrayRead(Xint, &rxint)); for (i = 0; i < Nint; i++) { tmp = 0.0; for (j = 0; j < 26; j++) tmp += rxint[i + j * Nint]; PetscCheck(PetscAbsScalar(tmp - 1.0) <= 1.e-10, PETSC_COMM_SELF, PETSC_ERR_PLIB, "Wrong Xint interpolation at i %" PetscInt_FMT " value %g", i, (double)PetscAbsScalar(tmp)); } PetscCall(MatDenseRestoreArrayRead(Xint, &rxint)); /* PetscCall(MatView(Xint,PETSC_VIEWER_STDOUT_WORLD)); */ #endif /* total vertices total faces total edges */ Ntotal = (mp + 1) * (np + 1) * (pp + 1) + mp * np * (pp + 1) + mp * pp * (np + 1) + np * pp * (mp + 1) + mp * (np + 1) * (pp + 1) + np * (mp + 1) * (pp + 1) + pp * (mp + 1) * (np + 1); /* For each vertex, edge, face on process (in the same orderings as used above) determine its local number including ghost points */ cnt = 0; gl[cnt++] = 0; { gl[cnt++] = 1; } gl[cnt++] = m - istart - 1; { gl[cnt++] = mwidth; { gl[cnt++] = mwidth + 1; } gl[cnt++] = mwidth + m - istart - 1; } gl[cnt++] = mwidth * (n - jstart - 1); { gl[cnt++] = mwidth * (n - jstart - 1) + 1; } gl[cnt++] = mwidth * (n - jstart - 1) + m - istart - 1; { gl[cnt++] = mwidth * nwidth; { gl[cnt++] = mwidth * nwidth + 1; } gl[cnt++] = mwidth * nwidth + m - istart - 1; { gl[cnt++] = mwidth * nwidth + mwidth; /* these are the interior nodes */ gl[cnt++] = mwidth * nwidth + mwidth + m - istart - 1; } gl[cnt++] = mwidth * nwidth + mwidth * (n - jstart - 1); { gl[cnt++] = mwidth * nwidth + mwidth * (n - jstart - 1) + 1; } gl[cnt++] = mwidth * nwidth + mwidth * (n - jstart - 1) + m - istart - 1; } gl[cnt++] = mwidth * nwidth * (p - kstart - 1); { gl[cnt++] = mwidth * nwidth * (p - kstart - 1) + 1; } gl[cnt++] = mwidth * nwidth * (p - kstart - 1) + m - istart - 1; { gl[cnt++] = mwidth * nwidth * (p - kstart - 1) + mwidth; { gl[cnt++] = mwidth * nwidth * (p - kstart - 1) + mwidth + 1; } gl[cnt++] = mwidth * nwidth * (p - kstart - 1) + mwidth + m - istart - 1; } gl[cnt++] = mwidth * nwidth * (p - kstart - 1) + mwidth * (n - jstart - 1); { gl[cnt++] = mwidth * nwidth * (p - kstart - 1) + mwidth * (n - jstart - 1) + 1; } gl[cnt++] = mwidth * nwidth * (p - kstart - 1) + mwidth * (n - jstart - 1) + m - istart - 1; /* PetscIntView(26,gl,PETSC_VIEWER_STDOUT_WORLD); */ /* convert that to global numbering and get them on all processes */ PetscCall(ISLocalToGlobalMappingApply(ltg, 26, gl, gl)); /* PetscIntView(26,gl,PETSC_VIEWER_STDOUT_WORLD); */ PetscCall(PetscMalloc1(26 * mp * np * pp, &globals)); PetscCallMPI(MPI_Allgather(gl, 26, MPIU_INT, globals, 26, MPIU_INT, PetscObjectComm((PetscObject)da))); /* Number the coarse grid points from 0 to Ntotal */ PetscCall(PetscHMapICreateWithSize(Ntotal / 3, &ht)); for (i = 0, cnt = 0; i < 26 * mp * np * pp; i++) { PetscHashIter it = 0; PetscBool missing = PETSC_TRUE; PetscCall(PetscHMapIPut(ht, globals[i] + 1, &it, &missing)); if (missing) { ++cnt; PetscCall(PetscHMapIIterSet(ht, it, cnt)); } } PetscCheck(cnt == Ntotal, PETSC_COMM_SELF, PETSC_ERR_PLIB, "Hash table size %" PetscInt_FMT " not equal to total number coarse grid points %" PetscInt_FMT, cnt, Ntotal); PetscCall(PetscFree(globals)); for (i = 0; i < 26; i++) { PetscCall(PetscHMapIGetWithDefault(ht, gl[i] + 1, 0, gl + i)); --(gl[i]); } PetscCall(PetscHMapIDestroy(&ht)); /* PetscIntView(26,gl,PETSC_VIEWER_STDOUT_WORLD); */ /* construct global interpolation matrix */ PetscCall(MatGetLocalSize(Aglobal, &Ng, NULL)); if (reuse == MAT_INITIAL_MATRIX) { PetscCall(MatCreateAIJ(PetscObjectComm((PetscObject)da), Ng, PETSC_DECIDE, PETSC_DECIDE, Ntotal, Nint + Nsurf, NULL, Nint + Nsurf, NULL, P)); } else { PetscCall(MatZeroEntries(*P)); } PetscCall(MatSetOption(*P, MAT_ROW_ORIENTED, PETSC_FALSE)); PetscCall(MatDenseGetArrayRead(Xint, &rxint)); PetscCall(MatSetValues(*P, Nint, IIint, 26, gl, rxint, INSERT_VALUES)); PetscCall(MatDenseRestoreArrayRead(Xint, &rxint)); PetscCall(MatDenseGetArrayRead(Xsurf, &rxint)); PetscCall(MatSetValues(*P, Nsurf, IIsurf, 26, gl, rxint, INSERT_VALUES)); PetscCall(MatDenseRestoreArrayRead(Xsurf, &rxint)); PetscCall(MatAssemblyBegin(*P, MAT_FINAL_ASSEMBLY)); PetscCall(MatAssemblyEnd(*P, MAT_FINAL_ASSEMBLY)); PetscCall(PetscFree2(IIint, IIsurf)); #if defined(PETSC_USE_DEBUG_foo) { Vec x, y; PetscScalar *yy; PetscCall(VecCreateMPI(PetscObjectComm((PetscObject)da), Ng, PETSC_DETERMINE, &y)); PetscCall(VecCreateMPI(PetscObjectComm((PetscObject)da), PETSC_DETERMINE, Ntotal, &x)); PetscCall(VecSet(x, 1.0)); PetscCall(MatMult(*P, x, y)); PetscCall(VecGetArray(y, &yy)); for (i = 0; i < Ng; i++) PetscCheck(PetscAbsScalar(yy[i] - 1.0) <= 1.e-10, PETSC_COMM_SELF, PETSC_ERR_PLIB, "Wrong p interpolation at i %" PetscInt_FMT " value %g", i, (double)PetscAbsScalar(yy[i])); PetscCall(VecRestoreArray(y, &yy)); PetscCall(VecDestroy(x)); PetscCall(VecDestroy(y)); } #endif PetscCall(MatDestroy(&Aii)); PetscCall(MatDestroy(&Ais)); PetscCall(MatDestroy(&Asi)); PetscCall(MatDestroy(&A)); PetscCall(ISDestroy(&is)); PetscCall(ISDestroy(&isint)); PetscCall(ISDestroy(&issurf)); PetscCall(MatDestroy(&Xint)); PetscCall(MatDestroy(&Xsurf)); PetscFunctionReturn(PETSC_SUCCESS); } /* DMDAGetFaceInterpolation - Gets the interpolation for a face based coarse space */ PetscErrorCode DMDAGetFaceInterpolation(PC pc, DM da, PC_Exotic *exotic, Mat Aglobal, MatReuse reuse, Mat *P) { PetscInt dim, i, j, k, m, n, p, dof, Nint, Nface, Nwire, Nsurf, *Iint, *Isurf, cint = 0, csurf = 0, istart, jstart, kstart, *II, N, c = 0; PetscInt mwidth, nwidth, pwidth, cnt, mp, np, pp, Ntotal, gl[6], *globals, Ng, *IIint, *IIsurf; Mat Xint, Xsurf, Xint_tmp; IS isint, issurf, is, row, col; ISLocalToGlobalMapping ltg; MPI_Comm comm; Mat A, Aii, Ais, Asi, *Aholder, iAii; MatFactorInfo info; PetscScalar *xsurf, *xint; const PetscScalar *rxint; #if defined(PETSC_USE_DEBUG_foo) PetscScalar tmp; #endif PetscHMapI ht; PetscFunctionBegin; PetscCall(DMDAGetInfo(da, &dim, NULL, NULL, NULL, &mp, &np, &pp, &dof, NULL, NULL, NULL, NULL, NULL)); PetscCheck(dof == 1, PetscObjectComm((PetscObject)da), PETSC_ERR_SUP, "Only for single field problems"); PetscCheck(dim == 3, PetscObjectComm((PetscObject)da), PETSC_ERR_SUP, "Only coded for 3d problems"); PetscCall(DMDAGetCorners(da, NULL, NULL, NULL, &m, &n, &p)); PetscCall(DMDAGetGhostCorners(da, &istart, &jstart, &kstart, &mwidth, &nwidth, &pwidth)); istart = istart ? -1 : 0; jstart = jstart ? -1 : 0; kstart = kstart ? -1 : 0; /* the columns of P are the interpolation of each coarse grid point (one for each vertex and edge) to all the local degrees of freedom (this includes the vertices, edges and faces). Xint are the subset of the interpolation into the interior Xface are the interpolation onto faces but not into the interior Xsurf are the interpolation onto the vertices and edges (the surfbasket) Xint Symbolically one could write P = (Xface) after interchanging the rows to match the natural ordering on the domain Xsurf */ N = (m - istart) * (n - jstart) * (p - kstart); Nint = (m - 2 - istart) * (n - 2 - jstart) * (p - 2 - kstart); Nface = 2 * ((m - 2 - istart) * (n - 2 - jstart) + (m - 2 - istart) * (p - 2 - kstart) + (n - 2 - jstart) * (p - 2 - kstart)); Nwire = 4 * ((m - 2 - istart) + (n - 2 - jstart) + (p - 2 - kstart)) + 8; Nsurf = Nface + Nwire; PetscCall(MatCreateSeqDense(MPI_COMM_SELF, Nint, 6, NULL, &Xint)); PetscCall(MatCreateSeqDense(MPI_COMM_SELF, Nsurf, 6, NULL, &Xsurf)); PetscCall(MatDenseGetArray(Xsurf, &xsurf)); /* Require that all 12 edges and 6 faces have at least one grid point. Otherwise some of the columns of Xsurf will be all zero (thus making the coarse matrix singular). */ PetscCheck(m - istart >= 3, PETSC_COMM_SELF, PETSC_ERR_SUP, "Number of grid points per process in X direction must be at least 3"); PetscCheck(n - jstart >= 3, PETSC_COMM_SELF, PETSC_ERR_SUP, "Number of grid points per process in Y direction must be at least 3"); PetscCheck(p - kstart >= 3, PETSC_COMM_SELF, PETSC_ERR_SUP, "Number of grid points per process in Z direction must be at least 3"); cnt = 0; for (j = 1; j < n - 1 - jstart; j++) { for (i = 1; i < m - istart - 1; i++) xsurf[cnt++ + 0 * Nsurf] = 1; } for (k = 1; k < p - 1 - kstart; k++) { for (i = 1; i < m - istart - 1; i++) xsurf[cnt++ + 1 * Nsurf] = 1; for (j = 1; j < n - 1 - jstart; j++) { xsurf[cnt++ + 2 * Nsurf] = 1; /* these are the interior nodes */ xsurf[cnt++ + 3 * Nsurf] = 1; } for (i = 1; i < m - istart - 1; i++) xsurf[cnt++ + 4 * Nsurf] = 1; } for (j = 1; j < n - 1 - jstart; j++) { for (i = 1; i < m - istart - 1; i++) xsurf[cnt++ + 5 * Nsurf] = 1; } #if defined(PETSC_USE_DEBUG_foo) for (i = 0; i < Nsurf; i++) { tmp = 0.0; for (j = 0; j < 6; j++) tmp += xsurf[i + j * Nsurf]; PetscCheck(PetscAbsScalar(tmp - 1.0) <= 1.e-10, PETSC_COMM_SELF, PETSC_ERR_PLIB, "Wrong Xsurf interpolation at i %" PetscInt_FMT " value %g", i, (double)PetscAbsScalar(tmp)); } #endif PetscCall(MatDenseRestoreArray(Xsurf, &xsurf)); /* PetscCall(MatView(Xsurf,PETSC_VIEWER_STDOUT_WORLD));*/ /* I are the indices for all the needed vertices (in global numbering) Iint are the indices for the interior values, I surf for the surface values (This is just for the part of the global matrix obtained with MatCreateSubMatrix(), it is NOT the local DMDA ordering.) IIint and IIsurf are the same as the Iint, Isurf except they are in the global numbering */ #define Endpoint(a, start, b) (a == 0 || a == (b - 1 - start)) PetscCall(PetscMalloc3(N, &II, Nint, &Iint, Nsurf, &Isurf)); PetscCall(PetscMalloc2(Nint, &IIint, Nsurf, &IIsurf)); for (k = 0; k < p - kstart; k++) { for (j = 0; j < n - jstart; j++) { for (i = 0; i < m - istart; i++) { II[c++] = i + j * mwidth + k * mwidth * nwidth; if (!Endpoint(i, istart, m) && !Endpoint(j, jstart, n) && !Endpoint(k, kstart, p)) { IIint[cint] = i + j * mwidth + k * mwidth * nwidth; Iint[cint++] = i + j * (m - istart) + k * (m - istart) * (n - jstart); } else { IIsurf[csurf] = i + j * mwidth + k * mwidth * nwidth; Isurf[csurf++] = i + j * (m - istart) + k * (m - istart) * (n - jstart); } } } } #undef Endpoint PetscCheck(c == N, PETSC_COMM_SELF, PETSC_ERR_PLIB, "c != N"); PetscCheck(cint == Nint, PETSC_COMM_SELF, PETSC_ERR_PLIB, "cint != Nint"); PetscCheck(csurf == Nsurf, PETSC_COMM_SELF, PETSC_ERR_PLIB, "csurf != Nsurf"); PetscCall(DMGetLocalToGlobalMapping(da, <g)); PetscCall(ISLocalToGlobalMappingApply(ltg, N, II, II)); PetscCall(ISLocalToGlobalMappingApply(ltg, Nint, IIint, IIint)); PetscCall(ISLocalToGlobalMappingApply(ltg, Nsurf, IIsurf, IIsurf)); PetscCall(PetscObjectGetComm((PetscObject)da, &comm)); PetscCall(ISCreateGeneral(comm, N, II, PETSC_COPY_VALUES, &is)); PetscCall(ISCreateGeneral(PETSC_COMM_SELF, Nint, Iint, PETSC_COPY_VALUES, &isint)); PetscCall(ISCreateGeneral(PETSC_COMM_SELF, Nsurf, Isurf, PETSC_COPY_VALUES, &issurf)); PetscCall(PetscFree3(II, Iint, Isurf)); PetscCall(ISSort(is)); PetscCall(MatCreateSubMatrices(Aglobal, 1, &is, &is, MAT_INITIAL_MATRIX, &Aholder)); A = *Aholder; PetscCall(PetscFree(Aholder)); PetscCall(MatCreateSubMatrix(A, isint, isint, MAT_INITIAL_MATRIX, &Aii)); PetscCall(MatCreateSubMatrix(A, isint, issurf, MAT_INITIAL_MATRIX, &Ais)); PetscCall(MatCreateSubMatrix(A, issurf, isint, MAT_INITIAL_MATRIX, &Asi)); /* Solve for the interpolation onto the interior Xint */ PetscCall(MatMatMult(Ais, Xsurf, MAT_INITIAL_MATRIX, PETSC_DETERMINE, &Xint_tmp)); PetscCall(MatScale(Xint_tmp, -1.0)); if (exotic->directSolve) { PetscCall(MatGetFactor(Aii, MATSOLVERPETSC, MAT_FACTOR_LU, &iAii)); PetscCall(MatFactorInfoInitialize(&info)); PetscCall(MatGetOrdering(Aii, MATORDERINGND, &row, &col)); PetscCall(MatLUFactorSymbolic(iAii, Aii, row, col, &info)); PetscCall(ISDestroy(&row)); PetscCall(ISDestroy(&col)); PetscCall(MatLUFactorNumeric(iAii, Aii, &info)); PetscCall(MatMatSolve(iAii, Xint_tmp, Xint)); PetscCall(MatDestroy(&iAii)); } else { Vec b, x; PetscScalar *xint_tmp; PetscCall(MatDenseGetArray(Xint, &xint)); PetscCall(VecCreateSeqWithArray(PETSC_COMM_SELF, 1, Nint, NULL, &x)); PetscCall(MatDenseGetArray(Xint_tmp, &xint_tmp)); PetscCall(VecCreateSeqWithArray(PETSC_COMM_SELF, 1, Nint, NULL, &b)); PetscCall(KSPSetOperators(exotic->ksp, Aii, Aii)); for (i = 0; i < 6; i++) { PetscCall(VecPlaceArray(x, xint + i * Nint)); PetscCall(VecPlaceArray(b, xint_tmp + i * Nint)); PetscCall(KSPSolve(exotic->ksp, b, x)); PetscCall(KSPCheckSolve(exotic->ksp, pc, x)); PetscCall(VecResetArray(x)); PetscCall(VecResetArray(b)); } PetscCall(MatDenseRestoreArray(Xint, &xint)); PetscCall(MatDenseRestoreArray(Xint_tmp, &xint_tmp)); PetscCall(VecDestroy(&x)); PetscCall(VecDestroy(&b)); } PetscCall(MatDestroy(&Xint_tmp)); #if defined(PETSC_USE_DEBUG_foo) PetscCall(MatDenseGetArrayRead(Xint, &rxint)); for (i = 0; i < Nint; i++) { tmp = 0.0; for (j = 0; j < 6; j++) tmp += rxint[i + j * Nint]; PetscCheck(PetscAbsScalar(tmp - 1.0) <= 1.e-10, PETSC_COMM_SELF, PETSC_ERR_PLIB, "Wrong Xint interpolation at i %" PetscInt_FMT " value %g", i, (double)PetscAbsScalar(tmp)); } PetscCall(MatDenseRestoreArrayRead(Xint, &rxint)); /* PetscCall(MatView(Xint,PETSC_VIEWER_STDOUT_WORLD)); */ #endif /* total faces */ Ntotal = mp * np * (pp + 1) + mp * pp * (np + 1) + np * pp * (mp + 1); /* For each vertex, edge, face on process (in the same orderings as used above) determine its local number including ghost points */ cnt = 0; { gl[cnt++] = mwidth + 1; } {{gl[cnt++] = mwidth * nwidth + 1; } { gl[cnt++] = mwidth * nwidth + mwidth; /* these are the interior nodes */ gl[cnt++] = mwidth * nwidth + mwidth + m - istart - 1; } { gl[cnt++] = mwidth * nwidth + mwidth * (n - jstart - 1) + 1; } } { gl[cnt++] = mwidth * nwidth * (p - kstart - 1) + mwidth + 1; } /* PetscIntView(6,gl,PETSC_VIEWER_STDOUT_WORLD); */ /* convert that to global numbering and get them on all processes */ PetscCall(ISLocalToGlobalMappingApply(ltg, 6, gl, gl)); /* PetscIntView(6,gl,PETSC_VIEWER_STDOUT_WORLD); */ PetscCall(PetscMalloc1(6 * mp * np * pp, &globals)); PetscCallMPI(MPI_Allgather(gl, 6, MPIU_INT, globals, 6, MPIU_INT, PetscObjectComm((PetscObject)da))); /* Number the coarse grid points from 0 to Ntotal */ PetscCall(PetscHMapICreateWithSize(Ntotal / 3, &ht)); for (i = 0, cnt = 0; i < 6 * mp * np * pp; i++) { PetscHashIter it = 0; PetscBool missing = PETSC_TRUE; PetscCall(PetscHMapIPut(ht, globals[i] + 1, &it, &missing)); if (missing) { ++cnt; PetscCall(PetscHMapIIterSet(ht, it, cnt)); } } PetscCheck(cnt == Ntotal, PETSC_COMM_SELF, PETSC_ERR_PLIB, "Hash table size %" PetscInt_FMT " not equal to total number coarse grid points %" PetscInt_FMT, cnt, Ntotal); PetscCall(PetscFree(globals)); for (i = 0; i < 6; i++) { PetscCall(PetscHMapIGetWithDefault(ht, gl[i] + 1, 0, gl + i)); --(gl[i]); } PetscCall(PetscHMapIDestroy(&ht)); /* PetscIntView(6,gl,PETSC_VIEWER_STDOUT_WORLD); */ /* construct global interpolation matrix */ PetscCall(MatGetLocalSize(Aglobal, &Ng, NULL)); if (reuse == MAT_INITIAL_MATRIX) { PetscCall(MatCreateAIJ(PetscObjectComm((PetscObject)da), Ng, PETSC_DECIDE, PETSC_DECIDE, Ntotal, Nint + Nsurf, NULL, Nint, NULL, P)); } else { PetscCall(MatZeroEntries(*P)); } PetscCall(MatSetOption(*P, MAT_ROW_ORIENTED, PETSC_FALSE)); PetscCall(MatDenseGetArrayRead(Xint, &rxint)); PetscCall(MatSetValues(*P, Nint, IIint, 6, gl, rxint, INSERT_VALUES)); PetscCall(MatDenseRestoreArrayRead(Xint, &rxint)); PetscCall(MatDenseGetArrayRead(Xsurf, &rxint)); PetscCall(MatSetValues(*P, Nsurf, IIsurf, 6, gl, rxint, INSERT_VALUES)); PetscCall(MatDenseRestoreArrayRead(Xsurf, &rxint)); PetscCall(MatAssemblyBegin(*P, MAT_FINAL_ASSEMBLY)); PetscCall(MatAssemblyEnd(*P, MAT_FINAL_ASSEMBLY)); PetscCall(PetscFree2(IIint, IIsurf)); #if defined(PETSC_USE_DEBUG_foo) { Vec x, y; PetscScalar *yy; PetscCall(VecCreateMPI(PetscObjectComm((PetscObject)da), Ng, PETSC_DETERMINE, &y)); PetscCall(VecCreateMPI(PetscObjectComm((PetscObject)da), PETSC_DETERMINE, Ntotal, &x)); PetscCall(VecSet(x, 1.0)); PetscCall(MatMult(*P, x, y)); PetscCall(VecGetArray(y, &yy)); for (i = 0; i < Ng; i++) PetscCheck(PetscAbsScalar(yy[i] - 1.0) <= 1.e-10, PETSC_COMM_SELF, PETSC_ERR_PLIB, "Wrong p interpolation at i %" PetscInt_FMT " value %g", i, (double)PetscAbsScalar(yy[i])); PetscCall(VecRestoreArray(y, &yy)); PetscCall(VecDestroy(x)); PetscCall(VecDestroy(y)); } #endif PetscCall(MatDestroy(&Aii)); PetscCall(MatDestroy(&Ais)); PetscCall(MatDestroy(&Asi)); PetscCall(MatDestroy(&A)); PetscCall(ISDestroy(&is)); PetscCall(ISDestroy(&isint)); PetscCall(ISDestroy(&issurf)); PetscCall(MatDestroy(&Xint)); PetscCall(MatDestroy(&Xsurf)); PetscFunctionReturn(PETSC_SUCCESS); } /*@ PCExoticSetType - Sets the type of coarse grid interpolation to use Logically Collective Input Parameters: + pc - the preconditioner context - type - either PC_EXOTIC_FACE or PC_EXOTIC_WIREBASKET (defaults to face) Notes: The face based interpolation has 1 degree of freedom per face and ignores the edge and vertex values completely in the coarse problem. For any seven point stencil the interpolation of a constant on all faces into the interior is that constant. The wirebasket interpolation has 1 degree of freedom per vertex, per edge and per face. A constant on the subdomain boundary is interpolated as that constant in the interior of the domain. The coarse grid matrix is obtained via the Galerkin computation A_c = R A R^T, hence if A is nonsingular A_c is also nonsingular. Both interpolations are suitable for only scalar problems. Level: intermediate .seealso: `PCEXOTIC`, `PCExoticType()` @*/ PetscErrorCode PCExoticSetType(PC pc, PCExoticType type) { PetscFunctionBegin; PetscValidHeaderSpecific(pc, PC_CLASSID, 1); PetscValidLogicalCollectiveEnum(pc, type, 2); PetscTryMethod(pc, "PCExoticSetType_C", (PC, PCExoticType), (pc, type)); PetscFunctionReturn(PETSC_SUCCESS); } static PetscErrorCode PCExoticSetType_Exotic(PC pc, PCExoticType type) { PC_MG *mg = (PC_MG *)pc->data; PC_Exotic *ctx = (PC_Exotic *)mg->innerctx; PetscFunctionBegin; ctx->type = type; PetscFunctionReturn(PETSC_SUCCESS); } PetscErrorCode PCSetUp_Exotic(PC pc) { Mat A; PC_MG *mg = (PC_MG *)pc->data; PC_Exotic *ex = (PC_Exotic *)mg->innerctx; MatReuse reuse = (ex->P) ? MAT_REUSE_MATRIX : MAT_INITIAL_MATRIX; PetscFunctionBegin; PetscCheck(pc->dm, PetscObjectComm((PetscObject)pc), PETSC_ERR_ARG_WRONGSTATE, "Need to call PCSetDM() before using this PC"); PetscCall(PCGetOperators(pc, NULL, &A)); PetscCheck(ex->type == PC_EXOTIC_FACE || ex->type == PC_EXOTIC_WIREBASKET, PetscObjectComm((PetscObject)pc), PETSC_ERR_PLIB, "Unknown exotic coarse space %d", ex->type); if (ex->type == PC_EXOTIC_FACE) { PetscCall(DMDAGetFaceInterpolation(pc, pc->dm, ex, A, reuse, &ex->P)); } else /* if (ex->type == PC_EXOTIC_WIREBASKET) */ { PetscCall(DMDAGetWireBasketInterpolation(pc, pc->dm, ex, A, reuse, &ex->P)); } PetscCall(PCMGSetInterpolation(pc, 1, ex->P)); /* if PC has attached DM we must remove it or the PCMG will use it to compute incorrect sized vectors and interpolations */ PetscCall(PCSetDM(pc, NULL)); PetscCall(PCSetUp_MG(pc)); PetscFunctionReturn(PETSC_SUCCESS); } PetscErrorCode PCDestroy_Exotic(PC pc) { PC_MG *mg = (PC_MG *)pc->data; PC_Exotic *ctx = (PC_Exotic *)mg->innerctx; PetscFunctionBegin; PetscCall(MatDestroy(&ctx->P)); PetscCall(KSPDestroy(&ctx->ksp)); PetscCall(PetscFree(ctx)); PetscCall(PetscObjectComposeFunction((PetscObject)pc, "PCExoticSetType_C", NULL)); PetscCall(PCDestroy_MG(pc)); PetscFunctionReturn(PETSC_SUCCESS); } PetscErrorCode PCView_Exotic(PC pc, PetscViewer viewer) { PC_MG *mg = (PC_MG *)pc->data; PetscBool iascii; PC_Exotic *ctx = (PC_Exotic *)mg->innerctx; PetscFunctionBegin; PetscCall(PetscObjectTypeCompare((PetscObject)viewer, PETSCVIEWERASCII, &iascii)); if (iascii) { PetscCall(PetscViewerASCIIPrintf(viewer, " Exotic type = %s\n", PCExoticTypes[ctx->type])); if (ctx->directSolve) { PetscCall(PetscViewerASCIIPrintf(viewer, " Using direct solver to construct interpolation\n")); } else { PetscViewer sviewer; PetscMPIInt rank; PetscCall(PetscViewerASCIIPrintf(viewer, " Using iterative solver to construct interpolation\n")); PetscCall(PetscViewerASCIIPushTab(viewer)); PetscCall(PetscViewerASCIIPushTab(viewer)); /* should not need to push this twice? */ PetscCall(PetscViewerGetSubViewer(viewer, PETSC_COMM_SELF, &sviewer)); PetscCallMPI(MPI_Comm_rank(PetscObjectComm((PetscObject)pc), &rank)); if (rank == 0) PetscCall(KSPView(ctx->ksp, sviewer)); PetscCall(PetscViewerRestoreSubViewer(viewer, PETSC_COMM_SELF, &sviewer)); PetscCall(PetscViewerASCIIPopTab(viewer)); PetscCall(PetscViewerASCIIPopTab(viewer)); } } PetscCall(PCView_MG(pc, viewer)); PetscFunctionReturn(PETSC_SUCCESS); } PetscErrorCode PCSetFromOptions_Exotic(PC pc, PetscOptionItems *PetscOptionsObject) { PetscBool flg; PC_MG *mg = (PC_MG *)pc->data; PCExoticType mgctype; PC_Exotic *ctx = (PC_Exotic *)mg->innerctx; PetscFunctionBegin; PetscOptionsHeadBegin(PetscOptionsObject, "Exotic coarse space options"); PetscCall(PetscOptionsEnum("-pc_exotic_type", "face or wirebasket", "PCExoticSetType", PCExoticTypes, (PetscEnum)ctx->type, (PetscEnum *)&mgctype, &flg)); if (flg) PetscCall(PCExoticSetType(pc, mgctype)); PetscCall(PetscOptionsBool("-pc_exotic_direct_solver", "use direct solver to construct interpolation", "None", ctx->directSolve, &ctx->directSolve, NULL)); if (!ctx->directSolve) { if (!ctx->ksp) { const char *prefix; PetscCall(KSPCreate(PETSC_COMM_SELF, &ctx->ksp)); PetscCall(KSPSetNestLevel(ctx->ksp, pc->kspnestlevel)); PetscCall(KSPSetErrorIfNotConverged(ctx->ksp, pc->erroriffailure)); PetscCall(PetscObjectIncrementTabLevel((PetscObject)ctx->ksp, (PetscObject)pc, 1)); PetscCall(PCGetOptionsPrefix(pc, &prefix)); PetscCall(KSPSetOptionsPrefix(ctx->ksp, prefix)); PetscCall(KSPAppendOptionsPrefix(ctx->ksp, "exotic_")); } PetscCall(KSPSetFromOptions(ctx->ksp)); } PetscOptionsHeadEnd(); PetscFunctionReturn(PETSC_SUCCESS); } /*MC PCEXOTIC - Two level overlapping Schwarz preconditioner with exotic (non-standard) coarse grid spaces This uses the `PCMG` infrastructure restricted to two levels and the face and wirebasket based coarse grid spaces. Level: advanced Notes: Must be used with `DMDA` By default this uses `KSPGMRES` on the fine grid smoother so this should be used with `KSPFGMRES` or the smoother changed to not use `KSPGMRES` References: + * - These coarse grid spaces originate in the work of Bramble, Pasciak and Schatz, "The Construction of Preconditioners for Elliptic Problems by Substructing IV", Mathematics of Computation, volume 53, 1989. . * - They were generalized slightly in "Domain Decomposition Method for Linear Elasticity", Ph. D. thesis, Barry Smith, New York University, 1990. . * - They were then explored in great detail in Dryja, Smith, Widlund, "Schwarz Analysis of Iterative Substructuring Methods for Elliptic Problems in Three Dimensions, SIAM Journal on Numerical Analysis, volume 31. 1994. These were developed in the context of iterative substructuring preconditioners. . * - They were then ingeniously applied as coarse grid spaces for overlapping Schwarz methods by Dohrmann and Widlund. They refer to them as GDSW (generalized Dryja, Smith, Widlund preconditioners). See, for example, Clark R. Dohrmann, Axel Klawonn, and Olof B. Widlund. Extending theory for domain decomposition algorithms to irregular subdomains. In Ulrich Langer, Marco Discacciati, David Keyes, Olof Widlund, and Walter Zulehner, editors, Proceedings of the 17th International Conference on Domain Decomposition Methods in Science and Engineering, held in Strobl, Austria, 2006, number 60 in Springer Verlag, Lecture Notes in Computational Science and Engineering, 2007. . * - Clark R. Dohrmann, Axel Klawonn, and Olof B. Widlund. A family of energy minimizing coarse spaces for overlapping Schwarz preconditioners. In Ulrich Langer, Marco Discacciati, David Keyes, Olof Widlund, and Walter Zulehner, editors, Proceedings of the 17th International Conference on Domain Decomposition Methods in Science and Engineering, held in Strobl, Austria, 2006, number 60 in Springer Verlag, Lecture Notes in Computational Science and Engineering, 2007 . * - Clark R. Dohrmann, Axel Klawonn, and Olof B. Widlund. Domain decomposition for less regular subdomains: Overlapping Schwarz in two dimensions. SIAM J. Numer. Anal., 46(4), 2008. - * - Clark R. Dohrmann and Olof B. Widlund. An overlapping Schwarz algorithm for almost incompressible elasticity. Technical Report TR2008 912, Department of Computer Science, Courant Institute of Mathematical Sciences, New York University, May 2008. URL: The usual `PCMG` options are supported, such as -mg_levels_pc_type -mg_coarse_pc_type and -pc_mg_type .seealso: `PCMG`, `PCSetDM()`, `PCExoticType`, `PCExoticSetType()` M*/ PETSC_EXTERN PetscErrorCode PCCreate_Exotic(PC pc) { PC_Exotic *ex; PC_MG *mg; PetscFunctionBegin; /* if type was previously mg; must manually destroy it because call to PCSetType(pc,PCMG) will not destroy it */ PetscTryTypeMethod(pc, destroy); pc->data = NULL; PetscCall(PetscFree(((PetscObject)pc)->type_name)); ((PetscObject)pc)->type_name = NULL; PetscCall(PCSetType(pc, PCMG)); PetscCall(PCMGSetLevels(pc, 2, NULL)); PetscCall(PCMGSetGalerkin(pc, PC_MG_GALERKIN_PMAT)); PetscCall(PetscNew(&ex)); ex->type = PC_EXOTIC_FACE; mg = (PC_MG *)pc->data; mg->innerctx = ex; pc->ops->setfromoptions = PCSetFromOptions_Exotic; pc->ops->view = PCView_Exotic; pc->ops->destroy = PCDestroy_Exotic; pc->ops->setup = PCSetUp_Exotic; PetscCall(PetscObjectComposeFunction((PetscObject)pc, "PCExoticSetType_C", PCExoticSetType_Exotic)); PetscFunctionReturn(PETSC_SUCCESS); }