1 /* 2 DMPatch, for domains covered by sets of patches. 3 */ 4 #if !defined(PETSCDMPATCH_H) 5 #define PETSCDMPATCH_H 6 7 #include <petscdm.h> 8 9 /* SUBMANSEC = DMPatch */ 10 11 /*S 12 DMPATCH - DM object that encapsulates a domain divided into many patches 13 14 Level: intermediate 15 16 .seealso: `DM`, `DMPatchCreate()` 17 S*/ 18 PETSC_EXTERN PetscErrorCode DMPatchCreate(MPI_Comm, DM*); 19 20 PETSC_EXTERN PetscErrorCode DMPatchZoom(DM,MatStencil,MatStencil,MPI_Comm,DM*,PetscSF*,PetscSF*); 21 PETSC_EXTERN PetscErrorCode DMPatchSolve(DM); 22 PETSC_EXTERN PetscErrorCode DMPatchGetPatchSize(DM,MatStencil*); 23 PETSC_EXTERN PetscErrorCode DMPatchSetPatchSize(DM,MatStencil); 24 PETSC_EXTERN PetscErrorCode DMPatchGetCommSize(DM,MatStencil*); 25 PETSC_EXTERN PetscErrorCode DMPatchSetCommSize(DM,MatStencil); 26 PETSC_EXTERN PetscErrorCode DMPatchGetCoarse(DM,DM*); 27 PETSC_EXTERN PetscErrorCode DMPatchCreateGrid(MPI_Comm,PetscInt,MatStencil,MatStencil,MatStencil,DM*); 28 29 /* 30 * We want each patch to consist of an entire DM, DMDA at first 31 - We cannot afford to store much more than the data from a single patch in memory 32 - No global PetscSection, only PetscLayout 33 - Optional scatters 34 * There is a storable coarse level, which will also be a traditional DM (DMDA here) 35 * The local and global vectors correspond to a ghosted patch 36 * Need a way to activate a patch 37 * Jack in sizes for l/g vectors 38 - Need routine for viewing a full global vector 39 - Jed handles solver 40 */ 41 42 #endif 43