| /petsc/src/ts/tests/ |
| H A D | ex30.c | 290 for (PetscInt grid = 0; grid < ctx->num_grids; grid++) { // add same particels for all grids in gridToParticles_private() local 292 …t_FMT ", index(b,g) %" PetscInt_FMT "\n", global_vertex_id, v_id, grid, LAND_PACK_IDX(v_id, grid)); in gridToParticles_private() 293 …grid], globSwarmArray[LAND_PACK_IDX(v_id, grid)], globXArray[LAND_PACK_IDX(v_id, grid)], t_fhat[gr… in gridToParticles_private() 304 for (PetscInt grid = 0; grid < ctx->num_grids; grid++) { in gridToParticles_private() local 307 DM sw = globSwarmArray[LAND_PACK_IDX(v_id, grid)]; in gridToParticles_private() 313 …PI * coords[p * dim + 0] : 1, w = fact * wp[p] * ctx->n_0 * ctx->masses[ctx->species_offset[grid]]; in gridToParticles_private() 324 for (PetscInt grid = 0; grid < ctx->num_grids; grid++) { in gridToParticles_private() local 327 DM sw = globSwarmArray[LAND_PACK_IDX(v_id, grid)]; in gridToParticles_private() 333 …dim + 0] : 1, w = fact * wp[p] * ctx->n_0 * ctx->masses[ctx->species_offset[grid]], ww = w * N_inv; in gridToParticles_private() 393 for (PetscInt grid = 0; grid < ctx->num_grids; grid++) { in PostStep() local [all …]
|
| /petsc/src/ts/utils/dmplexlandau/ |
| H A D | plexland.c | 37 for (PetscInt grid = 0; grid < maps[0].numgrids; grid++) { in LandauGPUMapsDestroy() local 38 PetscCall(PetscFree(maps[grid].c_maps)); in LandauGPUMapsDestroy() 39 PetscCall(PetscFree(maps[grid].gIdx)); in LandauGPUMapsDestroy() 116 …for (PetscInt grid = 0; grid < ctx->num_grids; grid++) PetscCall(DMCreateMatrix(ctx->plex[grid], &… in LandauFormJacobian_Internal() local 127 for (PetscInt grid = 0; grid < ctx->num_grids; grid++) { in LandauFormJacobian_Internal() local 129 PetscCheck(ctx->plex[grid] != NULL, ctx->comm, PETSC_ERR_ARG_WRONG, "Plex not created"); in LandauFormJacobian_Internal() 130 PetscCall(DMPlexGetHeightStratum(ctx->plex[grid], 0, &cStart, &cEnd)); in LandauFormJacobian_Internal() 131 numCells[grid] = cEnd - cStart; // grids can have different topology in LandauFormJacobian_Internal() 148 for (PetscInt grid = 0; grid < ctx->num_grids; grid++) { in LandauFormJacobian_Internal() local 149 PetscCall(DMGetLocalSection(ctx->plex[grid], §ion[grid])); in LandauFormJacobian_Internal() [all …]
|
| /petsc/src/snes/tutorials/ |
| H A D | ex30.c | 104 GridInfo *grid; member 128 GridInfo grid; in main() local 147 PetscCall(SetParams(¶m, &grid)); in main() 148 PetscCall(ReportParams(¶m, &grid)); in main() 153 …PetscCall(DMDACreate2d(comm, grid.bx, grid.by, grid.stencil, grid.ni, grid.nj, PETSC_DECIDE, PETSC… in main() 167 user->grid = &grid; in main() 309 GridInfo *grid = user->grid; in HorizVelocity() local 313 x = (i - grid->jlid) * grid->dx; in HorizVelocity() 314 z = (j - grid->jlid - 0.5) * grid->dz; in HorizVelocity() 326 GridInfo *grid = user->grid; in VertVelocity() local [all …]
|
| /petsc/doc/community/ |
| H A D | petsc_team.md | 29 .. grid:: 2 3 3 4 31 .. grid-item-card:: Mark Adams 35 .. grid-item-card:: Satish Balay 39 .. grid-item-card:: Jed Brown 43 .. grid-item-card:: Lisandro Dalcin 47 .. grid-item-card:: Toby Isaac 51 .. grid-item-card:: Pierre Jolivet 55 .. grid-item-card:: Matthew Knepley 59 .. grid-item-card:: Lois Curfman McInnes 63 .. grid-item-card:: Richard Tran Mills [all …]
|
| /petsc/src/ts/utils/dmplexlandau/kokkos/ |
| H A D | landau.kokkos.cxx | 93 …Maps maps[], pointInterpolationP4est (*pointMaps)[LANDAU_MAX_Q_FACE], PetscInt Nf[], PetscInt grid) in LandauKokkosCreateMatMaps() argument 96 …its<Kokkos::Unmanaged>> h_points((pointInterpolationP4est *)pointMaps, maps[grid].num_reduced); in LandauKokkosCreateMatMaps() 97 …Kokkos::MemoryTraits<Kokkos::Unmanaged>> h_gidx((LandauIdx *)maps[grid].gIdx, maps[grid].num_eleme… in LandauKokkosCreateMatMaps() 98 …intInterpolationP4est *[LANDAU_MAX_Q_FACE], Kokkos::LayoutRight>("points", maps[grid].num_reduced); in LandauKokkosCreateMatMaps() 99 …uIdx *[LANDAU_MAX_SPECIES][LANDAU_MAX_NQND], Kokkos::LayoutRight>("gIdx", maps[grid].num_elements); in LandauKokkosCreateMatMaps() 104 h_maps.num_elements = maps[grid].num_elements; in LandauKokkosCreateMatMaps() 105 h_maps.num_face = maps[grid].num_face; in LandauKokkosCreateMatMaps() 106 h_maps.num_reduced = maps[grid].num_reduced; in LandauKokkosCreateMatMaps() 107 h_maps.deviceType = maps[grid].deviceType; in LandauKokkosCreateMatMaps() 108 h_maps.numgrids = maps[grid].numgrids; in LandauKokkosCreateMatMaps() [all …]
|
| /petsc/src/binding/petsc4py/src/petsc4py/lib/_pytypes/viewer/ |
| H A D | petscpyvista.py | 105 grid = self.convertDMToPV(dm) 116 if scalars[1].shape[0] / scalars[2] == grid.n_cells: 117 grid.cell_data[scalars[0]] = scalars[1] 118 elif scalars[1].shape[0] / scalars[2] == grid.n_points: 120 grid.point_data[scalars[0]] = scalars[1].reshape(-1, scalars[2]) 124 grid.point_data[scalars[0]] = vecs 126 …%d) or vertices (%d)' % (scalars[0], scalars[1].shape[0], scalars[2], grid.n_cells, grid.n_points)) 129 grid = grid.warp_by_scalar(factor = self.warpFactor) 131 grid = grid.warp_by_vector(factor = self.warpFactor) 133 grid = grid.clip_box(self.clipBounds) [all …]
|
| /petsc/src/dm/tutorials/ |
| H A D | ex13f90.F90 | 36 PetscReal, allocatable :: f(:, :, :, :), grid(:, :, :, :) 91 allocate (grid(ndim, ib1:ibn, jb1:jbn, kb1:kbn)) 100 call petsc_to_local(CoordDM, coords, array, grid, ndim, stw) 101 xmin = grid(1, 1, 1, 1) 102 ymin = grid(2, 1, 1, 1) 103 zmin = grid(3, 1, 1, 1) 104 xmax = grid(1, imax, jmax, kmax) 105 ymax = grid(2, imax, jmax, kmax) 106 zmax = grid(3, imax, jmax, kmax) 107 call local_to_petsc(CoordDM, coords, array, grid, ndim, stw) [all …]
|
| /petsc/src/mat/impls/scalapack/ |
| H A D | matscalapack.c | 43 …rASCIIPrintf(viewer, "grid height=%d, grid width=%d\n", (int)a->grid->nprow, (int)a->grid->npcol)); in MatView_ScaLAPACK() 128 …, SCALAPACKinfog2l_(&gridx, &gcidx, a->desc, &a->grid->nprow, &a->grid->npcol, &a->grid->myrow, &a… in MatSetValues_ScaLAPACK() 129 if (rsrc == a->grid->myrow && csrc == a->grid->mycol) { in MatSetValues_ScaLAPACK() 172 …nit", SCALAPACKdescinit_(xdesc, &a->M, &one, &mb, &one, &zero, &zero, &a->grid->ictxcol, &xlld, &i… in MatMultXXXYYY_ScaLAPACK() 177 …nit", SCALAPACKdescinit_(ydesc, &one, &a->N, &one, &nb, &zero, &zero, &a->grid->ictxrow, &ylld, &i… in MatMultXXXYYY_ScaLAPACK() 181 lszx = SCALAPACKnumroc_(&a->M, &a->mb, &a->grid->myrow, &a->rsrc, &a->grid->nprow); in MatMultXXXYYY_ScaLAPACK() 182 lszy = SCALAPACKnumroc_(&a->N, &a->nb, &a->grid->mycol, &a->csrc, &a->grid->npcol); in MatMultXXXYYY_ScaLAPACK() 187 …, SCALAPACKdescinit_(x2desc, &a->M, &one, &a->mb, &one, &zero, &zero, &a->grid->ictxt, &xlld, &inf… in MatMultXXXYYY_ScaLAPACK() 189 …, SCALAPACKdescinit_(y2desc, &one, &a->N, &one, &a->nb, &zero, &zero, &a->grid->ictxt, &ylld, &inf… in MatMultXXXYYY_ScaLAPACK() 193 …, SCALAPACKgemr2d_(&a->M, &one, x, &one, &one, xdesc, x2d, &one, &one, x2desc, &a->grid->ictxcol)); in MatMultXXXYYY_ScaLAPACK() [all …]
|
| /petsc/src/tao/unconstrained/tutorials/ |
| H A D | adjointreads.m | 14 %xg - the grid 49 plot(grid,Init_ts,'ko-','Markersize',10); drawnow; 51 %plot(grid,temp,'c*','Markersize',10); drawnow; 52 plot(grid,exact,'r*','Markersize',10); drawnow; 53 %plot(grid,Curr_sol,'g-','Markersize',10); drawnow; 54 %plot(grid,Init_adj,'bo-','Markersize',10); drawnow; 56 xlabel('x (GLL grid)'); 76 % xlabel('x (GLL grid)'); 93 % grid on 107 % xlabel('GLL grid'); [all …]
|
| /petsc/src/ts/utils/dmplexlandau/tutorials/ |
| H A D | ex1.c | 23 PetscErrorCode landau_field_print_access_callback(DM dm, Vec x, PetscInt local_field, PetscInt grid… in landau_field_print_access_callback() argument 31 species = ctx->species_offset[grid] + local_field; in landau_field_print_access_callback() 32 val = (PetscScalar)(LAND_PACK_IDX(b_id, grid) + (species + 1) * 10); in landau_field_print_access_callback() 34 …_FMT " and local field %" PetscInt_FMT " with %" PetscInt_FMT " grids\n", grid, b_id, local_field,… in landau_field_print_access_callback() 93 …DM dm, Vec X, PetscReal time, PetscReal temps[], PetscReal ns[], PetscInt grid, PetscReal shifts[]… in SetMaxwellians() argument 100 …for (PetscInt ii = ctx->species_offset[grid], i0 = 0; ii < ctx->species_offset[grid + 1]; ii++, i0… in SetMaxwellians() 110 …ign(ctx->charges[ctx->species_offset[grid]]) * ctx->electronShift * ctx->m_0 / ctx->masses[ctx->sp… in SetMaxwellians() 114 …data[0].shift = ctx->electronShift * shifts[grid] * PetscSqrtReal(data[0].kT_m) / ctx->v_0; // shi… in SetMaxwellians() 256 for (PetscInt grid = 0; grid < ctx->num_grids; grid++) { in Monitor() local 258 DM dm = ctx->plex[grid]; in Monitor() [all …]
|
| /petsc/include/petsc/private/ |
| H A D | petscscalapack.h | 15 Mat_ScaLAPACK_Grid *grid; /* process grid */ member 121 …rc == _bb->csrc && _aa->grid->nprow == _bb->grid->nprow && _aa->grid->npcol == _bb->grid->npcol &&… 122 _aa->grid->mycol == _bb->grid->mycol, \
|
| /petsc/src/ts/utils/dmplexlandau/tutorials/output/ |
| H A D | ex2_simplex.out | 4 thermal T (K): e= 2.321e+07 i= 1.160e+07 0.000e+00. Normalization grid 0: v_0= 1.876e+07 ( 6.256e-… 5 Domain radius (AMR levels) grid 0: par= 5.000e+00 perp= 5.000e+00 (0) , 1: par= 5.835e-02 perp= 5.8… 20 …2.3596368842425e-18, energy= 7.1246540087660e+04 (m_i[0]/m_e = 3670.94, 48 cells on electron grid) 35 …9.7382234546064e-18, energy= 7.1246540087660e+04 (m_i[0]/m_e = 3670.94, 48 cells on electron grid) 38 …2.3596368842425e-18, energy= 7.1246540087660e+04 (m_i[0]/m_e = 3670.94, 48 cells on electron grid) 53 …9.7382234546064e-18, energy= 7.1246540087660e+04 (m_i[0]/m_e = 3670.94, 48 cells on electron grid)
|
| H A D | ex1_sphere.out | 4 thermal T (K): e= 2.321e+07 i= 1.160e+07 0.000e+00. Normalization grid 0: v_0= 1.876e+07 ( 6.256e-… 5 Domain radius (AMR levels) grid 0: par= 5.000e+00 perp= 5.000e+00 (1) , 1: par= 5.835e-02 perp= 5.8… 10 …5.6143870238028e-19, energy= 7.3772717719457e+04 (m_i[0]/m_e = 3670.94, 12 cells on electron grid) 15 …4.1036480213379e-18, energy= 7.3772717719441e+04 (m_i[0]/m_e = 3670.94, 12 cells on electron grid)
|
| H A D | ex2_0.out | 4 thermal T (K): e= 5.802e+07 i= 5.802e+07 0.000e+00. Normalization grid 0: v_0= 2.965e+07 ( 9.892e-… 5 Domain radius (AMR levels) grid 0: par= 5.000e+00 perp= 5.000e+00 (2) , 1: par= 8.252e-02 perp= 8.2… 10 …5.1936165265558e-18, energy= 2.4015413557925e+05 (m_i[0]/m_e = 3670.94, 14 cells on electron grid) 33 …1.1060501186683e-19, energy= 2.4015413557925e+05 (m_i[0]/m_e = 3670.94, 14 cells on electron grid) 37 …5.1936165265558e-18, energy= 2.4015413557925e+05 (m_i[0]/m_e = 3670.94, 14 cells on electron grid) 80 …4.5045672869187e-18, energy= 2.4015413557925e+05 (m_i[0]/m_e = 3670.94, 14 cells on electron grid)
|
| H A D | ex1_0.out | 4 thermal T (K): e= 3.481e+06 i= 2.321e+06 0.000e+00. Normalization grid 0: v_0= 7.264e+06 ( 2.423e-… 5 Domain radius (AMR levels) grid 0: par= 6.000e+00 perp= 6.000e+00 (0) , 1: par= 8.086e-02 perp= 8.0… 23 …1.1333771884393e-07, energy= 1.9821738917588e+04 (m_i[0]/m_e = 3670.94, 26 cells on electron grid) 32 …1.1333771884393e-07, energy= 1.9821738917588e+04 (m_i[0]/m_e = 3670.94, 26 cells on electron grid)
|
| H A D | ex1f90_0.out | 4 thermal T (K): e= 5.802e+07 i= 5.802e+07 5.802e+06. Normalization grid 1: v_0= 4.894e+05 ( 1.633e-… 5 Domain radius (AMR levels) grid 0: par= 3.029e+02 perp= 3.029e+02 (2) , 1: par= 5.000e+00 perp= 5.0… 20 …1.1601092271659e-17, energy= 2.3762108515975e+05 (m_i[0]/m_e = 3670.94, 14 cells on electron grid) 48 …4.9547999880240e-22, energy= 2.3762108515975e+05 (m_i[0]/m_e = 3670.94, 14 cells on electron grid)
|
| /petsc/src/benchmarks/ |
| H A D | benchmarkAssembly.py | 112 grid = [10]*len(procs) variable 114 grid = [1250]*len(procs) variable 117 grid = [100, 150, 200, 250, 300] variable 119 grid = range(150, 1350, 100) variable 120 procs = [1]*len(grid) 121 for n, p in zip(grid, procs):
|
| /petsc/src/ksp/ksp/tutorials/output/ |
| H A D | ex56_latebs-2.out | 25 Using externally compute Galerkin coarse grid matrices 29 Using parallel coarse grid solver (all coarse grid equations not put on one process) 36 Complexity: grid = 1.054 operator = 1.07125 41 Coarse grid solver -- level 0 ------------------------------- 113 Using externally compute Galerkin coarse grid matrices 117 Using parallel coarse grid solver (all coarse grid equations not put on one process) 124 Complexity: grid = 1.054 operator = 1.07125 129 Coarse grid solver -- level 0 ------------------------------- 211 Using externally compute Galerkin coarse grid matrices 215 Using parallel coarse grid solver (all coarse grid equations not put on one process) [all …]
|
| /petsc/src/ksp/ksp/tests/ |
| H A D | ex29.c | 118 PetscErrorCode FormJacobian_Grid(GridCtx *grid, Mat *J) in FormJacobian_Grid() argument 127 mx = grid->mx; in FormJacobian_Grid() 128 my = grid->my; in FormJacobian_Grid() 135 PetscCall(DMDAGetCorners(grid->da, &xs, &ys, 0, &xm, &ym, 0)); in FormJacobian_Grid() 136 PetscCall(DMDAGetGhostCorners(grid->da, &Xs, &Ys, 0, &Xm, &Ym, 0)); in FormJacobian_Grid() 137 PetscCall(DMGetLocalToGlobalMapping(grid->da, <ogm)); in FormJacobian_Grid()
|
| H A D | ex26.c | 148 PetscErrorCode FormJacobian_Grid(GridCtx *grid, Mat jac) in FormJacobian_Grid() argument 157 mx = grid->mx; in FormJacobian_Grid() 158 my = grid->my; in FormJacobian_Grid() 165 PetscCall(DMDAGetCorners(grid->da, &xs, &ys, 0, &xm, &ym, 0)); in FormJacobian_Grid() 166 PetscCall(DMDAGetGhostCorners(grid->da, &Xs, &Ys, 0, &Xm, &Ym, 0)); in FormJacobian_Grid() 167 PetscCall(DMGetLocalToGlobalMapping(grid->da, <ogm)); in FormJacobian_Grid()
|
| /petsc/src/mat/impls/aij/mpi/superlu_dist/ |
| H A D | superlu_dist.c | 94 gridinfo_t grid; member 126 …lVoid("SuperLU_DIST:pGetDiagU", pGetDiagU(F->rmap->N, &lu->LUstruct, &lu->grid, CASTDOUBLECOMPLEX … in MatSuperluDistGetDiagU_SuperLU_DIST() 142 gridinfo_t grid; member 162 PetscStackCallExternalVoid("SuperLU_DIST:superlu_gridexit", superlu_gridexit(&context->grid)); in Petsc_Superlu_dist_keyval_DeleteFn() 225 …PetscStackCallExternalVoid("SuperLU_DIST:Destroy_LU", sDestroy_LU(A->cmap->N, &lu->grid, &lu->sLUs… in MatDestroy_SuperLU_DIST() 231 …PetscStackCallExternalVoid("SuperLU_DIST:Destroy_LU", Destroy_LU(A->cmap->N, &lu->grid, &lu->LUstr… in MatDestroy_SuperLU_DIST() 242 PetscStackCallExternalVoid("SuperLU_DIST:superlu_gridexit", superlu_gridexit(&lu->grid)); in MatDestroy_SuperLU_DIST() 323 …u->options, &lu->A_sup, &lu->sScalePermstruct, lu->sbptr, (int)m, 1, &lu->grid, &lu->sLUstruct, &l… in MatSolve_SuperLU_DIST() 326 …&lu->A_sup, &lu->ScalePermstruct, CASTDOUBLECOMPLEX bptr, (int)m, 1, &lu->grid, &lu->LUstruct, &lu… in MatSolve_SuperLU_DIST() 329 …ernalVoid("SuperLU_DIST:PStatPrint", PStatPrint(&lu->options, &stat, &lu->grid)); /* Print the sta… in MatSolve_SuperLU_DIST() [all …]
|
| /petsc/src/ksp/ksp/tests/output/ |
| H A D | ex19_2.out | 1 Coarse grid size 5 by 5 2 Fine grid size 9 by 9
|
| H A D | ex19_1.out | 1 Coarse grid size 5 by 5 2 Fine grid size 9 by 9
|
| /petsc/src/ksp/ksp/tests/raja/ |
| H A D | ex1.raja.cxx | 93 void computeErr(double *I, grid_s grid); 410 void computeErr(double *I, grid_s grid) in computeErr() argument 412 RAJA::RangeSegment gridRange(0, grid.n); in computeErr() 418 int id = tx + grid.n * ty; in computeErr() 419 double x = grid.o + tx * grid.h; in computeErr() 420 double y = grid.o + ty * grid.h; in computeErr() 426 printf("Max error = %lg, h = %f \n", l2err, grid.h); in computeErr()
|
| /petsc/src/ts/tests/output/ |
| H A D | ex30_0.out | 4 thermal T (K): e= 2.321e+07 i= 1.160e+07 1.160e+07. Normalization grid 0: v_0= 1.876e+07 ( 6.256e-… 5 Domain radius (AMR levels) grid 0: par= 5.000e+00 perp= 5.000e+00 (0) , 1: par= 5.835e-02 perp= 5.8… 28 … 2.5070599309365e-16, energy= 7.2147560145244e+04 (m_i[0]/m_e = 3670.94, 8 cells on electron grid) 41 … 2.5088995200845e-16, energy= 7.2147560145244e+04 (m_i[0]/m_e = 3670.94, 8 cells on electron grid)
|