Home
last modified time | relevance | path

Searched refs:grid (Results 1 – 25 of 275) sorted by relevance

1234567891011

/petsc/src/ts/tests/
H A Dex30.c290 for (PetscInt grid = 0; grid < ctx->num_grids; grid++) { // add same particels for all grids in gridToParticles_private() local
292 …t_FMT ", index(b,g) %" PetscInt_FMT "\n", global_vertex_id, v_id, grid, LAND_PACK_IDX(v_id, grid)); in gridToParticles_private()
293grid], globSwarmArray[LAND_PACK_IDX(v_id, grid)], globXArray[LAND_PACK_IDX(v_id, grid)], t_fhat[gr… in gridToParticles_private()
304 for (PetscInt grid = 0; grid < ctx->num_grids; grid++) { in gridToParticles_private() local
307 DM sw = globSwarmArray[LAND_PACK_IDX(v_id, grid)]; in gridToParticles_private()
313 …PI * coords[p * dim + 0] : 1, w = fact * wp[p] * ctx->n_0 * ctx->masses[ctx->species_offset[grid]]; in gridToParticles_private()
324 for (PetscInt grid = 0; grid < ctx->num_grids; grid++) { in gridToParticles_private() local
327 DM sw = globSwarmArray[LAND_PACK_IDX(v_id, grid)]; in gridToParticles_private()
333 …dim + 0] : 1, w = fact * wp[p] * ctx->n_0 * ctx->masses[ctx->species_offset[grid]], ww = w * N_inv; in gridToParticles_private()
393 for (PetscInt grid = 0; grid < ctx->num_grids; grid++) { in PostStep() local
[all …]
/petsc/src/ts/utils/dmplexlandau/
H A Dplexland.c37 for (PetscInt grid = 0; grid < maps[0].numgrids; grid++) { in LandauGPUMapsDestroy() local
38 PetscCall(PetscFree(maps[grid].c_maps)); in LandauGPUMapsDestroy()
39 PetscCall(PetscFree(maps[grid].gIdx)); in LandauGPUMapsDestroy()
116 …for (PetscInt grid = 0; grid < ctx->num_grids; grid++) PetscCall(DMCreateMatrix(ctx->plex[grid], &… in LandauFormJacobian_Internal() local
127 for (PetscInt grid = 0; grid < ctx->num_grids; grid++) { in LandauFormJacobian_Internal() local
129 PetscCheck(ctx->plex[grid] != NULL, ctx->comm, PETSC_ERR_ARG_WRONG, "Plex not created"); in LandauFormJacobian_Internal()
130 PetscCall(DMPlexGetHeightStratum(ctx->plex[grid], 0, &cStart, &cEnd)); in LandauFormJacobian_Internal()
131 numCells[grid] = cEnd - cStart; // grids can have different topology in LandauFormJacobian_Internal()
148 for (PetscInt grid = 0; grid < ctx->num_grids; grid++) { in LandauFormJacobian_Internal() local
149 PetscCall(DMGetLocalSection(ctx->plex[grid], &section[grid])); in LandauFormJacobian_Internal()
[all …]
/petsc/src/snes/tutorials/
H A Dex30.c104 GridInfo *grid; member
128 GridInfo grid; in main() local
147 PetscCall(SetParams(&param, &grid)); in main()
148 PetscCall(ReportParams(&param, &grid)); in main()
153 …PetscCall(DMDACreate2d(comm, grid.bx, grid.by, grid.stencil, grid.ni, grid.nj, PETSC_DECIDE, PETSC… in main()
167 user->grid = &grid; in main()
309 GridInfo *grid = user->grid; in HorizVelocity() local
313 x = (i - grid->jlid) * grid->dx; in HorizVelocity()
314 z = (j - grid->jlid - 0.5) * grid->dz; in HorizVelocity()
326 GridInfo *grid = user->grid; in VertVelocity() local
[all …]
/petsc/doc/community/
H A Dpetsc_team.md29 .. grid:: 2 3 3 4
31 .. grid-item-card:: Mark Adams
35 .. grid-item-card:: Satish Balay
39 .. grid-item-card:: Jed Brown
43 .. grid-item-card:: Lisandro Dalcin
47 .. grid-item-card:: Toby Isaac
51 .. grid-item-card:: Pierre Jolivet
55 .. grid-item-card:: Matthew Knepley
59 .. grid-item-card:: Lois Curfman McInnes
63 .. grid-item-card:: Richard Tran Mills
[all …]
/petsc/src/ts/utils/dmplexlandau/kokkos/
H A Dlandau.kokkos.cxx93 …Maps maps[], pointInterpolationP4est (*pointMaps)[LANDAU_MAX_Q_FACE], PetscInt Nf[], PetscInt grid) in LandauKokkosCreateMatMaps() argument
96 …its<Kokkos::Unmanaged>> h_points((pointInterpolationP4est *)pointMaps, maps[grid].num_reduced); in LandauKokkosCreateMatMaps()
97 …Kokkos::MemoryTraits<Kokkos::Unmanaged>> h_gidx((LandauIdx *)maps[grid].gIdx, maps[grid].num_eleme… in LandauKokkosCreateMatMaps()
98 …intInterpolationP4est *[LANDAU_MAX_Q_FACE], Kokkos::LayoutRight>("points", maps[grid].num_reduced); in LandauKokkosCreateMatMaps()
99 …uIdx *[LANDAU_MAX_SPECIES][LANDAU_MAX_NQND], Kokkos::LayoutRight>("gIdx", maps[grid].num_elements); in LandauKokkosCreateMatMaps()
104 h_maps.num_elements = maps[grid].num_elements; in LandauKokkosCreateMatMaps()
105 h_maps.num_face = maps[grid].num_face; in LandauKokkosCreateMatMaps()
106 h_maps.num_reduced = maps[grid].num_reduced; in LandauKokkosCreateMatMaps()
107 h_maps.deviceType = maps[grid].deviceType; in LandauKokkosCreateMatMaps()
108 h_maps.numgrids = maps[grid].numgrids; in LandauKokkosCreateMatMaps()
[all …]
/petsc/src/binding/petsc4py/src/petsc4py/lib/_pytypes/viewer/
H A Dpetscpyvista.py105 grid = self.convertDMToPV(dm)
116 if scalars[1].shape[0] / scalars[2] == grid.n_cells:
117 grid.cell_data[scalars[0]] = scalars[1]
118 elif scalars[1].shape[0] / scalars[2] == grid.n_points:
120 grid.point_data[scalars[0]] = scalars[1].reshape(-1, scalars[2])
124 grid.point_data[scalars[0]] = vecs
126 …%d) or vertices (%d)' % (scalars[0], scalars[1].shape[0], scalars[2], grid.n_cells, grid.n_points))
129 grid = grid.warp_by_scalar(factor = self.warpFactor)
131 grid = grid.warp_by_vector(factor = self.warpFactor)
133 grid = grid.clip_box(self.clipBounds)
[all …]
/petsc/src/dm/tutorials/
H A Dex13f90.F9036 PetscReal, allocatable :: f(:, :, :, :), grid(:, :, :, :)
91 allocate (grid(ndim, ib1:ibn, jb1:jbn, kb1:kbn))
100 call petsc_to_local(CoordDM, coords, array, grid, ndim, stw)
101 xmin = grid(1, 1, 1, 1)
102 ymin = grid(2, 1, 1, 1)
103 zmin = grid(3, 1, 1, 1)
104 xmax = grid(1, imax, jmax, kmax)
105 ymax = grid(2, imax, jmax, kmax)
106 zmax = grid(3, imax, jmax, kmax)
107 call local_to_petsc(CoordDM, coords, array, grid, ndim, stw)
[all …]
/petsc/src/mat/impls/scalapack/
H A Dmatscalapack.c43 …rASCIIPrintf(viewer, "grid height=%d, grid width=%d\n", (int)a->grid->nprow, (int)a->grid->npcol)); in MatView_ScaLAPACK()
128 …, SCALAPACKinfog2l_(&gridx, &gcidx, a->desc, &a->grid->nprow, &a->grid->npcol, &a->grid->myrow, &a… in MatSetValues_ScaLAPACK()
129 if (rsrc == a->grid->myrow && csrc == a->grid->mycol) { in MatSetValues_ScaLAPACK()
172 …nit", SCALAPACKdescinit_(xdesc, &a->M, &one, &mb, &one, &zero, &zero, &a->grid->ictxcol, &xlld, &i… in MatMultXXXYYY_ScaLAPACK()
177 …nit", SCALAPACKdescinit_(ydesc, &one, &a->N, &one, &nb, &zero, &zero, &a->grid->ictxrow, &ylld, &i… in MatMultXXXYYY_ScaLAPACK()
181 lszx = SCALAPACKnumroc_(&a->M, &a->mb, &a->grid->myrow, &a->rsrc, &a->grid->nprow); in MatMultXXXYYY_ScaLAPACK()
182 lszy = SCALAPACKnumroc_(&a->N, &a->nb, &a->grid->mycol, &a->csrc, &a->grid->npcol); in MatMultXXXYYY_ScaLAPACK()
187 …, SCALAPACKdescinit_(x2desc, &a->M, &one, &a->mb, &one, &zero, &zero, &a->grid->ictxt, &xlld, &inf… in MatMultXXXYYY_ScaLAPACK()
189 …, SCALAPACKdescinit_(y2desc, &one, &a->N, &one, &a->nb, &zero, &zero, &a->grid->ictxt, &ylld, &inf… in MatMultXXXYYY_ScaLAPACK()
193 …, SCALAPACKgemr2d_(&a->M, &one, x, &one, &one, xdesc, x2d, &one, &one, x2desc, &a->grid->ictxcol)); in MatMultXXXYYY_ScaLAPACK()
[all …]
/petsc/src/tao/unconstrained/tutorials/
H A Dadjointreads.m14 %xg - the grid
49 plot(grid,Init_ts,'ko-','Markersize',10); drawnow;
51 %plot(grid,temp,'c*','Markersize',10); drawnow;
52 plot(grid,exact,'r*','Markersize',10); drawnow;
53 %plot(grid,Curr_sol,'g-','Markersize',10); drawnow;
54 %plot(grid,Init_adj,'bo-','Markersize',10); drawnow;
56 xlabel('x (GLL grid)');
76 % xlabel('x (GLL grid)');
93 % grid on
107 % xlabel('GLL grid');
[all …]
/petsc/src/ts/utils/dmplexlandau/tutorials/
H A Dex1.c23 PetscErrorCode landau_field_print_access_callback(DM dm, Vec x, PetscInt local_field, PetscInt grid in landau_field_print_access_callback() argument
31 species = ctx->species_offset[grid] + local_field; in landau_field_print_access_callback()
32 val = (PetscScalar)(LAND_PACK_IDX(b_id, grid) + (species + 1) * 10); in landau_field_print_access_callback()
34 …_FMT " and local field %" PetscInt_FMT " with %" PetscInt_FMT " grids\n", grid, b_id, local_field,… in landau_field_print_access_callback()
93 …DM dm, Vec X, PetscReal time, PetscReal temps[], PetscReal ns[], PetscInt grid, PetscReal shifts[]… in SetMaxwellians() argument
100 …for (PetscInt ii = ctx->species_offset[grid], i0 = 0; ii < ctx->species_offset[grid + 1]; ii++, i0… in SetMaxwellians()
110 …ign(ctx->charges[ctx->species_offset[grid]]) * ctx->electronShift * ctx->m_0 / ctx->masses[ctx->sp… in SetMaxwellians()
114 …data[0].shift = ctx->electronShift * shifts[grid] * PetscSqrtReal(data[0].kT_m) / ctx->v_0; // shi… in SetMaxwellians()
256 for (PetscInt grid = 0; grid < ctx->num_grids; grid++) { in Monitor() local
258 DM dm = ctx->plex[grid]; in Monitor()
[all …]
/petsc/include/petsc/private/
H A Dpetscscalapack.h15 Mat_ScaLAPACK_Grid *grid; /* process grid */ member
121 …rc == _bb->csrc && _aa->grid->nprow == _bb->grid->nprow && _aa->grid->npcol == _bb->grid->npcol &&…
122 _aa->grid->mycol == _bb->grid->mycol, \
/petsc/src/ts/utils/dmplexlandau/tutorials/output/
H A Dex2_simplex.out4 thermal T (K): e= 2.321e+07 i= 1.160e+07 0.000e+00. Normalization grid 0: v_0= 1.876e+07 ( 6.256e-…
5 Domain radius (AMR levels) grid 0: par= 5.000e+00 perp= 5.000e+00 (0) , 1: par= 5.835e-02 perp= 5.8…
20 …2.3596368842425e-18, energy= 7.1246540087660e+04 (m_i[0]/m_e = 3670.94, 48 cells on electron grid)
35 …9.7382234546064e-18, energy= 7.1246540087660e+04 (m_i[0]/m_e = 3670.94, 48 cells on electron grid)
38 …2.3596368842425e-18, energy= 7.1246540087660e+04 (m_i[0]/m_e = 3670.94, 48 cells on electron grid)
53 …9.7382234546064e-18, energy= 7.1246540087660e+04 (m_i[0]/m_e = 3670.94, 48 cells on electron grid)
H A Dex1_sphere.out4 thermal T (K): e= 2.321e+07 i= 1.160e+07 0.000e+00. Normalization grid 0: v_0= 1.876e+07 ( 6.256e-…
5 Domain radius (AMR levels) grid 0: par= 5.000e+00 perp= 5.000e+00 (1) , 1: par= 5.835e-02 perp= 5.8…
10 …5.6143870238028e-19, energy= 7.3772717719457e+04 (m_i[0]/m_e = 3670.94, 12 cells on electron grid)
15 …4.1036480213379e-18, energy= 7.3772717719441e+04 (m_i[0]/m_e = 3670.94, 12 cells on electron grid)
H A Dex2_0.out4 thermal T (K): e= 5.802e+07 i= 5.802e+07 0.000e+00. Normalization grid 0: v_0= 2.965e+07 ( 9.892e-…
5 Domain radius (AMR levels) grid 0: par= 5.000e+00 perp= 5.000e+00 (2) , 1: par= 8.252e-02 perp= 8.2…
10 …5.1936165265558e-18, energy= 2.4015413557925e+05 (m_i[0]/m_e = 3670.94, 14 cells on electron grid)
33 …1.1060501186683e-19, energy= 2.4015413557925e+05 (m_i[0]/m_e = 3670.94, 14 cells on electron grid)
37 …5.1936165265558e-18, energy= 2.4015413557925e+05 (m_i[0]/m_e = 3670.94, 14 cells on electron grid)
80 …4.5045672869187e-18, energy= 2.4015413557925e+05 (m_i[0]/m_e = 3670.94, 14 cells on electron grid)
H A Dex1_0.out4 thermal T (K): e= 3.481e+06 i= 2.321e+06 0.000e+00. Normalization grid 0: v_0= 7.264e+06 ( 2.423e-…
5 Domain radius (AMR levels) grid 0: par= 6.000e+00 perp= 6.000e+00 (0) , 1: par= 8.086e-02 perp= 8.0…
23 …1.1333771884393e-07, energy= 1.9821738917588e+04 (m_i[0]/m_e = 3670.94, 26 cells on electron grid)
32 …1.1333771884393e-07, energy= 1.9821738917588e+04 (m_i[0]/m_e = 3670.94, 26 cells on electron grid)
H A Dex1f90_0.out4 thermal T (K): e= 5.802e+07 i= 5.802e+07 5.802e+06. Normalization grid 1: v_0= 4.894e+05 ( 1.633e-…
5 Domain radius (AMR levels) grid 0: par= 3.029e+02 perp= 3.029e+02 (2) , 1: par= 5.000e+00 perp= 5.0…
20 …1.1601092271659e-17, energy= 2.3762108515975e+05 (m_i[0]/m_e = 3670.94, 14 cells on electron grid)
48 …4.9547999880240e-22, energy= 2.3762108515975e+05 (m_i[0]/m_e = 3670.94, 14 cells on electron grid)
/petsc/src/benchmarks/
H A DbenchmarkAssembly.py112 grid = [10]*len(procs) variable
114 grid = [1250]*len(procs) variable
117 grid = [100, 150, 200, 250, 300] variable
119 grid = range(150, 1350, 100) variable
120 procs = [1]*len(grid)
121 for n, p in zip(grid, procs):
/petsc/src/ksp/ksp/tutorials/output/
H A Dex56_latebs-2.out25 Using externally compute Galerkin coarse grid matrices
29 Using parallel coarse grid solver (all coarse grid equations not put on one process)
36 Complexity: grid = 1.054 operator = 1.07125
41 Coarse grid solver -- level 0 -------------------------------
113 Using externally compute Galerkin coarse grid matrices
117 Using parallel coarse grid solver (all coarse grid equations not put on one process)
124 Complexity: grid = 1.054 operator = 1.07125
129 Coarse grid solver -- level 0 -------------------------------
211 Using externally compute Galerkin coarse grid matrices
215 Using parallel coarse grid solver (all coarse grid equations not put on one process)
[all …]
/petsc/src/ksp/ksp/tests/
H A Dex29.c118 PetscErrorCode FormJacobian_Grid(GridCtx *grid, Mat *J) in FormJacobian_Grid() argument
127 mx = grid->mx; in FormJacobian_Grid()
128 my = grid->my; in FormJacobian_Grid()
135 PetscCall(DMDAGetCorners(grid->da, &xs, &ys, 0, &xm, &ym, 0)); in FormJacobian_Grid()
136 PetscCall(DMDAGetGhostCorners(grid->da, &Xs, &Ys, 0, &Xm, &Ym, 0)); in FormJacobian_Grid()
137 PetscCall(DMGetLocalToGlobalMapping(grid->da, &ltogm)); in FormJacobian_Grid()
H A Dex26.c148 PetscErrorCode FormJacobian_Grid(GridCtx *grid, Mat jac) in FormJacobian_Grid() argument
157 mx = grid->mx; in FormJacobian_Grid()
158 my = grid->my; in FormJacobian_Grid()
165 PetscCall(DMDAGetCorners(grid->da, &xs, &ys, 0, &xm, &ym, 0)); in FormJacobian_Grid()
166 PetscCall(DMDAGetGhostCorners(grid->da, &Xs, &Ys, 0, &Xm, &Ym, 0)); in FormJacobian_Grid()
167 PetscCall(DMGetLocalToGlobalMapping(grid->da, &ltogm)); in FormJacobian_Grid()
/petsc/src/mat/impls/aij/mpi/superlu_dist/
H A Dsuperlu_dist.c94 gridinfo_t grid; member
126 …lVoid("SuperLU_DIST:pGetDiagU", pGetDiagU(F->rmap->N, &lu->LUstruct, &lu->grid, CASTDOUBLECOMPLEX … in MatSuperluDistGetDiagU_SuperLU_DIST()
142 gridinfo_t grid; member
162 PetscStackCallExternalVoid("SuperLU_DIST:superlu_gridexit", superlu_gridexit(&context->grid)); in Petsc_Superlu_dist_keyval_DeleteFn()
225 …PetscStackCallExternalVoid("SuperLU_DIST:Destroy_LU", sDestroy_LU(A->cmap->N, &lu->grid, &lu->sLUs… in MatDestroy_SuperLU_DIST()
231 …PetscStackCallExternalVoid("SuperLU_DIST:Destroy_LU", Destroy_LU(A->cmap->N, &lu->grid, &lu->LUstr… in MatDestroy_SuperLU_DIST()
242 PetscStackCallExternalVoid("SuperLU_DIST:superlu_gridexit", superlu_gridexit(&lu->grid)); in MatDestroy_SuperLU_DIST()
323 …u->options, &lu->A_sup, &lu->sScalePermstruct, lu->sbptr, (int)m, 1, &lu->grid, &lu->sLUstruct, &l… in MatSolve_SuperLU_DIST()
326 …&lu->A_sup, &lu->ScalePermstruct, CASTDOUBLECOMPLEX bptr, (int)m, 1, &lu->grid, &lu->LUstruct, &lu… in MatSolve_SuperLU_DIST()
329 …ernalVoid("SuperLU_DIST:PStatPrint", PStatPrint(&lu->options, &stat, &lu->grid)); /* Print the sta… in MatSolve_SuperLU_DIST()
[all …]
/petsc/src/ksp/ksp/tests/output/
H A Dex19_2.out1 Coarse grid size 5 by 5
2 Fine grid size 9 by 9
H A Dex19_1.out1 Coarse grid size 5 by 5
2 Fine grid size 9 by 9
/petsc/src/ksp/ksp/tests/raja/
H A Dex1.raja.cxx93 void computeErr(double *I, grid_s grid);
410 void computeErr(double *I, grid_s grid) in computeErr() argument
412 RAJA::RangeSegment gridRange(0, grid.n); in computeErr()
418 int id = tx + grid.n * ty; in computeErr()
419 double x = grid.o + tx * grid.h; in computeErr()
420 double y = grid.o + ty * grid.h; in computeErr()
426 printf("Max error = %lg, h = %f \n", l2err, grid.h); in computeErr()
/petsc/src/ts/tests/output/
H A Dex30_0.out4 thermal T (K): e= 2.321e+07 i= 1.160e+07 1.160e+07. Normalization grid 0: v_0= 1.876e+07 ( 6.256e-…
5 Domain radius (AMR levels) grid 0: par= 5.000e+00 perp= 5.000e+00 (0) , 1: par= 5.835e-02 perp= 5.8…
28 … 2.5070599309365e-16, energy= 7.2147560145244e+04 (m_i[0]/m_e = 3670.94, 8 cells on electron grid)
41 … 2.5088995200845e-16, energy= 7.2147560145244e+04 (m_i[0]/m_e = 3670.94, 8 cells on electron grid)

1234567891011