| /petsc/src/sys/fileio/ |
| H A D | ftest.c | 53 gid_t *gid = NULL; in PetscTestOwnership() local 65 numGroups = getgroups(0, gid); in PetscTestOwnership() 67 PetscCall(PetscMalloc1(numGroups + 1, &gid)); in PetscTestOwnership() 74 gid[0] = getegid(); in PetscTestOwnership() 78 err = getgroups(numGroups, gid + 1); in PetscTestOwnership() 91 if (fgid == gid[g]) { in PetscTestOwnership() 99 PetscCall(PetscFree(gid)); in PetscTestOwnership()
|
| /petsc/src/mat/graphops/coarsen/impls/hem/ |
| H A D | hem.c | 70 node->gid = a_id; in PetscCDGetNewNode() 90 ail->new_node->gid = a_id; in PetscCDGetNewNode() 103 a_this->gid = a_id; in PetscCDIntNdSetID() 112 *a_gid = a_this->gid; in PetscCDIntNdGetID() 211 PetscCall(PetscSynchronizedPrintf(comm, " %" PetscInt_FMT, n->gid)); in PetscCDPrint() 365 PetscInt lsz, ii, kk, *idxs, jj, gid; in PetscCDGetASMBlocks() local 380 PetscCall(PetscCDIntNdGetID(n, &gid)); in PetscCDGetASMBlocks() 381 for (jj = 0; jj < a_bs; lsz++, jj++) idxs[lsz] = a_bs * gid + jj; in PetscCDGetASMBlocks() 501 for (PetscInt kk = 0, gid = Istart; kk < nloc; kk++, gid++) { in MatCoarsenApply_HEM_private() local 504 PetscCall(VecSetValues(vec, 1, &gid, &vval, INSERT_VALUES)); in MatCoarsenApply_HEM_private() [all …]
|
| /petsc/src/mat/graphops/coarsen/impls/mis/ |
| H A D | mis.c | 28 …PetscInt num_fine_ghosts, kk, n, ix, j, *idx, *ii, Iend, my0, nremoved, gid, lid, cpid, l… in MatCoarsenApply_MIS_private() local 60 for (kk = 0, gid = my0; kk < nloc; kk++, gid++) lid_gid[kk] = gid; in MatCoarsenApply_MIS_private() 108 gid = cpcol_gid[cpid]; in MatCoarsenApply_MIS_private() 110 …k(!MIS_IS_SELECTED(statej), PETSC_COMM_SELF, PETSC_ERR_SUP, "selected ghost: %" PetscInt_FMT, gid); in MatCoarsenApply_MIS_private() 111 … if (statej == MIS_NOT_DONE && gid >= Iend) { /* should be (pe>rank), use gid as pe proxy */ in MatCoarsenApply_MIS_private() 224 gid = icpcol_gid[cpid]; in MatCoarsenApply_MIS_private() 227 PetscCall(PetscCDAppendID(agg_lists, slid, gid)); in MatCoarsenApply_MIS_private()
|
| /petsc/src/mat/impls/sell/seq/seqhip/ |
| H A D | sellhip.hip.cxx | 192 PetscInt gid, row, start_slice, cid; in matmult_seqsell_tiled_kernel8() local 197 …gid = gridDim.x * WARP_SIZE * BLOCKY * iter + blockIdx.x * BLOCKY * WARP_SIZE + threadIdx.y * WARP… in matmult_seqsell_tiled_kernel8() 198 if (gid < nrows) y[gid] = 0.0; in matmult_seqsell_tiled_kernel8() 204 gid = cid * BLOCKY * WARP_SIZE + threadIdx.y * WARP_SIZE + threadIdx.x; in matmult_seqsell_tiled_kernel8() 210 while (gid < totalentries && gid >= sliidx[slice_id + 1]) slice_id++; in matmult_seqsell_tiled_kernel8() 213 if (row < nrows && gid < totalentries) t = aval[gid] * x[acolidx[gid]]; in matmult_seqsell_tiled_kernel8() 216 if (row < nrows && gid < totalentries && write) atomAdd(y[row], t); in matmult_seqsell_tiled_kernel8() 220 if (row < nrows) t += aval[gid] * x[acolidx[gid]]; in matmult_seqsell_tiled_kernel8() 247 PetscInt gid, row, start_slice, cid; in matmultadd_seqsell_tiled_kernel8() local 252 …gid = gridDim.x * WARP_SIZE * BLOCKY * iter + blockIdx.x * BLOCKY * WARP_SIZE + threadIdx.y * WARP… in matmultadd_seqsell_tiled_kernel8() [all …]
|
| /petsc/src/mat/impls/sell/seq/seqcuda/ |
| H A D | sellcuda.cu | 188 PetscInt gid, row, start_slice, cid; in matmult_seqsell_tiled_kernel8() local 193 … gid = gridDim.x * 32 * BLOCKY * iter + blockIdx.x * BLOCKY * 32 + threadIdx.y * 32 + threadIdx.x; in matmult_seqsell_tiled_kernel8() 194 if (gid < nrows) y[gid] = 0.0; in matmult_seqsell_tiled_kernel8() 200 gid = cid * BLOCKY * 32 + threadIdx.y * 32 + threadIdx.x; in matmult_seqsell_tiled_kernel8() 206 while (gid < totalentries && gid >= sliidx[slice_id + 1]) slice_id++; in matmult_seqsell_tiled_kernel8() 209 if (row < nrows && gid < totalentries) t = aval[gid] * x[acolidx[gid]]; in matmult_seqsell_tiled_kernel8() 212 if (row < nrows && gid < totalentries && write) atomAdd(y[row], t); in matmult_seqsell_tiled_kernel8() 216 if (row < nrows) t += aval[gid] * x[acolidx[gid]]; in matmult_seqsell_tiled_kernel8() 243 PetscInt gid, row, start_slice, cid; in matmultadd_seqsell_tiled_kernel8() local 248 … gid = gridDim.x * 32 * BLOCKY * iter + blockIdx.x * BLOCKY * 32 + threadIdx.y * 32 + threadIdx.x; in matmultadd_seqsell_tiled_kernel8() [all …]
|
| /petsc/src/mat/graphops/coarsen/impls/misk/ |
| H A D | misk.c | 74 …PetscInt num_fine_ghosts, kk, n, ix, j, *idx, *ai, Iend, my0, nremoved, gid, cpid, lidj, … in MatCoarsenApply_MISK_private() local 97 for (kk = 0, gid = my0; kk < nloc_inner; kk++, gid++) lid_gid[kk] = gid; in MatCoarsenApply_MISK_private() 145 gid = cpcol_gid[cpid]; in MatCoarsenApply_MISK_private() 146 if (cpcol_state[cpid] == MIS_NOT_DONE && gid >= Iend) { /* or pe>rank */ in MatCoarsenApply_MISK_private() 235 gid = icpcol_gid[cpid]; in MatCoarsenApply_MISK_private() 238 PetscCall(PetscCDAppendID(agg_lists, slid, gid)); in MatCoarsenApply_MISK_private()
|
| /petsc/src/ksp/pc/impls/gamg/ |
| H A D | util.c | 47 PetscInt gid = my0 + kk; in PCGAMGGetDataWithGhosts() local 51 PetscCall(VecSetValues(tmp_crds, 1, &gid, &crd, INSERT_VALUES)); in PCGAMGGetDataWithGhosts()
|
| H A D | agg.c | 974 PetscInt gid; in fixAggregatesWithSquare() local 976 PetscCall(PetscCDIntNdGetID(pos, &gid)); in fixAggregatesWithSquare() 977 if (gid == gidj) { in fixAggregatesWithSquare() 1018 PetscInt gid; in fixAggregatesWithSquare() local 1020 PetscCall(PetscCDIntNdGetID(pos, &gid)); in fixAggregatesWithSquare() 1021 if (lid + my0 == gid) { in fixAggregatesWithSquare() 1086 PetscInt gid = (PetscInt)PetscRealPart(cpcol_2_gid[cpid]); in fixAggregatesWithSquare() local 1088 PetscCall(PetscHMapISet(gid_cpid, gid, cpid)); in fixAggregatesWithSquare() 1103 PetscInt gid; in fixAggregatesWithSquare() local 1105 PetscCall(PetscCDIntNdGetID(pos, &gid)); in fixAggregatesWithSquare() [all …]
|
| /petsc/src/dm/partitioner/impls/multistage/ |
| H A D | mspart.c | 629 IS gid, rid; in PetscPartitionerPartition_Multistage() local 639 PetscCall(DMPlexCreateNumbering_Plex(dm, cStart, cEnd, part->height, NULL, sf, &gid)); in PetscPartitionerPartition_Multistage() 641 PetscCall(ISGetIndices(gid, &idxs1)); in PetscPartitionerPartition_Multistage() 642 PetscCall(ISGetLocalSize(gid, &cum)); in PetscPartitionerPartition_Multistage() 648 PetscCall(ISRestoreIndices(gid, &idxs1)); in PetscPartitionerPartition_Multistage() 651 PetscCall(ISDestroy(&gid)); in PetscPartitionerPartition_Multistage() 657 PetscCall(ISCreateGeneral(pcomm, numVertices, idxs2, PETSC_OWN_POINTER, &gid)); in PetscPartitionerPartition_Multistage() 658 PetscCall(ISRenumber(gid, NULL, NULL, &rid)); in PetscPartitionerPartition_Multistage() 659 PetscCall(ISLocalToGlobalMappingCreateIS(gid, &g2l)); in PetscPartitionerPartition_Multistage() 662 PetscCall(ISDestroy(&gid)); in PetscPartitionerPartition_Multistage()
|
| /petsc/src/mat/impls/sbaij/mpi/ |
| H A D | mmsbaij.c | 21 PetscInt gid, lid; in MatSetUpMultiply_MPISBAIJ() local 41 PetscHashIterGetKey(gid1_lid1, tpos, gid); in MatSetUpMultiply_MPISBAIJ() 44 gid--; in MatSetUpMultiply_MPISBAIJ() 46 garray[lid] = gid; in MatSetUpMultiply_MPISBAIJ()
|
| /petsc/src/mat/impls/sell/mpi/ |
| H A D | mmsell.c | 84 PetscInt gid, lid; in MatSetUpMultiply_MPISELL() local 113 PetscHashIterGetKey(gid1_lid1, tpos, gid); in MatSetUpMultiply_MPISELL() 116 gid--; in MatSetUpMultiply_MPISELL() 118 garray[lid] = gid; in MatSetUpMultiply_MPISELL()
|
| /petsc/src/mat/impls/baij/mpi/ |
| H A D | mmbaij.c | 18 PetscInt gid, lid; in MatSetUpMultiply_MPIBAIJ() local 41 PetscHashIterGetKey(gid1_lid1, tpos, gid); in MatSetUpMultiply_MPIBAIJ() 44 gid--; in MatSetUpMultiply_MPIBAIJ() 46 garray[lid] = gid; in MatSetUpMultiply_MPIBAIJ()
|
| /petsc/src/mat/impls/aij/mpi/ |
| H A D | mmaij.c | 19 PetscInt gid, lid; in MatSetUpMultiply_MPIAIJ() local 44 PetscHashIterGetKey(gid1_lid1, tpos, gid); in MatSetUpMultiply_MPIAIJ() 47 gid--; in MatSetUpMultiply_MPIAIJ() 49 garray[lid] = gid; in MatSetUpMultiply_MPIAIJ()
|
| /petsc/include/ |
| H A D | petscmatcoarsen.h | 37 PetscInt gid; member
|
| H A D | petsclandau.h | 206 LandauIdx gid; // Landau matrix index (<10,000) member
|
| /petsc/src/ts/utils/dmplexlandau/ |
| H A D | plexland.c | 268 PetscInt id = maps[grid].c_maps[idx][q].gid; in LandauFormJacobian_Internal() 497 if (maps[grid].c_maps[idx][q].gid < 0) break; in LandauFormJacobian_Internal() 498 rows0[q] = maps[grid].c_maps[idx][q].gid; in LandauFormJacobian_Internal() 512 if (maps[grid].c_maps[idx][q].gid < 0) break; in LandauFormJacobian_Internal() 513 cols0[q] = maps[grid].c_maps[idx][q].gid; in LandauFormJacobian_Internal() 1556 …s[grid].num_reduced][jj].scale == 0) pointMaps[maps[grid].num_reduced][jj].gid = -1; // 3D has Q a… in CreateStaticData() 1559 pointMaps[maps[grid].num_reduced][jj].gid = plex_batch[indices[f]]; in CreateStaticData() 1561 pointMaps[maps[grid].num_reduced][jj].gid = indices[f]; in CreateStaticData() 1568 pointMaps[maps[grid].num_reduced][jj].gid = -1; in CreateStaticData() 1607 maps[grid].c_maps[ej][q].gid = pointMaps[ej][q].gid; in CreateStaticData() [all …]
|
| /petsc/src/dm/impls/plex/ |
| H A D | plexpartition.c | 795 const PetscInt *gid, *clIdx; in PetscPartitionerDMPlexPartition() local 809 if (globalNumbering) PetscCall(ISGetIndices(globalNumbering, &gid)); in PetscPartitionerDMPlexPartition() 810 else gid = NULL; in PetscPartitionerDMPlexPartition() 815 if (gid && gid[p - pStart] < 0) continue; in PetscPartitionerDMPlexPartition() 834 if (globalNumbering) PetscCall(ISRestoreIndices(globalNumbering, &gid)); in PetscPartitionerDMPlexPartition()
|
| H A D | plex.c | 1531 IS gid, acis; in DMPlexView_Ascii() local 1559 PetscCall(DMPlexCreatePartitionerGraph(dm, cellHeight, &numVertices, &start, &adjacency, &gid)); in DMPlexView_Ascii() 1563 PetscCall(ISGetIndices(gid, &idxs)); in DMPlexView_Ascii() 1564 PetscCall(ISGetLocalSize(gid, &cum)); in DMPlexView_Ascii() 1570 PetscCall(ISRestoreIndices(gid, &idxs)); in DMPlexView_Ascii() 1572 PetscCall(ISDestroy(&gid)); in DMPlexView_Ascii() 1573 PetscCall(ISCreateGeneral(comm, numVertices, idxs2, PETSC_OWN_POINTER, &gid)); in DMPlexView_Ascii() 1592 PetscCall(ISLocalToGlobalMappingCreateIS(gid, &g2l)); in DMPlexView_Ascii() 1594 PetscCall(ISDestroy(&gid)); in DMPlexView_Ascii()
|
| /petsc/src/mat/impls/aij/mpi/kokkos/ |
| H A D | mpiaijkok.kokkos.cxx | 886 MatColIdxType *Fdj = Fdj_h.data(), *Foj = Foj_h.data(), gid; in MatMPIAIJKokkosBcastBegin() local 892 gid = Fj[Fi[i] + j]; in MatMPIAIJKokkosBcastBegin() 894 Foj[Foi[i] + j] = gid; in MatMPIAIJKokkosBcastBegin() 896 Fdj[Fdi[i] + j - nzLeft] = gid - cstart; in MatMPIAIJKokkosBcastBegin() 898 Foj[Foi[i] + j - len] = gid; in MatMPIAIJKokkosBcastBegin()
|
| /petsc/src/ts/utils/dmplexlandau/kokkos/ |
| H A D | landau.kokkos.cxx | 357 if (d_maps->c_maps[idx][q].gid < 0) break; in landau_mat_assemble() 369 if (d_maps->c_maps[idx][q].gid < 0) break; in landau_mat_assemble() 542 PetscInt id = maps[grid]->c_maps[idx][q].gid; in LandauKokkosJacobian()
|
| /petsc/src/ksp/pc/impls/bddc/ |
| H A D | bddcgraph.c | 823 PetscInt gid = idxs[i]; in PCBDDCGraphSetUp() local 831 for (j = 0; j < c; j++) graph->nodes[i].local_groups[j] = multi_root_subs[start[gid] + j]; in PCBDDCGraphSetUp()
|
| /petsc/src/mat/impls/aij/seq/ |
| H A D | aij.c | 3567 PetscInt gid, lid, i, ec, nz = aij->nz; in MatSeqAIJCompactOutExtraColumns_SeqAIJ() local 3588 PetscHashIterGetKey(gid1_lid1, tpos, gid); in MatSeqAIJCompactOutExtraColumns_SeqAIJ() 3591 gid--; in MatSeqAIJCompactOutExtraColumns_SeqAIJ() 3593 garray[lid] = gid; in MatSeqAIJCompactOutExtraColumns_SeqAIJ()
|
| /petsc/doc/ |
| H A D | petsc.bib | 5941 url = {http://www.gid.cimne.upc.es/2004/papers/p216.pdf},
|