1 #include <petsc/private/matimpl.h> /*I "petscmat.h" I*/
2 #include "hpl.h"
3
4 /*@
5 MatSetHPL - fills a `MATSEQDENSE` matrix using the HPL 2.3 random matrix generation routine
6
7 Collective
8
9 Input Parameters:
10 + A - the matrix
11 - iseed - the random number seed
12
13 Level: intermediate
14
15 .seealso: [](ch_matrices), `Mat`, `MatCreate()`
16 @*/
MatSetHPL(Mat A,int iseed)17 PetscErrorCode MatSetHPL(Mat A, int iseed)
18 {
19 PetscBool isDense;
20 PetscInt M, N, LDA;
21 PetscBLASInt bM, bN, bLDA;
22 PetscScalar *values;
23
24 PetscFunctionBegin;
25 PetscValidHeaderSpecific(A, MAT_CLASSID, 1);
26 PetscCall(PetscObjectTypeCompare((PetscObject)A, MATSEQDENSE, &isDense));
27 PetscCheck(isDense, PetscObjectComm((PetscObject)A), PETSC_ERR_SUP, "Only supports sequential dense matrices");
28 PetscCall(MatGetSize(A, &M, &N));
29 PetscCall(PetscBLASIntCast(M, &bM));
30 PetscCall(PetscBLASIntCast(N, &bN));
31 PetscCall(MatDenseGetLDA(A, &LDA));
32 PetscCall(PetscBLASIntCast(LDA, &bLDA));
33 PetscCall(MatDenseGetArrayWrite(A, &values));
34 PetscStackCallExternalVoid("HPL_dmatgen", HPL_dmatgen(bM, bN, values, bLDA, iseed));
35 PetscCall(MatDenseRestoreArrayWrite(A, &values));
36 PetscFunctionReturn(PETSC_SUCCESS);
37 }
38
39 #include <petsc/private/bmimpl.h>
40
41 typedef struct {
42 Mat A, F;
43 } PetscBench_HPL;
44
PetscBenchSetUp_HPL(PetscBench bm)45 static PetscErrorCode PetscBenchSetUp_HPL(PetscBench bm)
46 {
47 PetscBench_HPL *hp = (PetscBench_HPL *)bm->data;
48 PetscMPIInt rank;
49
50 PetscFunctionBegin;
51 PetscCallMPI(MPI_Comm_rank(PetscObjectComm((PetscObject)bm), &rank));
52 if (rank > 0) PetscFunctionReturn(PETSC_SUCCESS);
53 if (bm->size == PETSC_DECIDE) bm->size = 2500;
54 PetscCall(MatCreateSeqDense(PETSC_COMM_SELF, bm->size, bm->size, NULL, &hp->A));
55 PetscCall(MatSetHPL(hp->A, 0));
56 PetscCall(MatGetFactor(hp->A, MATSOLVERPETSC, MAT_FACTOR_LU, &hp->F));
57 PetscCall(MatLUFactorSymbolic(hp->F, hp->A, NULL, NULL, 0));
58 PetscFunctionReturn(PETSC_SUCCESS);
59 }
60
PetscBenchRun_HPL(PetscBench bm)61 static PetscErrorCode PetscBenchRun_HPL(PetscBench bm)
62 {
63 PetscBench_HPL *hp = (PetscBench_HPL *)bm->data;
64 PetscMPIInt rank;
65
66 PetscFunctionBegin;
67 PetscCallMPI(MPI_Comm_rank(PetscObjectComm((PetscObject)bm), &rank));
68 if (rank > 0) PetscFunctionReturn(PETSC_SUCCESS);
69 PetscCall(MatLUFactorNumeric(hp->F, hp->A, 0));
70 PetscFunctionReturn(PETSC_SUCCESS);
71 }
72
PetscBenchView_HPL(PetscBench bm,PetscViewer viewer)73 static PetscErrorCode PetscBenchView_HPL(PetscBench bm, PetscViewer viewer)
74 {
75 PetscEventPerfInfo *info;
76 PetscInt numThreads;
77
78 PetscFunctionBegin;
79 PetscCall(PetscLogHandlerGetEventPerfInfo(bm->lhdlr, 0, MAT_LUFactor, &info)); // because symbolic is trivial it does not log numeric!
80 PetscCall(PetscBLASGetNumThreads(&numThreads));
81 PetscCall(PetscViewerASCIIPrintf(viewer, "HPL Benchmark, number of threads %d, matrix size %d, flop rate %g (megaflops)\n", (int)numThreads, (int)bm->size, (double)(info->flops / (1.e7 * info->time))));
82 PetscFunctionReturn(PETSC_SUCCESS);
83 }
84
PetscBenchReset_HPL(PetscBench bm)85 static PetscErrorCode PetscBenchReset_HPL(PetscBench bm)
86 {
87 PetscBench_HPL *hp = (PetscBench_HPL *)bm->data;
88 PetscMPIInt rank;
89
90 PetscFunctionBegin;
91 PetscCallMPI(MPI_Comm_rank(PetscObjectComm((PetscObject)bm), &rank));
92 if (rank > 0) PetscFunctionReturn(PETSC_SUCCESS);
93 PetscCall(MatDestroy(&hp->F));
94 PetscCall(MatDestroy(&hp->A));
95 PetscFunctionReturn(PETSC_SUCCESS);
96 }
97
PetscBenchDestroy_HPL(PetscBench bm)98 static PetscErrorCode PetscBenchDestroy_HPL(PetscBench bm)
99 {
100 PetscFunctionBegin;
101 PetscCall(PetscFree(bm->data));
102 PetscFunctionReturn(PETSC_SUCCESS);
103 }
104
PetscBenchCreate_HPL(PetscBench bm)105 PETSC_INTERN PetscErrorCode PetscBenchCreate_HPL(PetscBench bm)
106 {
107 PetscBench_HPL *hp;
108
109 PetscFunctionBegin;
110 PetscCall(PetscNew(&hp));
111 bm->data = hp;
112 bm->ops->setup = PetscBenchSetUp_HPL;
113 bm->ops->run = PetscBenchRun_HPL;
114 bm->ops->view = PetscBenchView_HPL;
115 bm->ops->reset = PetscBenchReset_HPL;
116 bm->ops->destroy = PetscBenchDestroy_HPL;
117 PetscFunctionReturn(PETSC_SUCCESS);
118 }
119