1 #ifdef PETSC_RCS_HEADER 2 static char vcid[] = "$Id: mpidense.c,v 1.102 1999/01/27 19:47:10 bsmith Exp curfman $"; 3 #endif 4 5 /* 6 Basic functions for basic parallel dense matrices. 7 */ 8 9 #include "src/mat/impls/dense/mpi/mpidense.h" 10 #include "src/vec/vecimpl.h" 11 12 #undef __FUNC__ 13 #define __FUNC__ "MatSetValues_MPIDense" 14 int MatSetValues_MPIDense(Mat mat,int m,int *idxm,int n,int *idxn,Scalar *v,InsertMode addv) 15 { 16 Mat_MPIDense *A = (Mat_MPIDense *) mat->data; 17 int ierr, i, j, rstart = A->rstart, rend = A->rend, row; 18 int roworiented = A->roworiented; 19 20 PetscFunctionBegin; 21 for ( i=0; i<m; i++ ) { 22 if (idxm[i] < 0) continue; 23 if (idxm[i] >= A->M) SETERRQ(PETSC_ERR_ARG_OUTOFRANGE,0,"Row too large"); 24 if (idxm[i] >= rstart && idxm[i] < rend) { 25 row = idxm[i] - rstart; 26 if (roworiented) { 27 ierr = MatSetValues(A->A,1,&row,n,idxn,v+i*n,addv); CHKERRQ(ierr); 28 } else { 29 for ( j=0; j<n; j++ ) { 30 if (idxn[j] < 0) continue; 31 if (idxn[j] >= A->N) SETERRQ(PETSC_ERR_ARG_OUTOFRANGE,0,"Column too large"); 32 ierr = MatSetValues(A->A,1,&row,1,&idxn[j],v+i+j*m,addv); CHKERRQ(ierr); 33 } 34 } 35 } else { 36 if (roworiented) { 37 ierr = StashValues_Private(&A->stash,idxm[i],n,idxn,v+i*n,addv); CHKERRQ(ierr); 38 } else { /* must stash each seperately */ 39 row = idxm[i]; 40 for ( j=0; j<n; j++ ) { 41 ierr = StashValues_Private(&A->stash,row,1,&idxn[j],v+i+j*m,addv);CHKERRQ(ierr); 42 } 43 } 44 } 45 } 46 PetscFunctionReturn(0); 47 } 48 49 #undef __FUNC__ 50 #define __FUNC__ "MatGetValues_MPIDense" 51 int MatGetValues_MPIDense(Mat mat,int m,int *idxm,int n,int *idxn,Scalar *v) 52 { 53 Mat_MPIDense *mdn = (Mat_MPIDense *) mat->data; 54 int ierr, i, j, rstart = mdn->rstart, rend = mdn->rend, row; 55 56 PetscFunctionBegin; 57 for ( i=0; i<m; i++ ) { 58 if (idxm[i] < 0) SETERRQ(PETSC_ERR_ARG_OUTOFRANGE,0,"Negative row"); 59 if (idxm[i] >= mdn->M) SETERRQ(PETSC_ERR_ARG_OUTOFRANGE,0,"Row too large"); 60 if (idxm[i] >= rstart && idxm[i] < rend) { 61 row = idxm[i] - rstart; 62 for ( j=0; j<n; j++ ) { 63 if (idxn[j] < 0) SETERRQ(PETSC_ERR_ARG_OUTOFRANGE,0,"Negative column"); 64 if (idxn[j] >= mdn->N) { 65 SETERRQ(PETSC_ERR_ARG_OUTOFRANGE,0,"Column too large"); 66 } 67 ierr = MatGetValues(mdn->A,1,&row,1,&idxn[j],v+i*n+j); CHKERRQ(ierr); 68 } 69 } else { 70 SETERRQ(PETSC_ERR_SUP,0,"Only local values currently supported"); 71 } 72 } 73 PetscFunctionReturn(0); 74 } 75 76 #undef __FUNC__ 77 #define __FUNC__ "MatGetArray_MPIDense" 78 int MatGetArray_MPIDense(Mat A,Scalar **array) 79 { 80 Mat_MPIDense *a = (Mat_MPIDense *) A->data; 81 int ierr; 82 83 PetscFunctionBegin; 84 ierr = MatGetArray(a->A,array); CHKERRQ(ierr); 85 PetscFunctionReturn(0); 86 } 87 88 #undef __FUNC__ 89 #define __FUNC__ "MatRestoreArray_MPIDense" 90 int MatRestoreArray_MPIDense(Mat A,Scalar **array) 91 { 92 PetscFunctionBegin; 93 PetscFunctionReturn(0); 94 } 95 96 #undef __FUNC__ 97 #define __FUNC__ "MatAssemblyBegin_MPIDense" 98 int MatAssemblyBegin_MPIDense(Mat mat,MatAssemblyType mode) 99 { 100 Mat_MPIDense *mdn = (Mat_MPIDense *) mat->data; 101 MPI_Comm comm = mat->comm; 102 int size = mdn->size, *owners = mdn->rowners, rank = mdn->rank; 103 int *nprocs,i,j,idx,*procs,nsends,nreceives,nmax,*work; 104 int tag = mat->tag, *owner,*starts,count,ierr; 105 InsertMode addv; 106 MPI_Request *send_waits,*recv_waits; 107 Scalar *rvalues,*svalues; 108 109 PetscFunctionBegin; 110 /* make sure all processors are either in INSERTMODE or ADDMODE */ 111 ierr = MPI_Allreduce(&mat->insertmode,&addv,1,MPI_INT,MPI_BOR,comm);CHKERRQ(ierr); 112 if (addv == (ADD_VALUES|INSERT_VALUES)) { 113 SETERRQ(PETSC_ERR_ARG_WRONGSTATE,0,"Cannot mix adds/inserts on different procs"); 114 } 115 mat->insertmode = addv; /* in case this processor had no cache */ 116 117 /* first count number of contributors to each processor */ 118 nprocs = (int *) PetscMalloc( 2*size*sizeof(int) ); CHKPTRQ(nprocs); 119 PetscMemzero(nprocs,2*size*sizeof(int)); procs = nprocs + size; 120 owner = (int *) PetscMalloc( (mdn->stash.n+1)*sizeof(int) ); CHKPTRQ(owner); 121 for ( i=0; i<mdn->stash.n; i++ ) { 122 idx = mdn->stash.idx[i]; 123 for ( j=0; j<size; j++ ) { 124 if (idx >= owners[j] && idx < owners[j+1]) { 125 nprocs[j]++; procs[j] = 1; owner[i] = j; break; 126 } 127 } 128 } 129 nsends = 0; for ( i=0; i<size; i++ ) { nsends += procs[i];} 130 131 /* inform other processors of number of messages and max length*/ 132 work = (int *) PetscMalloc( size*sizeof(int) ); CHKPTRQ(work); 133 ierr = MPI_Allreduce(procs,work,size,MPI_INT,MPI_SUM,comm);CHKERRQ(ierr); 134 nreceives = work[rank]; 135 if (nreceives > size) SETERRQ(PETSC_ERR_PLIB,0,"Internal PETSc error"); 136 ierr = MPI_Allreduce(nprocs,work,size,MPI_INT,MPI_MAX,comm);CHKERRQ(ierr); 137 nmax = work[rank]; 138 PetscFree(work); 139 140 /* post receives: 141 1) each message will consist of ordered pairs 142 (global index,value) we store the global index as a double 143 to simplify the message passing. 144 2) since we don't know how long each individual message is we 145 allocate the largest needed buffer for each receive. Potentially 146 this is a lot of wasted space. 147 148 This could be done better. 149 */ 150 rvalues = (Scalar *) PetscMalloc(3*(nreceives+1)*(nmax+1)*sizeof(Scalar));CHKPTRQ(rvalues); 151 recv_waits = (MPI_Request *) PetscMalloc((nreceives+1)*sizeof(MPI_Request));CHKPTRQ(recv_waits); 152 for ( i=0; i<nreceives; i++ ) { 153 ierr = MPI_Irecv(rvalues+3*nmax*i,3*nmax,MPIU_SCALAR,MPI_ANY_SOURCE,tag,comm,recv_waits+i);CHKERRQ(ierr); 154 } 155 156 /* do sends: 157 1) starts[i] gives the starting index in svalues for stuff going to 158 the ith processor 159 */ 160 svalues = (Scalar *) PetscMalloc( 3*(mdn->stash.n+1)*sizeof(Scalar));CHKPTRQ(svalues); 161 send_waits = (MPI_Request *) PetscMalloc((nsends+1)*sizeof(MPI_Request));CHKPTRQ(send_waits); 162 starts = (int *) PetscMalloc( size*sizeof(int) ); CHKPTRQ(starts); 163 starts[0] = 0; 164 for ( i=1; i<size; i++ ) { starts[i] = starts[i-1] + nprocs[i-1];} 165 for ( i=0; i<mdn->stash.n; i++ ) { 166 svalues[3*starts[owner[i]]] = (Scalar) mdn->stash.idx[i]; 167 svalues[3*starts[owner[i]]+1] = (Scalar) mdn->stash.idy[i]; 168 svalues[3*(starts[owner[i]]++)+2] = mdn->stash.array[i]; 169 } 170 PetscFree(owner); 171 starts[0] = 0; 172 for ( i=1; i<size; i++ ) { starts[i] = starts[i-1] + nprocs[i-1];} 173 count = 0; 174 for ( i=0; i<size; i++ ) { 175 if (procs[i]) { 176 ierr = MPI_Isend(svalues+3*starts[i],3*nprocs[i],MPIU_SCALAR,i,tag,comm,send_waits+count++);CHKERRQ(ierr); 177 } 178 } 179 PetscFree(starts); PetscFree(nprocs); 180 181 /* Free cache space */ 182 PLogInfo(mat,"MatAssemblyBegin_MPIDense:Number of off-processor values %d\n",mdn->stash.n); 183 ierr = StashDestroy_Private(&mdn->stash); CHKERRQ(ierr); 184 185 mdn->svalues = svalues; mdn->rvalues = rvalues; 186 mdn->nsends = nsends; mdn->nrecvs = nreceives; 187 mdn->send_waits = send_waits; mdn->recv_waits = recv_waits; 188 mdn->rmax = nmax; 189 190 PetscFunctionReturn(0); 191 } 192 extern int MatSetUpMultiply_MPIDense(Mat); 193 194 #undef __FUNC__ 195 #define __FUNC__ "MatAssemblyEnd_MPIDense" 196 int MatAssemblyEnd_MPIDense(Mat mat,MatAssemblyType mode) 197 { 198 Mat_MPIDense *mdn = (Mat_MPIDense *) mat->data; 199 MPI_Status *send_status,recv_status; 200 int imdex, nrecvs=mdn->nrecvs, count=nrecvs, i, n, ierr, row, col; 201 Scalar *values,val; 202 InsertMode addv = mat->insertmode; 203 204 PetscFunctionBegin; 205 /* wait on receives */ 206 while (count) { 207 ierr = MPI_Waitany(nrecvs,mdn->recv_waits,&imdex,&recv_status);CHKERRQ(ierr); 208 /* unpack receives into our local space */ 209 values = mdn->rvalues + 3*imdex*mdn->rmax; 210 ierr = MPI_Get_count(&recv_status,MPIU_SCALAR,&n);CHKERRQ(ierr); 211 n = n/3; 212 for ( i=0; i<n; i++ ) { 213 row = (int) PetscReal(values[3*i]) - mdn->rstart; 214 col = (int) PetscReal(values[3*i+1]); 215 val = values[3*i+2]; 216 if (col >= 0 && col < mdn->N) { 217 MatSetValues(mdn->A,1,&row,1,&col,&val,addv); 218 } 219 else {SETERRQ(PETSC_ERR_ARG_OUTOFRANGE,0,"Invalid column");} 220 } 221 count--; 222 } 223 PetscFree(mdn->recv_waits); PetscFree(mdn->rvalues); 224 225 /* wait on sends */ 226 if (mdn->nsends) { 227 send_status = (MPI_Status *) PetscMalloc(mdn->nsends*sizeof(MPI_Status));CHKPTRQ(send_status); 228 ierr = MPI_Waitall(mdn->nsends,mdn->send_waits,send_status);CHKERRQ(ierr); 229 PetscFree(send_status); 230 } 231 PetscFree(mdn->send_waits); PetscFree(mdn->svalues); 232 233 ierr = MatAssemblyBegin(mdn->A,mode); CHKERRQ(ierr); 234 ierr = MatAssemblyEnd(mdn->A,mode); CHKERRQ(ierr); 235 236 if (!mat->was_assembled && mode == MAT_FINAL_ASSEMBLY) { 237 ierr = MatSetUpMultiply_MPIDense(mat); CHKERRQ(ierr); 238 } 239 PetscFunctionReturn(0); 240 } 241 242 #undef __FUNC__ 243 #define __FUNC__ "MatZeroEntries_MPIDense" 244 int MatZeroEntries_MPIDense(Mat A) 245 { 246 int ierr; 247 Mat_MPIDense *l = (Mat_MPIDense *) A->data; 248 249 PetscFunctionBegin; 250 ierr = MatZeroEntries(l->A);CHKERRQ(ierr); 251 PetscFunctionReturn(0); 252 } 253 254 #undef __FUNC__ 255 #define __FUNC__ "MatGetBlockSize_MPIDense" 256 int MatGetBlockSize_MPIDense(Mat A,int *bs) 257 { 258 PetscFunctionBegin; 259 *bs = 1; 260 PetscFunctionReturn(0); 261 } 262 263 /* the code does not do the diagonal entries correctly unless the 264 matrix is square and the column and row owerships are identical. 265 This is a BUG. The only way to fix it seems to be to access 266 mdn->A and mdn->B directly and not through the MatZeroRows() 267 routine. 268 */ 269 #undef __FUNC__ 270 #define __FUNC__ "MatZeroRows_MPIDense" 271 int MatZeroRows_MPIDense(Mat A,IS is,Scalar *diag) 272 { 273 Mat_MPIDense *l = (Mat_MPIDense *) A->data; 274 int i,ierr,N, *rows,*owners = l->rowners,size = l->size; 275 int *procs,*nprocs,j,found,idx,nsends,*work; 276 int nmax,*svalues,*starts,*owner,nrecvs,rank = l->rank; 277 int *rvalues,tag = A->tag,count,base,slen,n,*source; 278 int *lens,imdex,*lrows,*values; 279 MPI_Comm comm = A->comm; 280 MPI_Request *send_waits,*recv_waits; 281 MPI_Status recv_status,*send_status; 282 IS istmp; 283 284 PetscFunctionBegin; 285 ierr = ISGetSize(is,&N); CHKERRQ(ierr); 286 ierr = ISGetIndices(is,&rows); CHKERRQ(ierr); 287 288 /* first count number of contributors to each processor */ 289 nprocs = (int *) PetscMalloc( 2*size*sizeof(int) ); CHKPTRQ(nprocs); 290 PetscMemzero(nprocs,2*size*sizeof(int)); procs = nprocs + size; 291 owner = (int *) PetscMalloc((N+1)*sizeof(int)); CHKPTRQ(owner); /* see note*/ 292 for ( i=0; i<N; i++ ) { 293 idx = rows[i]; 294 found = 0; 295 for ( j=0; j<size; j++ ) { 296 if (idx >= owners[j] && idx < owners[j+1]) { 297 nprocs[j]++; procs[j] = 1; owner[i] = j; found = 1; break; 298 } 299 } 300 if (!found) SETERRQ(PETSC_ERR_ARG_OUTOFRANGE,0,"Index out of range"); 301 } 302 nsends = 0; for ( i=0; i<size; i++ ) { nsends += procs[i];} 303 304 /* inform other processors of number of messages and max length*/ 305 work = (int *) PetscMalloc( size*sizeof(int) ); CHKPTRQ(work); 306 ierr = MPI_Allreduce( procs, work,size,MPI_INT,MPI_SUM,comm);CHKERRQ(ierr); 307 nrecvs = work[rank]; 308 ierr = MPI_Allreduce( nprocs, work,size,MPI_INT,MPI_MAX,comm);CHKERRQ(ierr); 309 nmax = work[rank]; 310 PetscFree(work); 311 312 /* post receives: */ 313 rvalues = (int *) PetscMalloc((nrecvs+1)*(nmax+1)*sizeof(int));CHKPTRQ(rvalues); 314 recv_waits = (MPI_Request *) PetscMalloc((nrecvs+1)*sizeof(MPI_Request));CHKPTRQ(recv_waits); 315 for ( i=0; i<nrecvs; i++ ) { 316 ierr = MPI_Irecv(rvalues+nmax*i,nmax,MPI_INT,MPI_ANY_SOURCE,tag,comm,recv_waits+i);CHKERRQ(ierr); 317 } 318 319 /* do sends: 320 1) starts[i] gives the starting index in svalues for stuff going to 321 the ith processor 322 */ 323 svalues = (int *) PetscMalloc( (N+1)*sizeof(int) ); CHKPTRQ(svalues); 324 send_waits = (MPI_Request *) PetscMalloc((nsends+1)*sizeof(MPI_Request));CHKPTRQ(send_waits); 325 starts = (int *) PetscMalloc( (size+1)*sizeof(int) ); CHKPTRQ(starts); 326 starts[0] = 0; 327 for ( i=1; i<size; i++ ) { starts[i] = starts[i-1] + nprocs[i-1];} 328 for ( i=0; i<N; i++ ) { 329 svalues[starts[owner[i]]++] = rows[i]; 330 } 331 ISRestoreIndices(is,&rows); 332 333 starts[0] = 0; 334 for ( i=1; i<size+1; i++ ) { starts[i] = starts[i-1] + nprocs[i-1];} 335 count = 0; 336 for ( i=0; i<size; i++ ) { 337 if (procs[i]) { 338 ierr = MPI_Isend(svalues+starts[i],nprocs[i],MPI_INT,i,tag,comm,send_waits+count++);CHKERRQ(ierr); 339 } 340 } 341 PetscFree(starts); 342 343 base = owners[rank]; 344 345 /* wait on receives */ 346 lens = (int *) PetscMalloc( 2*(nrecvs+1)*sizeof(int) ); CHKPTRQ(lens); 347 source = lens + nrecvs; 348 count = nrecvs; slen = 0; 349 while (count) { 350 ierr = MPI_Waitany(nrecvs,recv_waits,&imdex,&recv_status);CHKERRQ(ierr); 351 /* unpack receives into our local space */ 352 ierr = MPI_Get_count(&recv_status,MPI_INT,&n);CHKERRQ(ierr); 353 source[imdex] = recv_status.MPI_SOURCE; 354 lens[imdex] = n; 355 slen += n; 356 count--; 357 } 358 PetscFree(recv_waits); 359 360 /* move the data into the send scatter */ 361 lrows = (int *) PetscMalloc( (slen+1)*sizeof(int) ); CHKPTRQ(lrows); 362 count = 0; 363 for ( i=0; i<nrecvs; i++ ) { 364 values = rvalues + i*nmax; 365 for ( j=0; j<lens[i]; j++ ) { 366 lrows[count++] = values[j] - base; 367 } 368 } 369 PetscFree(rvalues); PetscFree(lens); 370 PetscFree(owner); PetscFree(nprocs); 371 372 /* actually zap the local rows */ 373 ierr = ISCreateGeneral(PETSC_COMM_SELF,slen,lrows,&istmp);CHKERRQ(ierr); 374 PLogObjectParent(A,istmp); 375 PetscFree(lrows); 376 ierr = MatZeroRows(l->A,istmp,diag); CHKERRQ(ierr); 377 ierr = ISDestroy(istmp); CHKERRQ(ierr); 378 379 /* wait on sends */ 380 if (nsends) { 381 send_status = (MPI_Status *) PetscMalloc(nsends*sizeof(MPI_Status));CHKPTRQ(send_status); 382 ierr = MPI_Waitall(nsends,send_waits,send_status);CHKERRQ(ierr); 383 PetscFree(send_status); 384 } 385 PetscFree(send_waits); PetscFree(svalues); 386 387 PetscFunctionReturn(0); 388 } 389 390 #undef __FUNC__ 391 #define __FUNC__ "MatMult_MPIDense" 392 int MatMult_MPIDense(Mat mat,Vec xx,Vec yy) 393 { 394 Mat_MPIDense *mdn = (Mat_MPIDense *) mat->data; 395 int ierr; 396 397 PetscFunctionBegin; 398 ierr = VecScatterBegin(xx,mdn->lvec,INSERT_VALUES,SCATTER_FORWARD,mdn->Mvctx);CHKERRQ(ierr); 399 ierr = VecScatterEnd(xx,mdn->lvec,INSERT_VALUES,SCATTER_FORWARD,mdn->Mvctx);CHKERRQ(ierr); 400 ierr = MatMult_SeqDense(mdn->A,mdn->lvec,yy); CHKERRQ(ierr); 401 PetscFunctionReturn(0); 402 } 403 404 #undef __FUNC__ 405 #define __FUNC__ "MatMultAdd_MPIDense" 406 int MatMultAdd_MPIDense(Mat mat,Vec xx,Vec yy,Vec zz) 407 { 408 Mat_MPIDense *mdn = (Mat_MPIDense *) mat->data; 409 int ierr; 410 411 PetscFunctionBegin; 412 ierr = VecScatterBegin(xx,mdn->lvec,INSERT_VALUES,SCATTER_FORWARD,mdn->Mvctx);CHKERRQ(ierr); 413 ierr = VecScatterEnd(xx,mdn->lvec,INSERT_VALUES,SCATTER_FORWARD,mdn->Mvctx);CHKERRQ(ierr); 414 ierr = MatMultAdd_SeqDense(mdn->A,mdn->lvec,yy,zz); CHKERRQ(ierr); 415 PetscFunctionReturn(0); 416 } 417 418 #undef __FUNC__ 419 #define __FUNC__ "MatMultTrans_MPIDense" 420 int MatMultTrans_MPIDense(Mat A,Vec xx,Vec yy) 421 { 422 Mat_MPIDense *a = (Mat_MPIDense *) A->data; 423 int ierr; 424 Scalar zero = 0.0; 425 426 PetscFunctionBegin; 427 ierr = VecSet(&zero,yy); CHKERRQ(ierr); 428 ierr = MatMultTrans_SeqDense(a->A,xx,a->lvec); CHKERRQ(ierr); 429 ierr = VecScatterBegin(a->lvec,yy,ADD_VALUES,SCATTER_REVERSE,a->Mvctx); CHKERRQ(ierr); 430 ierr = VecScatterEnd(a->lvec,yy,ADD_VALUES,SCATTER_REVERSE,a->Mvctx); CHKERRQ(ierr); 431 PetscFunctionReturn(0); 432 } 433 434 #undef __FUNC__ 435 #define __FUNC__ "MatMultTransAdd_MPIDense" 436 int MatMultTransAdd_MPIDense(Mat A,Vec xx,Vec yy,Vec zz) 437 { 438 Mat_MPIDense *a = (Mat_MPIDense *) A->data; 439 int ierr; 440 441 PetscFunctionBegin; 442 ierr = VecCopy(yy,zz); CHKERRQ(ierr); 443 ierr = MatMultTrans_SeqDense(a->A,xx,a->lvec); CHKERRQ(ierr); 444 ierr = VecScatterBegin(a->lvec,zz,ADD_VALUES,SCATTER_REVERSE,a->Mvctx); CHKERRQ(ierr); 445 ierr = VecScatterEnd(a->lvec,zz,ADD_VALUES,SCATTER_REVERSE,a->Mvctx); CHKERRQ(ierr); 446 PetscFunctionReturn(0); 447 } 448 449 #undef __FUNC__ 450 #define __FUNC__ "MatGetDiagonal_MPIDense" 451 int MatGetDiagonal_MPIDense(Mat A,Vec v) 452 { 453 Mat_MPIDense *a = (Mat_MPIDense *) A->data; 454 Mat_SeqDense *aloc = (Mat_SeqDense *) a->A->data; 455 int ierr, len, i, n, m = a->m, radd; 456 Scalar *x, zero = 0.0; 457 458 PetscFunctionBegin; 459 VecSet(&zero,v); 460 ierr = VecGetArray(v,&x); CHKERRQ(ierr); 461 ierr = VecGetSize(v,&n); CHKERRQ(ierr); 462 if (n != a->M) SETERRQ(PETSC_ERR_ARG_SIZ,0,"Nonconforming mat and vec"); 463 len = PetscMin(aloc->m,aloc->n); 464 radd = a->rstart*m; 465 for ( i=0; i<len; i++ ) { 466 x[i] = aloc->v[radd + i*m + i]; 467 } 468 PetscFunctionReturn(0); 469 } 470 471 #undef __FUNC__ 472 #define __FUNC__ "MatDestroy_MPIDense" 473 int MatDestroy_MPIDense(Mat mat) 474 { 475 Mat_MPIDense *mdn = (Mat_MPIDense *) mat->data; 476 int ierr; 477 478 PetscFunctionBegin; 479 if (--mat->refct > 0) PetscFunctionReturn(0); 480 481 if (mat->mapping) { 482 ierr = ISLocalToGlobalMappingDestroy(mat->mapping); CHKERRQ(ierr); 483 } 484 if (mat->bmapping) { 485 ierr = ISLocalToGlobalMappingDestroy(mat->bmapping); CHKERRQ(ierr); 486 } 487 #if defined(USE_PETSC_LOG) 488 PLogObjectState((PetscObject)mat,"Rows=%d, Cols=%d",mdn->M,mdn->N); 489 #endif 490 PetscFree(mdn->rowners); 491 ierr = MatDestroy(mdn->A); CHKERRQ(ierr); 492 if (mdn->lvec) VecDestroy(mdn->lvec); 493 if (mdn->Mvctx) VecScatterDestroy(mdn->Mvctx); 494 if (mdn->factor) { 495 if (mdn->factor->temp) PetscFree(mdn->factor->temp); 496 if (mdn->factor->tag) PetscFree(mdn->factor->tag); 497 if (mdn->factor->pivots) PetscFree(mdn->factor->pivots); 498 PetscFree(mdn->factor); 499 } 500 PetscFree(mdn); 501 if (mat->rmap) { 502 ierr = MapDestroy(mat->rmap);CHKERRQ(ierr); 503 } 504 if (mat->cmap) { 505 ierr = MapDestroy(mat->cmap);CHKERRQ(ierr); 506 } 507 PLogObjectDestroy(mat); 508 PetscHeaderDestroy(mat); 509 PetscFunctionReturn(0); 510 } 511 512 #undef __FUNC__ 513 #define __FUNC__ "MatView_MPIDense_Binary" 514 static int MatView_MPIDense_Binary(Mat mat,Viewer viewer) 515 { 516 Mat_MPIDense *mdn = (Mat_MPIDense *) mat->data; 517 int ierr; 518 519 PetscFunctionBegin; 520 if (mdn->size == 1) { 521 ierr = MatView(mdn->A,viewer); CHKERRQ(ierr); 522 } 523 else SETERRQ(PETSC_ERR_SUP,0,"Only uniprocessor output supported"); 524 PetscFunctionReturn(0); 525 } 526 527 #undef __FUNC__ 528 #define __FUNC__ "MatView_MPIDense_ASCII" 529 static int MatView_MPIDense_ASCII(Mat mat,Viewer viewer) 530 { 531 Mat_MPIDense *mdn = (Mat_MPIDense *) mat->data; 532 int ierr, format, size = mdn->size, rank = mdn->rank; 533 FILE *fd; 534 ViewerType vtype; 535 536 PetscFunctionBegin; 537 ierr = ViewerGetType(viewer,&vtype);CHKERRQ(ierr); 538 ierr = ViewerASCIIGetPointer(viewer,&fd); CHKERRQ(ierr); 539 ierr = ViewerGetFormat(viewer,&format); 540 if (format == VIEWER_FORMAT_ASCII_INFO_LONG) { 541 MatInfo info; 542 ierr = MatGetInfo(mat,MAT_LOCAL,&info); 543 PetscSequentialPhaseBegin(mat->comm,1); 544 fprintf(fd," [%d] local rows %d nz %d nz alloced %d mem %d \n",rank,mdn->m, 545 (int)info.nz_used,(int)info.nz_allocated,(int)info.memory); 546 fflush(fd); 547 PetscSequentialPhaseEnd(mat->comm,1); 548 ierr = VecScatterView(mdn->Mvctx,viewer); CHKERRQ(ierr); 549 PetscFunctionReturn(0); 550 } 551 else if (format == VIEWER_FORMAT_ASCII_INFO) { 552 PetscFunctionReturn(0); 553 } 554 555 if (size == 1) { 556 ierr = MatView(mdn->A,viewer); CHKERRQ(ierr); 557 } else { 558 /* assemble the entire matrix onto first processor. */ 559 Mat A; 560 int M = mdn->M, N = mdn->N,m,row,i, nz, *cols; 561 Scalar *vals; 562 Mat_SeqDense *Amdn = (Mat_SeqDense*) mdn->A->data; 563 564 if (!rank) { 565 ierr = MatCreateMPIDense(mat->comm,M,N,M,N,PETSC_NULL,&A); CHKERRQ(ierr); 566 } else { 567 ierr = MatCreateMPIDense(mat->comm,0,N,M,N,PETSC_NULL,&A); CHKERRQ(ierr); 568 } 569 PLogObjectParent(mat,A); 570 571 /* Copy the matrix ... This isn't the most efficient means, 572 but it's quick for now */ 573 row = mdn->rstart; m = Amdn->m; 574 for ( i=0; i<m; i++ ) { 575 ierr = MatGetRow(mat,row,&nz,&cols,&vals); CHKERRQ(ierr); 576 ierr = MatSetValues(A,1,&row,nz,cols,vals,INSERT_VALUES); CHKERRQ(ierr); 577 ierr = MatRestoreRow(mat,row,&nz,&cols,&vals); CHKERRQ(ierr); 578 row++; 579 } 580 581 ierr = MatAssemblyBegin(A,MAT_FINAL_ASSEMBLY); CHKERRQ(ierr); 582 ierr = MatAssemblyEnd(A,MAT_FINAL_ASSEMBLY); CHKERRQ(ierr); 583 if (!rank) { 584 ierr = MatView(((Mat_MPIDense*)(A->data))->A,viewer); CHKERRQ(ierr); 585 } 586 ierr = MatDestroy(A); CHKERRQ(ierr); 587 } 588 PetscFunctionReturn(0); 589 } 590 591 #undef __FUNC__ 592 #define __FUNC__ "MatView_MPIDense" 593 int MatView_MPIDense(Mat mat,Viewer viewer) 594 { 595 int ierr; 596 ViewerType vtype; 597 598 ierr = ViewerGetType(viewer,&vtype); CHKERRQ(ierr); 599 if (PetscTypeCompare(vtype,ASCII_VIEWER)) { 600 ierr = MatView_MPIDense_ASCII(mat,viewer); CHKERRQ(ierr); 601 } else if (PetscTypeCompare(vtype,BINARY_VIEWER)) { 602 ierr = MatView_MPIDense_Binary(mat,viewer);CHKERRQ(ierr); 603 } else { 604 SETERRQ(1,1,"Viewer type not supported by PETSc object"); 605 } 606 PetscFunctionReturn(0); 607 } 608 609 #undef __FUNC__ 610 #define __FUNC__ "MatGetInfo_MPIDense" 611 int MatGetInfo_MPIDense(Mat A,MatInfoType flag,MatInfo *info) 612 { 613 Mat_MPIDense *mat = (Mat_MPIDense *) A->data; 614 Mat mdn = mat->A; 615 int ierr; 616 double isend[5], irecv[5]; 617 618 PetscFunctionBegin; 619 info->rows_global = (double)mat->M; 620 info->columns_global = (double)mat->N; 621 info->rows_local = (double)mat->m; 622 info->columns_local = (double)mat->N; 623 info->block_size = 1.0; 624 ierr = MatGetInfo(mdn,MAT_LOCAL,info); CHKERRQ(ierr); 625 isend[0] = info->nz_used; isend[1] = info->nz_allocated; isend[2] = info->nz_unneeded; 626 isend[3] = info->memory; isend[4] = info->mallocs; 627 if (flag == MAT_LOCAL) { 628 info->nz_used = isend[0]; 629 info->nz_allocated = isend[1]; 630 info->nz_unneeded = isend[2]; 631 info->memory = isend[3]; 632 info->mallocs = isend[4]; 633 } else if (flag == MAT_GLOBAL_MAX) { 634 ierr = MPI_Allreduce(isend,irecv,5,MPI_DOUBLE,MPI_MAX,A->comm);CHKERRQ(ierr); 635 info->nz_used = irecv[0]; 636 info->nz_allocated = irecv[1]; 637 info->nz_unneeded = irecv[2]; 638 info->memory = irecv[3]; 639 info->mallocs = irecv[4]; 640 } else if (flag == MAT_GLOBAL_SUM) { 641 ierr = MPI_Allreduce(isend,irecv,5,MPI_DOUBLE,MPI_SUM,A->comm);CHKERRQ(ierr); 642 info->nz_used = irecv[0]; 643 info->nz_allocated = irecv[1]; 644 info->nz_unneeded = irecv[2]; 645 info->memory = irecv[3]; 646 info->mallocs = irecv[4]; 647 } 648 info->fill_ratio_given = 0; /* no parallel LU/ILU/Cholesky */ 649 info->fill_ratio_needed = 0; 650 info->factor_mallocs = 0; 651 PetscFunctionReturn(0); 652 } 653 654 /* extern int MatLUFactorSymbolic_MPIDense(Mat,IS,IS,double,Mat*); 655 extern int MatLUFactorNumeric_MPIDense(Mat,Mat*); 656 extern int MatLUFactor_MPIDense(Mat,IS,IS,double); 657 extern int MatSolve_MPIDense(Mat,Vec,Vec); 658 extern int MatSolveAdd_MPIDense(Mat,Vec,Vec,Vec); 659 extern int MatSolveTrans_MPIDense(Mat,Vec,Vec); 660 extern int MatSolveTransAdd_MPIDense(Mat,Vec,Vec,Vec); */ 661 662 #undef __FUNC__ 663 #define __FUNC__ "MatSetOption_MPIDense" 664 int MatSetOption_MPIDense(Mat A,MatOption op) 665 { 666 Mat_MPIDense *a = (Mat_MPIDense *) A->data; 667 668 PetscFunctionBegin; 669 if (op == MAT_NO_NEW_NONZERO_LOCATIONS || 670 op == MAT_YES_NEW_NONZERO_LOCATIONS || 671 op == MAT_NEW_NONZERO_LOCATION_ERROR || 672 op == MAT_NEW_NONZERO_ALLOCATION_ERROR || 673 op == MAT_COLUMNS_SORTED || 674 op == MAT_COLUMNS_UNSORTED) { 675 MatSetOption(a->A,op); 676 } else if (op == MAT_ROW_ORIENTED) { 677 a->roworiented = 1; 678 MatSetOption(a->A,op); 679 } else if (op == MAT_ROWS_SORTED || 680 op == MAT_ROWS_UNSORTED || 681 op == MAT_SYMMETRIC || 682 op == MAT_STRUCTURALLY_SYMMETRIC || 683 op == MAT_YES_NEW_DIAGONALS || 684 op == MAT_USE_HASH_TABLE) { 685 PLogInfo(A,"MatSetOption_MPIDense:Option ignored\n"); 686 } else if (op == MAT_COLUMN_ORIENTED) { 687 a->roworiented = 0; MatSetOption(a->A,op); 688 } else if (op == MAT_NO_NEW_DIAGONALS) { 689 SETERRQ(PETSC_ERR_SUP,0,"MAT_NO_NEW_DIAGONALS"); 690 } else { 691 SETERRQ(PETSC_ERR_SUP,0,"unknown option"); 692 } 693 PetscFunctionReturn(0); 694 } 695 696 #undef __FUNC__ 697 #define __FUNC__ "MatGetSize_MPIDense" 698 int MatGetSize_MPIDense(Mat A,int *m,int *n) 699 { 700 Mat_MPIDense *mat = (Mat_MPIDense *) A->data; 701 702 PetscFunctionBegin; 703 *m = mat->M; *n = mat->N; 704 PetscFunctionReturn(0); 705 } 706 707 #undef __FUNC__ 708 #define __FUNC__ "MatGetLocalSize_MPIDense" 709 int MatGetLocalSize_MPIDense(Mat A,int *m,int *n) 710 { 711 Mat_MPIDense *mat = (Mat_MPIDense *) A->data; 712 713 PetscFunctionBegin; 714 *m = mat->m; *n = mat->N; 715 PetscFunctionReturn(0); 716 } 717 718 #undef __FUNC__ 719 #define __FUNC__ "MatGetOwnershipRange_MPIDense" 720 int MatGetOwnershipRange_MPIDense(Mat A,int *m,int *n) 721 { 722 Mat_MPIDense *mat = (Mat_MPIDense *) A->data; 723 724 PetscFunctionBegin; 725 *m = mat->rstart; *n = mat->rend; 726 PetscFunctionReturn(0); 727 } 728 729 #undef __FUNC__ 730 #define __FUNC__ "MatGetRow_MPIDense" 731 int MatGetRow_MPIDense(Mat A,int row,int *nz,int **idx,Scalar **v) 732 { 733 Mat_MPIDense *mat = (Mat_MPIDense *) A->data; 734 int lrow, rstart = mat->rstart, rend = mat->rend,ierr; 735 736 PetscFunctionBegin; 737 if (row < rstart || row >= rend) SETERRQ(PETSC_ERR_SUP,0,"only local rows") 738 lrow = row - rstart; 739 ierr = MatGetRow(mat->A,lrow,nz,idx,v);CHKERRQ(ierr); 740 PetscFunctionReturn(0); 741 } 742 743 #undef __FUNC__ 744 #define __FUNC__ "MatRestoreRow_MPIDense" 745 int MatRestoreRow_MPIDense(Mat mat,int row,int *nz,int **idx,Scalar **v) 746 { 747 PetscFunctionBegin; 748 if (idx) PetscFree(*idx); 749 if (v) PetscFree(*v); 750 PetscFunctionReturn(0); 751 } 752 753 #undef __FUNC__ 754 #define __FUNC__ "MatNorm_MPIDense" 755 int MatNorm_MPIDense(Mat A,NormType type,double *norm) 756 { 757 Mat_MPIDense *mdn = (Mat_MPIDense *) A->data; 758 Mat_SeqDense *mat = (Mat_SeqDense*) mdn->A->data; 759 int ierr, i, j; 760 double sum = 0.0; 761 Scalar *v = mat->v; 762 763 PetscFunctionBegin; 764 if (mdn->size == 1) { 765 ierr = MatNorm(mdn->A,type,norm); CHKERRQ(ierr); 766 } else { 767 if (type == NORM_FROBENIUS) { 768 for (i=0; i<mat->n*mat->m; i++ ) { 769 #if defined(USE_PETSC_COMPLEX) 770 sum += PetscReal(PetscConj(*v)*(*v)); v++; 771 #else 772 sum += (*v)*(*v); v++; 773 #endif 774 } 775 ierr = MPI_Allreduce(&sum,norm,1,MPI_DOUBLE,MPI_SUM,A->comm);CHKERRQ(ierr); 776 *norm = sqrt(*norm); 777 PLogFlops(2*mat->n*mat->m); 778 } else if (type == NORM_1) { 779 double *tmp, *tmp2; 780 tmp = (double *) PetscMalloc( 2*mdn->N*sizeof(double) ); CHKPTRQ(tmp); 781 tmp2 = tmp + mdn->N; 782 PetscMemzero(tmp,2*mdn->N*sizeof(double)); 783 *norm = 0.0; 784 v = mat->v; 785 for ( j=0; j<mat->n; j++ ) { 786 for ( i=0; i<mat->m; i++ ) { 787 tmp[j] += PetscAbsScalar(*v); v++; 788 } 789 } 790 ierr = MPI_Allreduce(tmp,tmp2,mdn->N,MPI_DOUBLE,MPI_SUM,A->comm);CHKERRQ(ierr); 791 for ( j=0; j<mdn->N; j++ ) { 792 if (tmp2[j] > *norm) *norm = tmp2[j]; 793 } 794 PetscFree(tmp); 795 PLogFlops(mat->n*mat->m); 796 } else if (type == NORM_INFINITY) { /* max row norm */ 797 double ntemp; 798 ierr = MatNorm(mdn->A,type,&ntemp); CHKERRQ(ierr); 799 ierr = MPI_Allreduce(&ntemp,norm,1,MPI_DOUBLE,MPI_MAX,A->comm);CHKERRQ(ierr); 800 } else { 801 SETERRQ(PETSC_ERR_SUP,0,"No support for two norm"); 802 } 803 } 804 PetscFunctionReturn(0); 805 } 806 807 #undef __FUNC__ 808 #define __FUNC__ "MatTranspose_MPIDense" 809 int MatTranspose_MPIDense(Mat A,Mat *matout) 810 { 811 Mat_MPIDense *a = (Mat_MPIDense *) A->data; 812 Mat_SeqDense *Aloc = (Mat_SeqDense *) a->A->data; 813 Mat B; 814 int M = a->M, N = a->N, m, n, *rwork, rstart = a->rstart; 815 int j, i, ierr; 816 Scalar *v; 817 818 PetscFunctionBegin; 819 if (matout == PETSC_NULL && M != N) { 820 SETERRQ(PETSC_ERR_SUP,0,"Supports square matrix only in-place"); 821 } 822 ierr = MatCreateMPIDense(A->comm,PETSC_DECIDE,PETSC_DECIDE,N,M,PETSC_NULL,&B);CHKERRQ(ierr); 823 824 m = Aloc->m; n = Aloc->n; v = Aloc->v; 825 rwork = (int *) PetscMalloc(n*sizeof(int)); CHKPTRQ(rwork); 826 for ( j=0; j<n; j++ ) { 827 for (i=0; i<m; i++) rwork[i] = rstart + i; 828 ierr = MatSetValues(B,1,&j,m,rwork,v,INSERT_VALUES); CHKERRQ(ierr); 829 v += m; 830 } 831 PetscFree(rwork); 832 ierr = MatAssemblyBegin(B,MAT_FINAL_ASSEMBLY); CHKERRQ(ierr); 833 ierr = MatAssemblyEnd(B,MAT_FINAL_ASSEMBLY); CHKERRQ(ierr); 834 if (matout != PETSC_NULL) { 835 *matout = B; 836 } else { 837 PetscOps *Abops; 838 MatOps Aops; 839 840 /* This isn't really an in-place transpose, but free data struct from a */ 841 PetscFree(a->rowners); 842 ierr = MatDestroy(a->A); CHKERRQ(ierr); 843 if (a->lvec) VecDestroy(a->lvec); 844 if (a->Mvctx) VecScatterDestroy(a->Mvctx); 845 PetscFree(a); 846 847 /* 848 This is horrible, horrible code. We need to keep the 849 A pointers for the bops and ops but copy everything 850 else from C. 851 */ 852 Abops = A->bops; 853 Aops = A->ops; 854 PetscMemcpy(A,B,sizeof(struct _p_Mat)); 855 A->bops = Abops; 856 A->ops = Aops; 857 858 PetscHeaderDestroy(B); 859 } 860 PetscFunctionReturn(0); 861 } 862 863 #include "pinclude/blaslapack.h" 864 #undef __FUNC__ 865 #define __FUNC__ "MatScale_MPIDense" 866 int MatScale_MPIDense(Scalar *alpha,Mat inA) 867 { 868 Mat_MPIDense *A = (Mat_MPIDense *) inA->data; 869 Mat_SeqDense *a = (Mat_SeqDense *) A->A->data; 870 int one = 1, nz; 871 872 PetscFunctionBegin; 873 nz = a->m*a->n; 874 BLscal_( &nz, alpha, a->v, &one ); 875 PLogFlops(nz); 876 PetscFunctionReturn(0); 877 } 878 879 static int MatDuplicate_MPIDense(Mat,MatDuplicateOption,Mat *); 880 extern int MatGetSubMatrices_MPIDense(Mat,int,IS *,IS *,MatReuse,Mat **); 881 882 /* -------------------------------------------------------------------*/ 883 static struct _MatOps MatOps_Values = {MatSetValues_MPIDense, 884 MatGetRow_MPIDense, 885 MatRestoreRow_MPIDense, 886 MatMult_MPIDense, 887 MatMultAdd_MPIDense, 888 MatMultTrans_MPIDense, 889 MatMultTransAdd_MPIDense, 890 0, 891 0, 892 0, 893 0, 894 0, 895 0, 896 0, 897 MatTranspose_MPIDense, 898 MatGetInfo_MPIDense,0, 899 MatGetDiagonal_MPIDense, 900 0, 901 MatNorm_MPIDense, 902 MatAssemblyBegin_MPIDense, 903 MatAssemblyEnd_MPIDense, 904 0, 905 MatSetOption_MPIDense, 906 MatZeroEntries_MPIDense, 907 MatZeroRows_MPIDense, 908 0, 909 0, 910 0, 911 0, 912 MatGetSize_MPIDense, 913 MatGetLocalSize_MPIDense, 914 MatGetOwnershipRange_MPIDense, 915 0, 916 0, 917 MatGetArray_MPIDense, 918 MatRestoreArray_MPIDense, 919 MatDuplicate_MPIDense, 920 0, 921 0, 922 0, 923 0, 924 0, 925 MatGetSubMatrices_MPIDense, 926 0, 927 MatGetValues_MPIDense, 928 0, 929 0, 930 MatScale_MPIDense, 931 0, 932 0, 933 0, 934 MatGetBlockSize_MPIDense, 935 0, 936 0, 937 0, 938 0, 939 0, 940 0, 941 0, 942 0, 943 0, 944 0, 945 0, 946 0, 947 MatGetMaps_Petsc}; 948 949 #undef __FUNC__ 950 #define __FUNC__ "MatCreateMPIDense" 951 /*@C 952 MatCreateMPIDense - Creates a sparse parallel matrix in dense format. 953 954 Collective on MPI_Comm 955 956 Input Parameters: 957 + comm - MPI communicator 958 . m - number of local rows (or PETSC_DECIDE to have calculated if M is given) 959 . n - number of local columns (or PETSC_DECIDE to have calculated if N is given) 960 . M - number of global rows (or PETSC_DECIDE to have calculated if m is given) 961 . N - number of global columns (or PETSC_DECIDE to have calculated if n is given) 962 - data - optional location of matrix data. Set data=PETSC_NULL for PETSc 963 to control all matrix memory allocation. 964 965 Output Parameter: 966 . A - the matrix 967 968 Notes: 969 The dense format is fully compatible with standard Fortran 77 970 storage by columns. 971 972 The data input variable is intended primarily for Fortran programmers 973 who wish to allocate their own matrix memory space. Most users should 974 set data=PETSC_NULL. 975 976 The user MUST specify either the local or global matrix dimensions 977 (possibly both). 978 979 Currently, the only parallel dense matrix decomposition is by rows, 980 so that n=N and each submatrix owns all of the global columns. 981 982 Level: intermediate 983 984 .keywords: matrix, dense, parallel 985 986 .seealso: MatCreate(), MatCreateSeqDense(), MatSetValues() 987 @*/ 988 int MatCreateMPIDense(MPI_Comm comm,int m,int n,int M,int N,Scalar *data,Mat *A) 989 { 990 Mat mat; 991 Mat_MPIDense *a; 992 int ierr, i,flg; 993 994 PetscFunctionBegin; 995 /* Note: For now, when data is specified above, this assumes the user correctly 996 allocates the local dense storage space. We should add error checking. */ 997 998 *A = 0; 999 PetscHeaderCreate(mat,_p_Mat,struct _MatOps,MAT_COOKIE,MATMPIDENSE,"Mat",comm,MatDestroy,MatView); 1000 PLogObjectCreate(mat); 1001 mat->data = (void *) (a = PetscNew(Mat_MPIDense)); CHKPTRQ(a); 1002 PetscMemcpy(mat->ops,&MatOps_Values,sizeof(struct _MatOps)); 1003 mat->ops->destroy = MatDestroy_MPIDense; 1004 mat->ops->view = MatView_MPIDense; 1005 mat->factor = 0; 1006 mat->mapping = 0; 1007 1008 a->factor = 0; 1009 mat->insertmode = NOT_SET_VALUES; 1010 MPI_Comm_rank(comm,&a->rank); 1011 MPI_Comm_size(comm,&a->size); 1012 1013 if (M == PETSC_DECIDE) {ierr = MPI_Allreduce(&m,&M,1,MPI_INT,MPI_SUM,comm);CHKERRQ(ierr);} 1014 if (m == PETSC_DECIDE) {m = M/a->size + ((M % a->size) > a->rank);} 1015 1016 /* 1017 The computation of n is wrong below, n should represent the number of local 1018 rows in the right (column vector) 1019 */ 1020 1021 /* each row stores all columns */ 1022 if (N == PETSC_DECIDE) N = n; 1023 if (n == PETSC_DECIDE) {n = N/a->size + ((N % a->size) > a->rank);} 1024 /* if (n != N) SETERRQ(PETSC_ERR_SUP,0,"For now, only n=N is supported"); */ 1025 a->N = mat->N = N; 1026 a->M = mat->M = M; 1027 a->m = mat->m = m; 1028 a->n = mat->n = n; 1029 1030 /* the information in the maps duplicates the information computed below, eventually 1031 we should remove the duplicate information that is not contained in the maps */ 1032 ierr = MapCreateMPI(comm,m,M,&mat->rmap);CHKERRQ(ierr); 1033 ierr = MapCreateMPI(comm,n,N,&mat->cmap);CHKERRQ(ierr); 1034 1035 /* build local table of row and column ownerships */ 1036 a->rowners = (int *) PetscMalloc(2*(a->size+2)*sizeof(int)); CHKPTRQ(a->rowners); 1037 a->cowners = a->rowners + a->size + 1; 1038 PLogObjectMemory(mat,2*(a->size+2)*sizeof(int)+sizeof(struct _p_Mat)+sizeof(Mat_MPIDense)); 1039 ierr = MPI_Allgather(&m,1,MPI_INT,a->rowners+1,1,MPI_INT,comm);CHKERRQ(ierr); 1040 a->rowners[0] = 0; 1041 for ( i=2; i<=a->size; i++ ) { 1042 a->rowners[i] += a->rowners[i-1]; 1043 } 1044 a->rstart = a->rowners[a->rank]; 1045 a->rend = a->rowners[a->rank+1]; 1046 ierr = MPI_Allgather(&n,1,MPI_INT,a->cowners+1,1,MPI_INT,comm);CHKERRQ(ierr); 1047 a->cowners[0] = 0; 1048 for ( i=2; i<=a->size; i++ ) { 1049 a->cowners[i] += a->cowners[i-1]; 1050 } 1051 1052 ierr = MatCreateSeqDense(PETSC_COMM_SELF,m,N,data,&a->A); CHKERRQ(ierr); 1053 PLogObjectParent(mat,a->A); 1054 1055 /* build cache for off array entries formed */ 1056 ierr = StashBuild_Private(&a->stash); CHKERRQ(ierr); 1057 1058 /* stuff used for matrix vector multiply */ 1059 a->lvec = 0; 1060 a->Mvctx = 0; 1061 a->roworiented = 1; 1062 1063 *A = mat; 1064 ierr = OptionsHasName(PETSC_NULL,"-help",&flg); CHKERRQ(ierr); 1065 if (flg) { 1066 ierr = MatPrintHelp(mat); CHKERRQ(ierr); 1067 } 1068 PetscFunctionReturn(0); 1069 } 1070 1071 #undef __FUNC__ 1072 #define __FUNC__ "MatDuplicate_MPIDense" 1073 static int MatDuplicate_MPIDense(Mat A,MatDuplicateOption cpvalues,Mat *newmat) 1074 { 1075 Mat mat; 1076 Mat_MPIDense *a,*oldmat = (Mat_MPIDense *) A->data; 1077 int ierr; 1078 FactorCtx *factor; 1079 1080 PetscFunctionBegin; 1081 *newmat = 0; 1082 PetscHeaderCreate(mat,_p_Mat,struct _MatOps,MAT_COOKIE,MATMPIDENSE,"Mat",A->comm,MatDestroy,MatView); 1083 PLogObjectCreate(mat); 1084 mat->data = (void *) (a = PetscNew(Mat_MPIDense)); CHKPTRQ(a); 1085 PetscMemcpy(mat->ops,&MatOps_Values,sizeof(struct _MatOps)); 1086 mat->ops->destroy = MatDestroy_MPIDense; 1087 mat->ops->view = MatView_MPIDense; 1088 mat->factor = A->factor; 1089 mat->assembled = PETSC_TRUE; 1090 1091 a->m = mat->m = oldmat->m; 1092 a->n = mat->n = oldmat->n; 1093 a->M = mat->M = oldmat->M; 1094 a->N = mat->N = oldmat->N; 1095 if (oldmat->factor) { 1096 a->factor = (FactorCtx *) (factor = PetscNew(FactorCtx)); CHKPTRQ(factor); 1097 /* copy factor contents ... add this code! */ 1098 } else a->factor = 0; 1099 1100 a->rstart = oldmat->rstart; 1101 a->rend = oldmat->rend; 1102 a->size = oldmat->size; 1103 a->rank = oldmat->rank; 1104 mat->insertmode = NOT_SET_VALUES; 1105 1106 a->rowners = (int *) PetscMalloc((a->size+1)*sizeof(int)); CHKPTRQ(a->rowners); 1107 PLogObjectMemory(mat,(a->size+1)*sizeof(int)+sizeof(struct _p_Mat)+sizeof(Mat_MPIDense)); 1108 PetscMemcpy(a->rowners,oldmat->rowners,(a->size+1)*sizeof(int)); 1109 ierr = StashInitialize_Private(&a->stash); CHKERRQ(ierr); 1110 1111 ierr = VecDuplicate(oldmat->lvec,&a->lvec); CHKERRQ(ierr); 1112 PLogObjectParent(mat,a->lvec); 1113 ierr = VecScatterCopy(oldmat->Mvctx,&a->Mvctx); CHKERRQ(ierr); 1114 PLogObjectParent(mat,a->Mvctx); 1115 ierr = MatDuplicate(oldmat->A,cpvalues,&a->A); CHKERRQ(ierr); 1116 PLogObjectParent(mat,a->A); 1117 *newmat = mat; 1118 PetscFunctionReturn(0); 1119 } 1120 1121 #include "sys.h" 1122 1123 #undef __FUNC__ 1124 #define __FUNC__ "MatLoad_MPIDense_DenseInFile" 1125 int MatLoad_MPIDense_DenseInFile(MPI_Comm comm,int fd,int M, int N, Mat *newmat) 1126 { 1127 int *rowners, i,size,rank,m,ierr,nz,j; 1128 Scalar *array,*vals,*vals_ptr; 1129 MPI_Status status; 1130 1131 PetscFunctionBegin; 1132 MPI_Comm_rank(comm,&rank); 1133 MPI_Comm_size(comm,&size); 1134 1135 /* determine ownership of all rows */ 1136 m = M/size + ((M % size) > rank); 1137 rowners = (int *) PetscMalloc((size+2)*sizeof(int)); CHKPTRQ(rowners); 1138 ierr = MPI_Allgather(&m,1,MPI_INT,rowners+1,1,MPI_INT,comm);CHKERRQ(ierr); 1139 rowners[0] = 0; 1140 for ( i=2; i<=size; i++ ) { 1141 rowners[i] += rowners[i-1]; 1142 } 1143 1144 ierr = MatCreateMPIDense(comm,m,PETSC_DECIDE,M,N,PETSC_NULL,newmat);CHKERRQ(ierr); 1145 ierr = MatGetArray(*newmat,&array); CHKERRQ(ierr); 1146 1147 if (!rank) { 1148 vals = (Scalar *) PetscMalloc( m*N*sizeof(Scalar) ); CHKPTRQ(vals); 1149 1150 /* read in my part of the matrix numerical values */ 1151 ierr = PetscBinaryRead(fd,vals,m*N,PETSC_SCALAR); CHKERRQ(ierr); 1152 1153 /* insert into matrix-by row (this is why cannot directly read into array */ 1154 vals_ptr = vals; 1155 for ( i=0; i<m; i++ ) { 1156 for ( j=0; j<N; j++ ) { 1157 array[i + j*m] = *vals_ptr++; 1158 } 1159 } 1160 1161 /* read in other processors and ship out */ 1162 for ( i=1; i<size; i++ ) { 1163 nz = (rowners[i+1] - rowners[i])*N; 1164 ierr = PetscBinaryRead(fd,vals,nz,PETSC_SCALAR); CHKERRQ(ierr); 1165 ierr = MPI_Send(vals,nz,MPIU_SCALAR,i,(*newmat)->tag,comm);CHKERRQ(ierr); 1166 } 1167 } else { 1168 /* receive numeric values */ 1169 vals = (Scalar*) PetscMalloc( m*N*sizeof(Scalar) ); CHKPTRQ(vals); 1170 1171 /* receive message of values*/ 1172 ierr = MPI_Recv(vals,m*N,MPIU_SCALAR,0,(*newmat)->tag,comm,&status);CHKERRQ(ierr); 1173 1174 /* insert into matrix-by row (this is why cannot directly read into array */ 1175 vals_ptr = vals; 1176 for ( i=0; i<m; i++ ) { 1177 for ( j=0; j<N; j++ ) { 1178 array[i + j*m] = *vals_ptr++; 1179 } 1180 } 1181 } 1182 PetscFree(rowners); 1183 PetscFree(vals); 1184 ierr = MatAssemblyBegin(*newmat,MAT_FINAL_ASSEMBLY); CHKERRQ(ierr); 1185 ierr = MatAssemblyEnd(*newmat,MAT_FINAL_ASSEMBLY); CHKERRQ(ierr); 1186 PetscFunctionReturn(0); 1187 } 1188 1189 1190 #undef __FUNC__ 1191 #define __FUNC__ "MatLoad_MPIDense" 1192 int MatLoad_MPIDense(Viewer viewer,MatType type,Mat *newmat) 1193 { 1194 Mat A; 1195 Scalar *vals,*svals; 1196 MPI_Comm comm = ((PetscObject)viewer)->comm; 1197 MPI_Status status; 1198 int header[4],rank,size,*rowlengths = 0,M,N,m,*rowners,maxnz,*cols; 1199 int *ourlens,*sndcounts = 0,*procsnz = 0, *offlens,jj,*mycols,*smycols; 1200 int tag = ((PetscObject)viewer)->tag; 1201 int i, nz, ierr, j,rstart, rend, fd; 1202 1203 PetscFunctionBegin; 1204 MPI_Comm_size(comm,&size); MPI_Comm_rank(comm,&rank); 1205 if (!rank) { 1206 ierr = ViewerBinaryGetDescriptor(viewer,&fd); CHKERRQ(ierr); 1207 ierr = PetscBinaryRead(fd,(char *)header,4,PETSC_INT); CHKERRQ(ierr); 1208 if (header[0] != MAT_COOKIE) SETERRQ(PETSC_ERR_FILE_UNEXPECTED,0,"not matrix object"); 1209 } 1210 1211 ierr = MPI_Bcast(header+1,3,MPI_INT,0,comm);CHKERRQ(ierr); 1212 M = header[1]; N = header[2]; nz = header[3]; 1213 1214 /* 1215 Handle case where matrix is stored on disk as a dense matrix 1216 */ 1217 if (nz == MATRIX_BINARY_FORMAT_DENSE) { 1218 ierr = MatLoad_MPIDense_DenseInFile(comm,fd,M,N,newmat);CHKERRQ(ierr); 1219 PetscFunctionReturn(0); 1220 } 1221 1222 /* determine ownership of all rows */ 1223 m = M/size + ((M % size) > rank); 1224 rowners = (int *) PetscMalloc((size+2)*sizeof(int)); CHKPTRQ(rowners); 1225 ierr = MPI_Allgather(&m,1,MPI_INT,rowners+1,1,MPI_INT,comm);CHKERRQ(ierr); 1226 rowners[0] = 0; 1227 for ( i=2; i<=size; i++ ) { 1228 rowners[i] += rowners[i-1]; 1229 } 1230 rstart = rowners[rank]; 1231 rend = rowners[rank+1]; 1232 1233 /* distribute row lengths to all processors */ 1234 ourlens = (int*) PetscMalloc( 2*(rend-rstart)*sizeof(int) ); CHKPTRQ(ourlens); 1235 offlens = ourlens + (rend-rstart); 1236 if (!rank) { 1237 rowlengths = (int*) PetscMalloc( M*sizeof(int) ); CHKPTRQ(rowlengths); 1238 ierr = PetscBinaryRead(fd,rowlengths,M,PETSC_INT); CHKERRQ(ierr); 1239 sndcounts = (int*) PetscMalloc( size*sizeof(int) ); CHKPTRQ(sndcounts); 1240 for ( i=0; i<size; i++ ) sndcounts[i] = rowners[i+1] - rowners[i]; 1241 ierr = MPI_Scatterv(rowlengths,sndcounts,rowners,MPI_INT,ourlens,rend-rstart,MPI_INT,0,comm);CHKERRQ(ierr); 1242 PetscFree(sndcounts); 1243 } else { 1244 ierr = MPI_Scatterv(0,0,0,MPI_INT,ourlens,rend-rstart,MPI_INT, 0,comm);CHKERRQ(ierr); 1245 } 1246 1247 if (!rank) { 1248 /* calculate the number of nonzeros on each processor */ 1249 procsnz = (int*) PetscMalloc( size*sizeof(int) ); CHKPTRQ(procsnz); 1250 PetscMemzero(procsnz,size*sizeof(int)); 1251 for ( i=0; i<size; i++ ) { 1252 for ( j=rowners[i]; j< rowners[i+1]; j++ ) { 1253 procsnz[i] += rowlengths[j]; 1254 } 1255 } 1256 PetscFree(rowlengths); 1257 1258 /* determine max buffer needed and allocate it */ 1259 maxnz = 0; 1260 for ( i=0; i<size; i++ ) { 1261 maxnz = PetscMax(maxnz,procsnz[i]); 1262 } 1263 cols = (int *) PetscMalloc( maxnz*sizeof(int) ); CHKPTRQ(cols); 1264 1265 /* read in my part of the matrix column indices */ 1266 nz = procsnz[0]; 1267 mycols = (int *) PetscMalloc( nz*sizeof(int) ); CHKPTRQ(mycols); 1268 ierr = PetscBinaryRead(fd,mycols,nz,PETSC_INT); CHKERRQ(ierr); 1269 1270 /* read in every one elses and ship off */ 1271 for ( i=1; i<size; i++ ) { 1272 nz = procsnz[i]; 1273 ierr = PetscBinaryRead(fd,cols,nz,PETSC_INT); CHKERRQ(ierr); 1274 ierr = MPI_Send(cols,nz,MPI_INT,i,tag,comm);CHKERRQ(ierr); 1275 } 1276 PetscFree(cols); 1277 } else { 1278 /* determine buffer space needed for message */ 1279 nz = 0; 1280 for ( i=0; i<m; i++ ) { 1281 nz += ourlens[i]; 1282 } 1283 mycols = (int*) PetscMalloc( nz*sizeof(int) ); CHKPTRQ(mycols); 1284 1285 /* receive message of column indices*/ 1286 ierr = MPI_Recv(mycols,nz,MPI_INT,0,tag,comm,&status);CHKERRQ(ierr); 1287 ierr = MPI_Get_count(&status,MPI_INT,&maxnz);CHKERRQ(ierr); 1288 if (maxnz != nz) SETERRQ(PETSC_ERR_FILE_UNEXPECTED,0,"something is wrong with file"); 1289 } 1290 1291 /* loop over local rows, determining number of off diagonal entries */ 1292 PetscMemzero(offlens,m*sizeof(int)); 1293 jj = 0; 1294 for ( i=0; i<m; i++ ) { 1295 for ( j=0; j<ourlens[i]; j++ ) { 1296 if (mycols[jj] < rstart || mycols[jj] >= rend) offlens[i]++; 1297 jj++; 1298 } 1299 } 1300 1301 /* create our matrix */ 1302 for ( i=0; i<m; i++ ) { 1303 ourlens[i] -= offlens[i]; 1304 } 1305 ierr = MatCreateMPIDense(comm,m,PETSC_DECIDE,M,N,PETSC_NULL,newmat);CHKERRQ(ierr); 1306 A = *newmat; 1307 for ( i=0; i<m; i++ ) { 1308 ourlens[i] += offlens[i]; 1309 } 1310 1311 if (!rank) { 1312 vals = (Scalar *) PetscMalloc( maxnz*sizeof(Scalar) ); CHKPTRQ(vals); 1313 1314 /* read in my part of the matrix numerical values */ 1315 nz = procsnz[0]; 1316 ierr = PetscBinaryRead(fd,vals,nz,PETSC_SCALAR); CHKERRQ(ierr); 1317 1318 /* insert into matrix */ 1319 jj = rstart; 1320 smycols = mycols; 1321 svals = vals; 1322 for ( i=0; i<m; i++ ) { 1323 ierr = MatSetValues(A,1,&jj,ourlens[i],smycols,svals,INSERT_VALUES);CHKERRQ(ierr); 1324 smycols += ourlens[i]; 1325 svals += ourlens[i]; 1326 jj++; 1327 } 1328 1329 /* read in other processors and ship out */ 1330 for ( i=1; i<size; i++ ) { 1331 nz = procsnz[i]; 1332 ierr = PetscBinaryRead(fd,vals,nz,PETSC_SCALAR); CHKERRQ(ierr); 1333 ierr = MPI_Send(vals,nz,MPIU_SCALAR,i,A->tag,comm);CHKERRQ(ierr); 1334 } 1335 PetscFree(procsnz); 1336 } else { 1337 /* receive numeric values */ 1338 vals = (Scalar*) PetscMalloc( nz*sizeof(Scalar) ); CHKPTRQ(vals); 1339 1340 /* receive message of values*/ 1341 ierr = MPI_Recv(vals,nz,MPIU_SCALAR,0,A->tag,comm,&status);CHKERRQ(ierr); 1342 ierr = MPI_Get_count(&status,MPIU_SCALAR,&maxnz);CHKERRQ(ierr); 1343 if (maxnz != nz) SETERRQ(PETSC_ERR_FILE_UNEXPECTED,0,"something is wrong with file"); 1344 1345 /* insert into matrix */ 1346 jj = rstart; 1347 smycols = mycols; 1348 svals = vals; 1349 for ( i=0; i<m; i++ ) { 1350 ierr = MatSetValues(A,1,&jj,ourlens[i],smycols,svals,INSERT_VALUES);CHKERRQ(ierr); 1351 smycols += ourlens[i]; 1352 svals += ourlens[i]; 1353 jj++; 1354 } 1355 } 1356 PetscFree(ourlens); PetscFree(vals); PetscFree(mycols); PetscFree(rowners); 1357 1358 ierr = MatAssemblyBegin(A,MAT_FINAL_ASSEMBLY); CHKERRQ(ierr); 1359 ierr = MatAssemblyEnd(A,MAT_FINAL_ASSEMBLY); CHKERRQ(ierr); 1360 PetscFunctionReturn(0); 1361 } 1362 1363 1364 1365 1366 1367