#include /* Note that the API for using PETSCVIEWERVTK is totally wrong since its use requires including the private vtkvimpl.h file. The code should be refactored. */ #include <../src/sys/classes/viewer/impls/vtk/vtkvimpl.h> /* Helper function which determines if any DMDA fields are named. This is used as a proxy for the user's intention to use DMDA fields as distinct scalar-valued fields as opposed to a single vector-valued field */ static PetscErrorCode DMDAGetFieldsNamed(DM da,PetscBool *fieldsnamed) { PetscErrorCode ierr; PetscInt f,bs; PetscFunctionBegin; *fieldsnamed = PETSC_FALSE; ierr = DMDAGetDof(da,&bs);CHKERRQ(ierr); for (f=0; fdata; PetscViewerVTKObjectLink link; FILE *fp; PetscMPIInt rank,size,tag; DMDALocalInfo info; PetscInt dim,mx,my,mz,cdim,bs,boffset,maxnnodes,maxbs,i,j,k,r; PetscInt rloc[6],(*grloc)[6] = NULL; PetscScalar *array,*array2; PetscErrorCode ierr; PetscFunctionBegin; ierr = PetscObjectGetComm((PetscObject)da,&comm);CHKERRQ(ierr); if (PetscDefined(USE_COMPLEX)) SETERRQ(PETSC_COMM_SELF,PETSC_ERR_SUP,"Complex values not supported"); ierr = MPI_Comm_size(comm,&size);CHKERRMPI(ierr); ierr = MPI_Comm_rank(comm,&rank);CHKERRMPI(ierr); ierr = DMDAGetInfo(da,&dim,&mx,&my,&mz,NULL,NULL,NULL,&bs,NULL,NULL,NULL,NULL,NULL);CHKERRQ(ierr); ierr = DMDAGetLocalInfo(da,&info);CHKERRQ(ierr); ierr = DMGetCoordinates(da,&Coords);CHKERRQ(ierr); if (Coords) { PetscInt csize; ierr = VecGetSize(Coords,&csize);CHKERRQ(ierr); if (csize % (mx*my*mz)) SETERRQ(PETSC_COMM_SELF,PETSC_ERR_PLIB,"Coordinate vector size mismatch"); cdim = csize/(mx*my*mz); if (cdim < dim || cdim > 3) SETERRQ(PETSC_COMM_SELF,PETSC_ERR_PLIB,"Coordinate vector size mismatch"); } else { cdim = dim; } ierr = PetscFOpen(comm,vtk->filename,"wb",&fp);CHKERRQ(ierr); ierr = PetscFPrintf(comm,fp,"\n");CHKERRQ(ierr); ierr = PetscFPrintf(comm,fp,"\n",byte_order);CHKERRQ(ierr); ierr = PetscFPrintf(comm,fp," \n",0,mx-1,0,my-1,0,mz-1);CHKERRQ(ierr); if (rank == 0) {ierr = PetscMalloc1(size*6,&grloc);CHKERRQ(ierr);} rloc[0] = info.xs; rloc[1] = info.xm; rloc[2] = info.ys; rloc[3] = info.ym; rloc[4] = info.zs; rloc[5] = info.zm; ierr = MPI_Gather(rloc,6,MPIU_INT,&grloc[0][0],6,MPIU_INT,0,comm);CHKERRMPI(ierr); /* Write XML header */ maxnnodes = 0; /* Used for the temporary array size on rank 0 */ maxbs = 0; /* Used for the temporary array size on rank 0 */ boffset = 0; /* Offset into binary file */ for (r=0; r\n",xs,xs+xm-1,ys,ys+ym-1,zs,zs+zm-1);CHKERRQ(ierr); ierr = PetscFPrintf(comm,fp," \n");CHKERRQ(ierr); ierr = PetscFPrintf(comm,fp," \n",precision,boffset);CHKERRQ(ierr); boffset += 3*nnodes*sizeof(PetscScalar) + sizeof(int); ierr = PetscFPrintf(comm,fp," \n");CHKERRQ(ierr); ierr = PetscFPrintf(comm,fp," \n");CHKERRQ(ierr); for (link=vtk->link; link; link=link->next) { Vec X = (Vec)link->vec; PetscInt bs,f; DM daCurr; PetscBool fieldsnamed; const char *vecname = "Unnamed"; ierr = VecGetDM(X,&daCurr);CHKERRQ(ierr); ierr = DMDAGetInfo(daCurr,NULL,NULL,NULL,NULL,NULL,NULL,NULL,&bs,NULL,NULL,NULL,NULL,NULL);CHKERRQ(ierr); maxbs = PetscMax(maxbs,bs); if (((PetscObject)X)->name || link != vtk->link) { ierr = PetscObjectGetName((PetscObject)X,&vecname);CHKERRQ(ierr); } /* If any fields are named, add scalar fields. Otherwise, add a vector field */ ierr = DMDAGetFieldsNamed(daCurr,&fieldsnamed);CHKERRQ(ierr); if (fieldsnamed) { for (f=0; f\n",precision,vecname,fieldname,boffset);CHKERRQ(ierr); boffset += nnodes*sizeof(PetscScalar) + sizeof(int); } } else { ierr = PetscFPrintf(comm,fp," \n",precision,vecname,bs,boffset);CHKERRQ(ierr); boffset += bs*nnodes*sizeof(PetscScalar) + sizeof(int); } } ierr = PetscFPrintf(comm,fp," \n");CHKERRQ(ierr); ierr = PetscFPrintf(comm,fp," \n");CHKERRQ(ierr); } ierr = PetscFPrintf(comm,fp," \n");CHKERRQ(ierr); ierr = PetscFPrintf(comm,fp," \n");CHKERRQ(ierr); ierr = PetscFPrintf(comm,fp,"_");CHKERRQ(ierr); /* Now write the arrays. */ tag = ((PetscObject)viewer)->tag; ierr = PetscMalloc2(maxnnodes*PetscMax(3,maxbs),&array,maxnnodes*PetscMax(3,maxbs),&array2);CHKERRQ(ierr); for (r=0; r 1 ? array[Iloc*cdim + 1] : 0.0; array2[Iloc*3+2] = cdim > 2 ? array[Iloc*cdim + 2] : 0.0; } } } } else if (r == rank) { ierr = MPI_Send((void*)coords,nnodes*cdim,MPIU_SCALAR,0,tag,comm);CHKERRMPI(ierr); } ierr = VecRestoreArrayRead(Coords,&coords);CHKERRQ(ierr); } else { /* Fabricate some coordinates using grid index */ for (k=0; klink; link; link=link->next) { Vec X = (Vec)link->vec; const PetscScalar *x; PetscInt bs,f; DM daCurr; PetscBool fieldsnamed; ierr = VecGetDM(X,&daCurr);CHKERRQ(ierr); ierr = DMDAGetInfo(daCurr,NULL,NULL,NULL,NULL, NULL,NULL,NULL,&bs,NULL,NULL,NULL,NULL,NULL);CHKERRQ(ierr); ierr = VecGetArrayRead(X,&x);CHKERRQ(ierr); if (rank == 0) { if (r) { PetscMPIInt nn; ierr = MPI_Recv(array,nnodes*bs,MPIU_SCALAR,r,tag,comm,&status);CHKERRMPI(ierr); ierr = MPI_Get_count(&status,MPIU_SCALAR,&nn);CHKERRMPI(ierr); if (nn != nnodes*bs) SETERRQ1(PETSC_COMM_SELF,PETSC_ERR_PLIB,"Array size mismatch receiving from rank %D",r); } else { ierr = PetscArraycpy(array,x,nnodes*bs);CHKERRQ(ierr); } /* If any fields are named, add scalar fields. Otherwise, add a vector field */ ierr = DMDAGetFieldsNamed(daCurr,&fieldsnamed);CHKERRQ(ierr); if (fieldsnamed) { for (f=0; f\n");CHKERRQ(ierr); ierr = PetscFPrintf(comm,fp,"\n");CHKERRQ(ierr); ierr = PetscFClose(comm,fp);CHKERRQ(ierr); PetscFunctionReturn(0); } static PetscErrorCode DMDAVTKWriteAll_VTR(DM da,PetscViewer viewer) { const char *byte_order = PetscBinaryBigEndian() ? "BigEndian" : "LittleEndian"; #if defined(PETSC_USE_REAL_SINGLE) const char precision[] = "Float32"; #elif defined(PETSC_USE_REAL_DOUBLE) const char precision[] = "Float64"; #else const char precision[] = "UnknownPrecision"; #endif MPI_Comm comm; PetscViewer_VTK *vtk = (PetscViewer_VTK*)viewer->data; PetscViewerVTKObjectLink link; FILE *fp; PetscMPIInt rank,size,tag; DMDALocalInfo info; PetscInt dim,mx,my,mz,boffset,maxnnodes,maxbs,i,j,k,r; PetscInt rloc[6],(*grloc)[6] = NULL; PetscScalar *array,*array2; PetscErrorCode ierr; PetscFunctionBegin; ierr = PetscObjectGetComm((PetscObject)da,&comm);CHKERRQ(ierr); if (PetscDefined(USE_COMPLEX)) SETERRQ(PETSC_COMM_SELF,PETSC_ERR_SUP,"Complex values not supported"); ierr = MPI_Comm_size(comm,&size);CHKERRMPI(ierr); ierr = MPI_Comm_rank(comm,&rank);CHKERRMPI(ierr); ierr = DMDAGetInfo(da,&dim,&mx,&my,&mz,NULL,NULL,NULL,NULL,NULL,NULL,NULL,NULL,NULL);CHKERRQ(ierr); ierr = DMDAGetLocalInfo(da,&info);CHKERRQ(ierr); ierr = PetscFOpen(comm,vtk->filename,"wb",&fp);CHKERRQ(ierr); ierr = PetscFPrintf(comm,fp,"\n");CHKERRQ(ierr); ierr = PetscFPrintf(comm,fp,"\n",byte_order);CHKERRQ(ierr); ierr = PetscFPrintf(comm,fp," \n",0,mx-1,0,my-1,0,mz-1);CHKERRQ(ierr); if (rank == 0) {ierr = PetscMalloc1(size*6,&grloc);CHKERRQ(ierr);} rloc[0] = info.xs; rloc[1] = info.xm; rloc[2] = info.ys; rloc[3] = info.ym; rloc[4] = info.zs; rloc[5] = info.zm; ierr = MPI_Gather(rloc,6,MPIU_INT,&grloc[0][0],6,MPIU_INT,0,comm);CHKERRMPI(ierr); /* Write XML header */ maxnnodes = 0; /* Used for the temporary array size on rank 0 */ maxbs = 0; /* Used for the temporary array size on rank 0 */ boffset = 0; /* Offset into binary file */ for (r=0; r\n",xs,xs+xm-1,ys,ys+ym-1,zs,zs+zm-1);CHKERRQ(ierr); ierr = PetscFPrintf(comm,fp," \n");CHKERRQ(ierr); ierr = PetscFPrintf(comm,fp," \n",precision,boffset);CHKERRQ(ierr); boffset += xm*sizeof(PetscScalar) + sizeof(int); ierr = PetscFPrintf(comm,fp," \n",precision,boffset);CHKERRQ(ierr); boffset += ym*sizeof(PetscScalar) + sizeof(int); ierr = PetscFPrintf(comm,fp," \n",precision,boffset);CHKERRQ(ierr); boffset += zm*sizeof(PetscScalar) + sizeof(int); ierr = PetscFPrintf(comm,fp," \n");CHKERRQ(ierr); ierr = PetscFPrintf(comm,fp," \n");CHKERRQ(ierr); for (link=vtk->link; link; link=link->next) { Vec X = (Vec)link->vec; PetscInt bs,f; DM daCurr; PetscBool fieldsnamed; const char *vecname = "Unnamed"; ierr = VecGetDM(X,&daCurr);CHKERRQ(ierr); ierr = DMDAGetInfo(daCurr,NULL,NULL,NULL,NULL,NULL,NULL,NULL,&bs,NULL,NULL,NULL,NULL,NULL);CHKERRQ(ierr); maxbs = PetscMax(maxbs,bs); if (((PetscObject)X)->name || link != vtk->link) { ierr = PetscObjectGetName((PetscObject)X,&vecname);CHKERRQ(ierr); } /* If any fields are named, add scalar fields. Otherwise, add a vector field */ ierr = DMDAGetFieldsNamed(daCurr,&fieldsnamed);CHKERRQ(ierr); if (fieldsnamed) { for (f=0; f\n",precision,vecname,fieldname,boffset);CHKERRQ(ierr); boffset += nnodes*sizeof(PetscScalar) + sizeof(int); } } else { ierr = PetscFPrintf(comm,fp," \n",precision,vecname,bs,boffset);CHKERRQ(ierr); boffset += bs*nnodes*sizeof(PetscScalar) + sizeof(int); } } ierr = PetscFPrintf(comm,fp," \n");CHKERRQ(ierr); ierr = PetscFPrintf(comm,fp," \n");CHKERRQ(ierr); } ierr = PetscFPrintf(comm,fp," \n");CHKERRQ(ierr); ierr = PetscFPrintf(comm,fp," \n");CHKERRQ(ierr); ierr = PetscFPrintf(comm,fp,"_");CHKERRQ(ierr); /* Now write the arrays. */ tag = ((PetscObject)viewer)->tag; ierr = PetscMalloc2(PetscMax(maxnnodes*maxbs,info.xm+info.ym+info.zm),&array,PetscMax(maxnnodes*maxbs,info.xm+info.ym+info.zm),&array2);CHKERRQ(ierr); for (r=0; r 1 ? coords[Iloc*dim + 1] : 0; } /* z coordinates */ for (i=0, j=0, k=0; k 2 ? coords[Iloc*dim + 2] : 0; } } } else if (r == rank) { xm = info.xm; ym = info.ym; zm = info.zm; for (j=0, k=0, i=0; i 1 ? coords[Iloc*dim + 1] : 0; } for (i=0, j=0, k=0; k 2 ? coords[Iloc*dim + 2] : 0; } ierr = MPI_Send((void*)array2,xm+ym+zm,MPIU_SCALAR,0,tag,comm);CHKERRMPI(ierr); } ierr = VecRestoreArrayRead(Coords,&coords);CHKERRQ(ierr); } else { /* Fabricate some coordinates using grid index */ for (i=0; ilink; link; link=link->next) { Vec X = (Vec)link->vec; const PetscScalar *x; PetscInt bs,f; DM daCurr; PetscBool fieldsnamed; ierr = VecGetDM(X,&daCurr);CHKERRQ(ierr); ierr = DMDAGetInfo(daCurr,NULL,NULL,NULL,NULL,NULL,NULL,NULL,&bs,NULL,NULL,NULL,NULL,NULL);CHKERRQ(ierr); ierr = VecGetArrayRead(X,&x);CHKERRQ(ierr); if (rank == 0) { if (r) { PetscMPIInt nn; ierr = MPI_Recv(array,nnodes*bs,MPIU_SCALAR,r,tag,comm,&status);CHKERRMPI(ierr); ierr = MPI_Get_count(&status,MPIU_SCALAR,&nn);CHKERRMPI(ierr); if (nn != nnodes*bs) SETERRQ1(PETSC_COMM_SELF,PETSC_ERR_PLIB,"Array size mismatch receiving from rank %D",r); } else { ierr = PetscArraycpy(array,x,nnodes*bs);CHKERRQ(ierr); } /* If any fields are named, add scalar fields. Otherwise, add a vector field */ ierr = DMDAGetFieldsNamed(daCurr,&fieldsnamed);CHKERRQ(ierr); if (fieldsnamed) { for (f=0; f\n");CHKERRQ(ierr); ierr = PetscFPrintf(comm,fp,"\n");CHKERRQ(ierr); ierr = PetscFClose(comm,fp);CHKERRQ(ierr); PetscFunctionReturn(0); } /*@C DMDAVTKWriteAll - Write a file containing all the fields that have been provided to the viewer Collective Input Parameters: + odm - DM specifying the grid layout, passed as a PetscObject - viewer - viewer of type VTK Level: developer Notes: This function is a callback used by the VTK viewer to actually write the file. The reason for this odd model is that the VTK file format does not provide any way to write one field at a time. Instead, metadata for the entire file needs to be available up-front before you can start writing the file. If any fields have been named (see e.g. DMDASetFieldName()), then individual scalar fields are written. Otherwise, a single multi-dof (vector) field is written. .seealso: PETSCVIEWERVTK @*/ PetscErrorCode DMDAVTKWriteAll(PetscObject odm,PetscViewer viewer) { DM dm = (DM)odm; PetscBool isvtk; PetscErrorCode ierr; PetscFunctionBegin; PetscValidHeaderSpecific(dm,DM_CLASSID,1); PetscValidHeaderSpecific(viewer,PETSC_VIEWER_CLASSID,2); ierr = PetscObjectTypeCompare((PetscObject)viewer,PETSCVIEWERVTK,&isvtk);CHKERRQ(ierr); if (!isvtk) SETERRQ1(PetscObjectComm((PetscObject)viewer),PETSC_ERR_ARG_INCOMP,"Cannot use viewer type %s",((PetscObject)viewer)->type_name); switch (viewer->format) { case PETSC_VIEWER_VTK_VTS: ierr = DMDAVTKWriteAll_VTS(dm,viewer);CHKERRQ(ierr); break; case PETSC_VIEWER_VTK_VTR: ierr = DMDAVTKWriteAll_VTR(dm,viewer);CHKERRQ(ierr); break; default: SETERRQ1(PetscObjectComm((PetscObject)dm),PETSC_ERR_SUP,"No support for format '%s'",PetscViewerFormats[viewer->format]); } PetscFunctionReturn(0); }