本文整理汇总了C++中PetscMalloc函数的典型用法代码示例。如果您正苦于以下问题:C++ PetscMalloc函数的具体用法?C++ PetscMalloc怎么用?C++ PetscMalloc使用的例子?那么恭喜您, 这里精选的函数代码示例或许可以为您提供帮助。
在下文中一共展示了PetscMalloc函数的15个代码示例,这些例子默认根据受欢迎程度排序。您可以为喜欢或者感觉有用的代码点赞,您的评价将有助于系统推荐出更棒的C++代码示例。
示例1: DMDAGetElements_2D
static PetscErrorCode DMDAGetElements_2D(DM dm,PetscInt *nel,PetscInt *nen,const PetscInt *e[])
{
PetscErrorCode ierr;
DM_DA *da = (DM_DA*)dm->data;
PetscInt i,xs,xe,Xs,Xe;
PetscInt j,ys,ye,Ys,Ye;
PetscInt cnt=0, cell[4], ns=2, nn=3;
PetscInt c, split[] = {0,1,3,
2,3,1};
PetscFunctionBegin;
if (!da->e) {
if (da->elementtype == DMDA_ELEMENT_P1) {ns=2; nn=3;}
if (da->elementtype == DMDA_ELEMENT_Q1) {ns=1; nn=4;}
ierr = DMDAGetCorners(dm,&xs,&ys,0,&xe,&ye,0);CHKERRQ(ierr);
ierr = DMDAGetGhostCorners(dm,&Xs,&Ys,0,&Xe,&Ye,0);CHKERRQ(ierr);
xe += xs; Xe += Xs; if (xs != Xs) xs -= 1;
ye += ys; Ye += Ys; if (ys != Ys) ys -= 1;
da->ne = ns*(xe - xs - 1)*(ye - ys - 1);
ierr = PetscMalloc((1 + nn*da->ne)*sizeof(PetscInt),&da->e);CHKERRQ(ierr);
for (j=ys; j<ye-1; j++) {
for (i=xs; i<xe-1; i++) {
cell[0] = (i-Xs ) + (j-Ys )*(Xe-Xs);
cell[1] = (i-Xs+1) + (j-Ys )*(Xe-Xs);
cell[2] = (i-Xs+1) + (j-Ys+1)*(Xe-Xs);
cell[3] = (i-Xs ) + (j-Ys+1)*(Xe-Xs);
if (da->elementtype == DMDA_ELEMENT_P1) {
for (c=0; c<ns*nn; c++)
da->e[cnt++] = cell[split[c]];
}
if (da->elementtype == DMDA_ELEMENT_Q1) {
for (c=0; c<ns*nn; c++)
da->e[cnt++] = cell[c];
}
}
}
}
*nel = da->ne;
*nen = nn;
*e = da->e;
PetscFunctionReturn(0);
}
示例2: InitializeVectors
PetscErrorCode InitializeVectors( UserContext* uc)
{
PetscErrorCode ierr;
PetscFunctionBegin;
ierr = VecCreate(PETSC_COMM_WORLD, &uc->b); CHKERRQ(ierr);
ierr = VecSetSizes(uc->b, uc->numNodes, uc->numNodes); CHKERRQ(ierr);
ierr = VecSetType(uc->b, VECSEQ); CHKERRQ(ierr);
ierr = VecDuplicate(uc->b,&uc->p);CHKERRQ(ierr);
ierr = VecDuplicate(uc->b,&uc->u);CHKERRQ(ierr);
ierr = VecDuplicate(uc->b,&uc->v);CHKERRQ(ierr);
ierr = VecDuplicate(uc->b,&uc->px);CHKERRQ(ierr);
ierr = VecDuplicate(uc->b,&uc->py);CHKERRQ(ierr);
ierr = VecDuplicate(uc->b,&uc->c);CHKERRQ(ierr);
ierr = PetscMalloc(uc->n * sizeof(PetscReal), &uc->imageResult ); CHKERRQ(ierr);
PetscFunctionReturn(0);
}
示例3: insertnode
void insertnode(LIST *ilist, PetscInt Node)
{
node *_new;
node *current;
current = ilist->head;
PetscTruth Exist = PETSC_FALSE;
while(current) {
if (Node == current->Node) {
Exist = PETSC_TRUE;
}
if (Exist) break;
current = current->next;
}
if (!Exist) {
PetscMalloc(sizeof(node), &_new);
_new->next = ilist->head;
_new->Node = Node;
ilist->head = _new;
}
}
示例4: SNESMonitorSet
/*@C
SNESMonitorSetRatio - Sets SNES to use a monitor that prints the
ratio of the function norm at each iteration.
Collective on SNES
Input Parameters:
+ snes - the SNES context
- viewer - ASCII viewer to print output
Level: intermediate
.keywords: SNES, nonlinear, monitor, norm
.seealso: SNESMonitorSet(), SNESMonitorSolution(), SNESMonitorDefault()
@*/
PetscErrorCode SNESMonitorSetRatio(SNES snes,PetscViewer viewer)
{
PetscErrorCode ierr;
SNESMonitorRatioContext *ctx;
PetscReal *history;
PetscFunctionBegin;
if (!viewer) {
ierr = PetscViewerASCIIOpen(((PetscObject)snes)->comm,"stdout",&viewer);CHKERRQ(ierr);
ierr = PetscObjectReference((PetscObject)viewer);CHKERRQ(ierr);
}
ierr = PetscNewLog(snes,SNESMonitorRatioContext,&ctx);CHKERRQ(ierr);
ierr = SNESGetConvergenceHistory(snes,&history,PETSC_NULL,PETSC_NULL);CHKERRQ(ierr);
if (!history) {
ierr = PetscMalloc(100*sizeof(PetscReal),&ctx->history);CHKERRQ(ierr);
ierr = SNESSetConvergenceHistory(snes,ctx->history,0,100,PETSC_TRUE);CHKERRQ(ierr);
}
ctx->viewer = viewer;
ierr = SNESMonitorSet(snes,SNESMonitorRatio,ctx,SNESMonitorRatioDestroy);CHKERRQ(ierr);
PetscFunctionReturn(0);
}
示例5: VecDuplicate_Nest
static PetscErrorCode VecDuplicate_Nest(Vec x,Vec *y)
{
Vec_Nest *bx = (Vec_Nest*)x->data;
Vec Y;
Vec *sub;
PetscInt i;
PetscErrorCode ierr;
PetscFunctionBegin;
ierr = PetscMalloc(sizeof(Vec)*bx->nb,&sub);CHKERRQ(ierr);
for (i=0; i<bx->nb; i++) {
ierr = VecDuplicate(bx->v[i],&sub[i]);CHKERRQ(ierr);
}
ierr = VecCreateNest(PetscObjectComm((PetscObject)x),bx->nb,bx->is,sub,&Y);CHKERRQ(ierr);
for (i=0; i<bx->nb; i++) {
ierr = VecDestroy(&sub[i]);CHKERRQ(ierr);
}
ierr = PetscFree(sub);CHKERRQ(ierr);
*y = Y;
PetscFunctionReturn(0);
}
示例6: PetscCDGetMIS
PetscErrorCode PetscCDGetMIS(PetscCoarsenData *ail, IS *a_mis)
{
PetscErrorCode ierr;
PetscCDIntNd *n;
PetscInt ii,kk;
PetscInt *permute;
PetscFunctionBegin;
for (ii=kk=0;ii<ail->size;ii++){
n = ail->array[ii];
if (n) kk++;
}
ierr = PetscMalloc(kk*sizeof(PetscInt), &permute);CHKERRQ(ierr);
for (ii=kk=0;ii<ail->size;ii++){
n = ail->array[ii];
if (n) permute[kk++] = ii;
}
ierr = ISCreateGeneral(PETSC_COMM_SELF, kk, permute, PETSC_OWN_POINTER, a_mis);CHKERRQ(ierr);
PetscFunctionReturn(0);
}
示例7: DMLibMeshGetVariables
PetscErrorCode DMLibMeshGetVariables(DM dm, PetscInt *n, char*** varnames)
{
PetscErrorCode ierr;
PetscFunctionBegin;
PetscValidHeaderSpecific(dm,DM_CLASSID,1);
PetscBool islibmesh;
PetscInt i;
ierr = PetscObjectTypeCompare((PetscObject)dm, DMLIBMESH,&islibmesh);
if(!islibmesh) SETERRQ2(((PetscObject)dm)->comm, PETSC_ERR_ARG_WRONG, "Got DM oftype %s, not of type %s", ((PetscObject)dm)->type_name, DMLIBMESH);
DM_libMesh *dlm = (DM_libMesh *)(dm->data);
PetscValidPointer(n,2);
*n = dlm->varids->size();
if(!varnames) PetscFunctionReturn(0);
ierr = PetscMalloc(*n*sizeof(char*), varnames); CHKERRQ(ierr);
i = 0;
for(std::map<std::string, unsigned int>::const_iterator it = dlm->varids->begin(); it != dlm->varids->end(); ++it){
ierr = PetscStrallocpy(it->first.c_str(), *varnames+i); CHKERRQ(ierr);
++i;
}
PetscFunctionReturn(0);
}
示例8: vsnprintf
/*@C
PetscVSNPrintf - The PETSc version of vsnprintf(). Converts a PETSc format string into a standard C format string and then puts all the
function arguments into a string using the format statement.
Input Parameters:
+ str - location to put result
. len - the amount of space in str
+ format - the PETSc format string
- fullLength - the amount of space in str actually used.
Developer Notes: this function may be called from an error handler, if an error occurs when it is called by the error handler than likely
a recursion will occur and possible crash.
Level: developer
@*/
PetscErrorCode PetscVSNPrintf(char *str,size_t len,const char *format,size_t *fullLength,va_list Argp)
{
char *newformat;
char formatbuf[8*1024];
size_t oldLength,length;
int fullLengthInt;
PetscErrorCode ierr;
PetscFunctionBegin;
ierr = PetscStrlen(format, &oldLength);CHKERRQ(ierr);
if (oldLength < 8*1024) {
newformat = formatbuf;
oldLength = 8*1024-1;
} else {
oldLength = PETSC_MAX_LENGTH_FORMAT(oldLength);
ierr = PetscMalloc(oldLength * sizeof(char), &newformat);CHKERRQ(ierr);
}
PetscFormatConvert(format,newformat,oldLength);
ierr = PetscStrlen(newformat, &length);CHKERRQ(ierr);
#if 0
if (length > len) {
newformat[len] = '\0';
}
#endif
#if defined(PETSC_HAVE_VSNPRINTF_CHAR)
fullLengthInt = vsnprintf(str,len,newformat,(char *)Argp);
#elif defined(PETSC_HAVE_VSNPRINTF)
fullLengthInt = vsnprintf(str,len,newformat,Argp);
#elif defined(PETSC_HAVE__VSNPRINTF)
fullLengthInt = _vsnprintf(str,len,newformat,Argp);
#else
#error "vsnprintf not found"
#endif
if (fullLengthInt < 0) SETERRQ(PETSC_COMM_SELF,PETSC_ERR_SYS,"vsnprintf() failed");
if (fullLength) *fullLength = (size_t)fullLengthInt;
if (oldLength >= 8*1024) {
ierr = PetscFree(newformat);CHKERRQ(ierr);
}
PetscFunctionReturn(0);
}
示例9: printf
/*@C
PetscSynchronizedPrintf - Prints synchronized output from several processors.
Output of the first processor is followed by that of the second, etc.
Not Collective
Input Parameters:
+ comm - the communicator
- format - the usual printf() format string
Level: intermediate
Notes:
REQUIRES a intervening call to PetscSynchronizedFlush() for the information
from all the processors to be printed.
Fortran Note:
The call sequence is PetscSynchronizedPrintf(MPI_Comm, character(*), PetscErrorCode ierr) from Fortran.
That is, you can only pass a single character string from Fortran.
.seealso: PetscSynchronizedFlush(), PetscSynchronizedFPrintf(), PetscFPrintf(),
PetscPrintf(), PetscViewerASCIIPrintf(), PetscViewerASCIISynchronizedPrintf()
@*/
PetscErrorCode PetscSynchronizedPrintf(MPI_Comm comm,const char format[],...)
{
PetscErrorCode ierr;
PetscMPIInt rank;
PetscFunctionBegin;
ierr = MPI_Comm_rank(comm,&rank);CHKERRQ(ierr);
/* First processor prints immediately to stdout */
if (!rank) {
va_list Argp;
va_start(Argp,format);
ierr = (*PetscVFPrintf)(PETSC_STDOUT,format,Argp);CHKERRQ(ierr);
if (petsc_history) {
va_start(Argp,format);
ierr = (*PetscVFPrintf)(petsc_history,format,Argp);CHKERRQ(ierr);
}
va_end(Argp);
} else { /* other processors add to local queue */
va_list Argp;
PrintfQueue next;
size_t fullLength = 8191;
ierr = PetscNew(struct _PrintfQueue,&next);CHKERRQ(ierr);
if (petsc_printfqueue) {petsc_printfqueue->next = next; petsc_printfqueue = next; petsc_printfqueue->next = 0;}
else {petsc_printfqueuebase = petsc_printfqueue = next;}
petsc_printfqueuelength++;
next->size = -1;
while((PetscInt)fullLength >= next->size) {
next->size = fullLength+1;
ierr = PetscMalloc(next->size * sizeof(char), &next->string);CHKERRQ(ierr);
va_start(Argp,format);
ierr = PetscMemzero(next->string,next->size);CHKERRQ(ierr);
ierr = PetscVSNPrintf(next->string,next->size,format, &fullLength,Argp);CHKERRQ(ierr);
va_end(Argp);
}
}
PetscFunctionReturn(0);
}
示例10: MatCreateFFT
/*@
MatCreateFFT - Creates a matrix object that provides FFT via an external package
Collective on MPI_Comm
Input Parameter:
+ comm - MPI communicator
. ndim - the ndim-dimensional transform
. dim - array of size ndim, dim[i] contains the vector length in the i-dimension
- type - package type, e.g., FFTW or FFTCU
Output Parameter:
. A - the matrix
Options Database Keys:
+ -mat_fft_type - set FFT type
Level: intermediate
@*/
PetscErrorCode MatCreateFFT(MPI_Comm comm,PetscInt ndim,const PetscInt dim[],MatType mattype,Mat *A)
{
PetscErrorCode ierr;
PetscMPIInt size;
Mat FFT;
PetscInt N,i;
Mat_FFT *fft;
PetscFunctionBegin;
if (ndim < 1) SETERRQ1(comm,PETSC_ERR_USER,"ndim %d must be > 0",ndim);
ierr = MPI_Comm_size(comm, &size);CHKERRQ(ierr);
ierr = MatCreate(comm,&FFT);CHKERRQ(ierr);
ierr = PetscNewLog(FFT,Mat_FFT,&fft);CHKERRQ(ierr);
FFT->data = (void*)fft;
N = 1;
for (i=0; i<ndim; i++) {
if (dim[i] < 1) SETERRQ2(PETSC_COMM_SELF,PETSC_ERR_USER,"dim[%d]=%d must be > 0",i,dim[i]);
N *= dim[i];
}
ierr = PetscMalloc(ndim*sizeof(PetscInt),&fft->dim);CHKERRQ(ierr);
ierr = PetscMemcpy(fft->dim,dim,ndim*sizeof(PetscInt));CHKERRQ(ierr);
fft->ndim = ndim;
fft->n = PETSC_DECIDE;
fft->N = N;
fft->data = NULL;
ierr = MatSetType(FFT,mattype);CHKERRQ(ierr);
FFT->ops->destroy = MatDestroy_FFT;
/* get runtime options */
ierr = PetscOptionsBegin(PetscObjectComm((PetscObject)FFT),((PetscObject)FFT)->prefix,"FFT Options","Mat");CHKERRQ(ierr);
PetscOptionsEnd();
*A = FFT;
PetscFunctionReturn(0);
}
示例11: PetscCommBuildTwoSided_Allreduce
static PetscErrorCode PetscCommBuildTwoSided_Allreduce(MPI_Comm comm,PetscMPIInt count,MPI_Datatype dtype,PetscMPIInt nto,const PetscMPIInt *toranks,const void *todata,PetscMPIInt *nfrom,PetscMPIInt **fromranks,void *fromdata)
{
PetscErrorCode ierr;
PetscMPIInt size,*iflags,nrecvs,tag,*franks,i;
MPI_Aint lb,unitbytes;
char *tdata,*fdata;
MPI_Request *reqs,*sendreqs;
MPI_Status *statuses;
PetscFunctionBegin;
ierr = MPI_Comm_size(comm,&size);CHKERRQ(ierr);
ierr = PetscCalloc1(size,&iflags);CHKERRQ(ierr);
for (i=0; i<nto; i++) iflags[toranks[i]] = 1;
ierr = PetscGatherNumberOfMessages(comm,iflags,NULL,&nrecvs);CHKERRQ(ierr);
ierr = PetscFree(iflags);CHKERRQ(ierr);
ierr = PetscCommDuplicate(comm,&comm,&tag);CHKERRQ(ierr);
ierr = MPI_Type_get_extent(dtype,&lb,&unitbytes);CHKERRQ(ierr);
if (lb != 0) SETERRQ1(comm,PETSC_ERR_SUP,"Datatype with nonzero lower bound %ld\n",(long)lb);
ierr = PetscMalloc(nrecvs*count*unitbytes,&fdata);CHKERRQ(ierr);
tdata = (char*)todata;
ierr = PetscMalloc2(nto+nrecvs,&reqs,nto+nrecvs,&statuses);CHKERRQ(ierr);
sendreqs = reqs + nrecvs;
for (i=0; i<nrecvs; i++) {
ierr = MPI_Irecv((void*)(fdata+count*unitbytes*i),count,dtype,MPI_ANY_SOURCE,tag,comm,reqs+i);CHKERRQ(ierr);
}
for (i=0; i<nto; i++) {
ierr = MPI_Isend((void*)(tdata+count*unitbytes*i),count,dtype,toranks[i],tag,comm,sendreqs+i);CHKERRQ(ierr);
}
ierr = MPI_Waitall(nto+nrecvs,reqs,statuses);CHKERRQ(ierr);
ierr = PetscMalloc1(nrecvs,&franks);CHKERRQ(ierr);
for (i=0; i<nrecvs; i++) franks[i] = statuses[i].MPI_SOURCE;
ierr = PetscFree2(reqs,statuses);CHKERRQ(ierr);
ierr = PetscCommDestroy(&comm);CHKERRQ(ierr);
*nfrom = nrecvs;
*fromranks = franks;
*(void**)fromdata = fdata;
PetscFunctionReturn(0);
}
示例12: Monitor
PetscErrorCode Monitor(TS ts,PetscInt step,PetscReal time,Vec global,void *ctx)
{
VecScatter scatter;
IS from,to;
PetscInt i,n,*idx;
Vec tmp_vec;
PetscErrorCode ierr;
PetscScalar *tmp;
/* Get the size of the vector */
ierr = VecGetSize(global,&n);CHKERRQ(ierr);
/* Set the index sets */
ierr = PetscMalloc(n*sizeof(PetscInt),&idx);CHKERRQ(ierr);
for(i=0; i<n; i++) idx[i]=i;
/* Create local sequential vectors */
ierr = VecCreateSeq(PETSC_COMM_SELF,n,&tmp_vec);CHKERRQ(ierr);
/* Create scatter context */
ierr = ISCreateGeneral(PETSC_COMM_SELF,n,idx,PETSC_COPY_VALUES,&from);CHKERRQ(ierr);
ierr = ISCreateGeneral(PETSC_COMM_SELF,n,idx,PETSC_COPY_VALUES,&to);CHKERRQ(ierr);
ierr = VecScatterCreate(global,from,tmp_vec,to,&scatter);CHKERRQ(ierr);
ierr = VecScatterBegin(scatter,global,tmp_vec,INSERT_VALUES,SCATTER_FORWARD);CHKERRQ(ierr);
ierr = VecScatterEnd(scatter,global,tmp_vec,INSERT_VALUES,SCATTER_FORWARD);CHKERRQ(ierr);
ierr = VecGetArray(tmp_vec,&tmp);CHKERRQ(ierr);
ierr = PetscPrintf(PETSC_COMM_WORLD,"At t =%14.6e u = %14.6e %14.6e %14.6e \n",
time,PetscRealPart(tmp[0]),PetscRealPart(tmp[1]),PetscRealPart(tmp[2]));CHKERRQ(ierr);
ierr = PetscPrintf(PETSC_COMM_WORLD,"At t =%14.6e errors = %14.6e %14.6e %14.6e \n",
time,PetscRealPart(tmp[0]-solx(time)),PetscRealPart(tmp[1]-soly(time)),PetscRealPart(tmp[2]-solz(time)));CHKERRQ(ierr);
ierr = VecRestoreArray(tmp_vec,&tmp);CHKERRQ(ierr);
ierr = VecScatterDestroy(&scatter);CHKERRQ(ierr);
ierr = ISDestroy(&from);CHKERRQ(ierr);
ierr = ISDestroy(&to);CHKERRQ(ierr);
ierr = PetscFree(idx);CHKERRQ(ierr);
ierr = VecDestroy(&tmp_vec);CHKERRQ(ierr);
return 0;
}
示例13: DMPlexDistribute
/*@
DMPlexDistributeData - Distribute field data to match a given PetscSF, usually the SF from mesh distribution
Collective on DM
Input Parameters:
+ dm - The DMPlex object
. pointSF - The PetscSF describing the communication pattern
. originalSection - The PetscSection for existing data layout
. datatype - The type of data
- originalData - The existing data
Output Parameters:
+ newSection - The PetscSF describing the new data layout
- newData - The new data
Level: developer
.seealso: DMPlexDistribute(), DMPlexDistributeField()
@*/
PetscErrorCode DMPlexDistributeData(DM dm, PetscSF pointSF, PetscSection originalSection, MPI_Datatype datatype, void *originalData, PetscSection newSection, void **newData)
{
PetscSF fieldSF;
PetscInt *remoteOffsets, fieldSize;
PetscMPIInt dataSize;
PetscErrorCode ierr;
PetscFunctionBegin;
ierr = PetscLogEventBegin(DMPLEX_DistributeData,dm,0,0,0);CHKERRQ(ierr);
ierr = PetscSFDistributeSection(pointSF, originalSection, &remoteOffsets, newSection);CHKERRQ(ierr);
ierr = PetscSectionGetStorageSize(newSection, &fieldSize);CHKERRQ(ierr);
ierr = MPI_Type_size(datatype, &dataSize);CHKERRQ(ierr);
ierr = PetscMalloc(fieldSize * dataSize, newData);CHKERRQ(ierr);
ierr = PetscSFCreateSectionSF(pointSF, originalSection, remoteOffsets, newSection, &fieldSF);CHKERRQ(ierr);
ierr = PetscSFBcastBegin(fieldSF, datatype, originalData, *newData);CHKERRQ(ierr);
ierr = PetscSFBcastEnd(fieldSF, datatype, originalData, *newData);CHKERRQ(ierr);
ierr = PetscSFDestroy(&fieldSF);CHKERRQ(ierr);
ierr = PetscLogEventEnd(DMPLEX_DistributeData,dm,0,0,0);CHKERRQ(ierr);
PetscFunctionReturn(0);
}
示例14: PetscViewersCreate
/*@C
PetscViewersGetViewer - Gets a PetscViewer from a PetscViewer collection
Not Collective, but PetscViewer will be collective object on PetscViewers
Input Parameter:
+ viewers - object created with PetscViewersCreate()
- n - number of PetscViewer you want
Output Parameter:
. viewer - the PetscViewer
Level: intermediate
Concepts: PetscViewer^array of
.seealso: PetscViewersCreate(), PetscViewersDestroy()
@*/
PetscErrorCode PetscViewersGetViewer(PetscViewers viewers,PetscInt n,PetscViewer *viewer)
{
PetscErrorCode ierr;
PetscFunctionBegin;
if (n < 0) SETERRQ1(PETSC_COMM_SELF,PETSC_ERR_ARG_OUTOFRANGE,"Cannot access using a negative index - %d\n",n);
if (n >= viewers->n) {
PetscViewer *v;
int newn = n + 64; /* add 64 new ones at a time */
ierr = PetscMalloc(newn*sizeof(PetscViewer),&v);CHKERRQ(ierr);
ierr = PetscMemzero(v,newn*sizeof(PetscViewer));CHKERRQ(ierr);
ierr = PetscMemcpy(v,viewers->viewer,viewers->n*sizeof(PetscViewer));CHKERRQ(ierr);
ierr = PetscFree(viewers->viewer);CHKERRQ(ierr);
viewers->viewer = v;
}
if (!viewers->viewer[n]) {
ierr = PetscViewerCreate(viewers->comm,&viewers->viewer[n]);CHKERRQ(ierr);
}
*viewer = viewers->viewer[n];
PetscFunctionReturn(0);
}
示例15: time
/*@
PetscSequentialPhaseBegin - Begins a sequential section of code.
Collective on MPI_Comm
Input Parameters:
+ comm - Communicator to sequentialize.
- ng - Number in processor group. This many processes are allowed to execute
at the same time (usually 1)
Level: intermediate
Notes:
PetscSequentialPhaseBegin() and PetscSequentialPhaseEnd() provide a
way to force a section of code to be executed by the processes in
rank order. Typically, this is done with
.vb
PetscSequentialPhaseBegin(comm, 1);
<code to be executed sequentially>
PetscSequentialPhaseEnd(comm, 1);
.ve
Often, the sequential code contains output statements (e.g., printf) to
be executed. Note that you may need to flush the I/O buffers before
calling PetscSequentialPhaseEnd(). Also, note that some systems do
not propagate I/O in any order to the controling terminal (in other words,
even if you flush the output, you may not get the data in the order
that you want).
.seealso: PetscSequentialPhaseEnd()
Concepts: sequential stage
@*/
PetscErrorCode PetscSequentialPhaseBegin(MPI_Comm comm,int ng)
{
PetscErrorCode ierr;
PetscMPIInt size;
MPI_Comm local_comm,*addr_local_comm;
PetscFunctionBegin;
ierr = MPI_Comm_size(comm,&size);CHKERRQ(ierr);
if (size == 1) PetscFunctionReturn(0);
/* Get the private communicator for the sequential operations */
if (Petsc_Seq_keyval == MPI_KEYVAL_INVALID) {
ierr = MPI_Keyval_create(MPI_NULL_COPY_FN,MPI_NULL_DELETE_FN,&Petsc_Seq_keyval,0);CHKERRQ(ierr);
}
ierr = MPI_Comm_dup(comm,&local_comm);CHKERRQ(ierr);
ierr = PetscMalloc(sizeof(MPI_Comm),&addr_local_comm);CHKERRQ(ierr);
*addr_local_comm = local_comm;
ierr = MPI_Attr_put(comm,Petsc_Seq_keyval,(void*)addr_local_comm);CHKERRQ(ierr);
ierr = PetscSequentialPhaseBegin_Private(local_comm,ng);CHKERRQ(ierr);
PetscFunctionReturn(0);
}