Actual source code: mpitr.c
2: /*
3: Code for tracing mistakes in MPI usage. For example, sends that are never received,
4: nonblocking messages that are not correctly waited for, etc.
5: */
7: #include <petscsys.h>
9: #if defined(PETSC_USE_LOG) && !defined(PETSC_HAVE_MPIUNI)
11: /*@C
12: PetscMPIDump - Dumps a listing of incomplete MPI operations, such as sends that
13: have never been received, etc.
15: Collective on PETSC_COMM_WORLD
17: Input Parameter:
18: . fp - file pointer. If fp is NULL, stdout is assumed.
20: Options Database Key:
21: . -mpidump - Dumps MPI incompleteness during call to PetscFinalize()
23: Level: developer
25: .seealso: PetscMallocDump()
26: @*/
27: PetscErrorCode PetscMPIDump(FILE *fd)
28: {
29: PetscMPIInt rank;
30: double tsends,trecvs,work;
31: int err;
33: MPI_Comm_rank(PETSC_COMM_WORLD,&rank);
34: if (!fd) fd = PETSC_STDOUT;
36: /* Did we wait on all the non-blocking sends and receives? */
37: PetscSequentialPhaseBegin(PETSC_COMM_WORLD,1);
38: if (petsc_irecv_ct + petsc_isend_ct != petsc_sum_of_waits_ct) {
39: PetscFPrintf(PETSC_COMM_SELF,fd,"[%d]You have not waited on all non-blocking sends and receives",rank);
40: PetscFPrintf(PETSC_COMM_SELF,fd,"[%d]Number non-blocking sends %g receives %g number of waits %g\n",rank,petsc_isend_ct,petsc_irecv_ct,petsc_sum_of_waits_ct);
41: err = fflush(fd);
43: }
44: PetscSequentialPhaseEnd(PETSC_COMM_WORLD,1);
45: /* Did we receive all the messages that we sent? */
46: work = petsc_irecv_ct + petsc_recv_ct;
47: MPI_Reduce(&work,&trecvs,1,MPI_DOUBLE,MPI_SUM,0,PETSC_COMM_WORLD);
48: work = petsc_isend_ct + petsc_send_ct;
49: MPI_Reduce(&work,&tsends,1,MPI_DOUBLE,MPI_SUM,0,PETSC_COMM_WORLD);
50: if (rank == 0 && tsends != trecvs) {
51: PetscFPrintf(PETSC_COMM_SELF,fd,"Total number sends %g not equal receives %g\n",tsends,trecvs);
52: err = fflush(fd);
54: }
55: return 0;
56: }
58: #else
60: PetscErrorCode PetscMPIDump(FILE *fd)
61: {
62: return 0;
63: }
65: #endif
67: #if defined(PETSC_HAVE_MPI_PROCESS_SHARED_MEMORY)
68: /*
69: OpenMPI version of MPI_Win_allocate_shared() does not provide __float128 alignment so we provide
70: a utility that insures alignment up to data item size.
71: */
72: PetscErrorCode MPIU_Win_allocate_shared(MPI_Aint sz,PetscMPIInt szind,MPI_Info info,MPI_Comm comm,void *ptr,MPI_Win *win)
73: {
74: float *tmp;
76: MPI_Win_allocate_shared(16+sz,szind,info,comm,&tmp,win);
77: tmp += ((size_t)tmp) % szind ? szind/4 - ((((size_t)tmp) % szind)/4) : 0;
78: *(void**)ptr = (void*)tmp;
79: return 0;
80: }
82: PETSC_EXTERN PetscErrorCode MPIU_Win_shared_query(MPI_Win win,PetscMPIInt rank,MPI_Aint *sz,PetscMPIInt *szind,void *ptr)
83: {
84: float *tmp;
86: MPI_Win_shared_query(win,rank,sz,szind,&tmp);
88: tmp += ((size_t)tmp) % *szind ? *szind/4 - ((((size_t)tmp) % *szind)/4) : 0;
89: *(void**)ptr = (void*)tmp;
90: return 0;
91: }
93: #endif