Actual source code: mpiu.c
2: #include <petscsys.h>
3: #include <petsc/private/petscimpl.h>
4: /*
5: Note that tag of 0 is ok because comm is a private communicator
6: generated below just for these routines.
7: */
9: PETSC_INTERN PetscErrorCode PetscSequentialPhaseBegin_Private(MPI_Comm comm,int ng)
10: {
12: PetscMPIInt rank,size,tag = 0;
13: MPI_Status status;
16: MPI_Comm_size(comm,&size);
17: if (size == 1) return(0);
18: MPI_Comm_rank(comm,&rank);
19: if (rank) {
20: MPI_Recv(NULL,0,MPI_INT,rank-1,tag,comm,&status);
21: }
22: /* Send to the next process in the group unless we are the last process */
23: if ((rank % ng) < ng - 1 && rank != size - 1) {
24: MPI_Send(NULL,0,MPI_INT,rank + 1,tag,comm);
25: }
26: return(0);
27: }
29: PETSC_INTERN PetscErrorCode PetscSequentialPhaseEnd_Private(MPI_Comm comm,int ng)
30: {
32: PetscMPIInt rank,size,tag = 0;
33: MPI_Status status;
36: MPI_Comm_rank(comm,&rank);
37: MPI_Comm_size(comm,&size);
38: if (size == 1) return(0);
40: /* Send to the first process in the next group */
41: if ((rank % ng) == ng - 1 || rank == size - 1) {
42: MPI_Send(NULL,0,MPI_INT,(rank + 1) % size,tag,comm);
43: }
44: if (rank == 0) {
45: MPI_Recv(NULL,0,MPI_INT,size-1,tag,comm,&status);
46: }
47: return(0);
48: }
50: /* ---------------------------------------------------------------------*/
51: /*
52: The variable Petsc_Seq_keyval is used to indicate an MPI attribute that
53: is attached to a communicator that manages the sequential phase code below.
54: */
55: PetscMPIInt Petsc_Seq_keyval = MPI_KEYVAL_INVALID;
57: /*@
58: PetscSequentialPhaseBegin - Begins a sequential section of code.
60: Collective
62: Input Parameters:
63: + comm - Communicator to sequentialize.
64: - ng - Number in processor group. This many processes are allowed to execute
65: at the same time (usually 1)
67: Level: intermediate
69: Notes:
70: PetscSequentialPhaseBegin() and PetscSequentialPhaseEnd() provide a
71: way to force a section of code to be executed by the processes in
72: rank order. Typically, this is done with
73: .vb
74: PetscSequentialPhaseBegin(comm, 1);
75: <code to be executed sequentially>
76: PetscSequentialPhaseEnd(comm, 1);
77: .ve
79: Often, the sequential code contains output statements (e.g., printf) to
80: be executed. Note that you may need to flush the I/O buffers before
81: calling PetscSequentialPhaseEnd(). Also, note that some systems do
82: not propagate I/O in any order to the controling terminal (in other words,
83: even if you flush the output, you may not get the data in the order
84: that you want).
86: .seealso: PetscSequentialPhaseEnd()
88: @*/
89: PetscErrorCode PetscSequentialPhaseBegin(MPI_Comm comm,int ng)
90: {
92: PetscMPIInt size;
93: MPI_Comm local_comm,*addr_local_comm;
96: PetscSysInitializePackage();
97: MPI_Comm_size(comm,&size);
98: if (size == 1) return(0);
100: /* Get the private communicator for the sequential operations */
101: if (Petsc_Seq_keyval == MPI_KEYVAL_INVALID) {
102: MPI_Comm_create_keyval(MPI_COMM_NULL_COPY_FN,MPI_COMM_NULL_DELETE_FN,&Petsc_Seq_keyval,NULL);
103: }
105: MPI_Comm_dup(comm,&local_comm);
106: PetscMalloc1(1,&addr_local_comm);
108: *addr_local_comm = local_comm;
110: MPI_Comm_set_attr(comm,Petsc_Seq_keyval,(void*)addr_local_comm);
111: PetscSequentialPhaseBegin_Private(local_comm,ng);
112: return(0);
113: }
115: /*@
116: PetscSequentialPhaseEnd - Ends a sequential section of code.
118: Collective
120: Input Parameters:
121: + comm - Communicator to sequentialize.
122: - ng - Number in processor group. This many processes are allowed to execute
123: at the same time (usually 1)
125: Level: intermediate
127: Notes:
128: See PetscSequentialPhaseBegin() for more details.
130: .seealso: PetscSequentialPhaseBegin()
132: @*/
133: PetscErrorCode PetscSequentialPhaseEnd(MPI_Comm comm,int ng)
134: {
136: PetscMPIInt size,flag;
137: MPI_Comm local_comm,*addr_local_comm;
140: MPI_Comm_size(comm,&size);
141: if (size == 1) return(0);
143: MPI_Comm_get_attr(comm,Petsc_Seq_keyval,(void**)&addr_local_comm,&flag);
144: if (!flag) SETERRQ(PETSC_COMM_SELF,PETSC_ERR_ARG_INCOMP,"Wrong MPI communicator; must pass in one used with PetscSequentialPhaseBegin()");
145: local_comm = *addr_local_comm;
147: PetscSequentialPhaseEnd_Private(local_comm,ng);
149: PetscFree(addr_local_comm);
150: MPI_Comm_free(&local_comm);
151: MPI_Comm_delete_attr(comm,Petsc_Seq_keyval);
152: return(0);
153: }
155: /*@C
156: PetscGlobalMinMaxInt - Get the global min/max from local min/max input
158: Collective
160: Input Parameter:
161: . minMaxVal - An array with the local min and max
163: Output Parameter:
164: . minMaxValGlobal - An array with the global min and max
166: Level: beginner
168: .seealso: PetscSplitOwnership()
169: @*/
170: PetscErrorCode PetscGlobalMinMaxInt(MPI_Comm comm, PetscInt minMaxVal[2], PetscInt minMaxValGlobal[2])
171: {
175: minMaxVal[1] = -minMaxVal[1];
176: MPI_Allreduce(minMaxVal, minMaxValGlobal, 2, MPIU_INT, MPI_MIN, comm);
177: minMaxValGlobal[1] = -minMaxValGlobal[1];
178: return(0);
179: }
181: /*@C
182: PetscGlobalMinMaxReal - Get the global min/max from local min/max input
184: Collective
186: Input Parameter:
187: . minMaxVal - An array with the local min and max
189: Output Parameter:
190: . minMaxValGlobal - An array with the global min and max
192: Level: beginner
194: .seealso: PetscSplitOwnership()
195: @*/
196: PetscErrorCode PetscGlobalMinMaxReal(MPI_Comm comm, PetscReal minMaxVal[2], PetscReal minMaxValGlobal[2])
197: {
201: minMaxVal[1] = -minMaxVal[1];
202: MPI_Allreduce(minMaxVal, minMaxValGlobal, 2, MPIU_REAL, MPI_MIN, comm);
203: minMaxValGlobal[1] = -minMaxValGlobal[1];
204: return(0);
205: }