1e5c89e4eSSatish Balay 2c6db04a5SJed Brown #include <petscsys.h> /*I "petscsys.h" I*/ 38ae1f613SSatish Balay #include <petsc/private/petscimpl.h> 4e5c89e4eSSatish Balay /* 5e5c89e4eSSatish Balay Note that tag of 0 is ok because comm is a private communicator 6e5c89e4eSSatish Balay generated below just for these routines. 7e5c89e4eSSatish Balay */ 8e5c89e4eSSatish Balay 9e5c89e4eSSatish Balay PetscErrorCode PetscSequentialPhaseBegin_Private(MPI_Comm comm,int ng) 10e5c89e4eSSatish Balay { 11e5c89e4eSSatish Balay PetscErrorCode ierr; 12e5c89e4eSSatish Balay PetscMPIInt rank,size,tag = 0; 13e5c89e4eSSatish Balay MPI_Status status; 14e5c89e4eSSatish Balay 15e5c89e4eSSatish Balay PetscFunctionBegin; 16e5c89e4eSSatish Balay ierr = MPI_Comm_size(comm,&size);CHKERRQ(ierr); 17e5c89e4eSSatish Balay if (size == 1) PetscFunctionReturn(0); 18e5c89e4eSSatish Balay ierr = MPI_Comm_rank(comm,&rank);CHKERRQ(ierr); 19e5c89e4eSSatish Balay if (rank) { 20e5c89e4eSSatish Balay ierr = MPI_Recv(0,0,MPI_INT,rank-1,tag,comm,&status);CHKERRQ(ierr); 21e5c89e4eSSatish Balay } 22e5c89e4eSSatish Balay /* Send to the next process in the group unless we are the last process */ 23e5c89e4eSSatish Balay if ((rank % ng) < ng - 1 && rank != size - 1) { 24e5c89e4eSSatish Balay ierr = MPI_Send(0,0,MPI_INT,rank + 1,tag,comm);CHKERRQ(ierr); 25e5c89e4eSSatish Balay } 26e5c89e4eSSatish Balay PetscFunctionReturn(0); 27e5c89e4eSSatish Balay } 28e5c89e4eSSatish Balay 29e5c89e4eSSatish Balay PetscErrorCode PetscSequentialPhaseEnd_Private(MPI_Comm comm,int ng) 30e5c89e4eSSatish Balay { 31e5c89e4eSSatish Balay PetscErrorCode ierr; 32e5c89e4eSSatish Balay PetscMPIInt rank,size,tag = 0; 33e5c89e4eSSatish Balay MPI_Status status; 34e5c89e4eSSatish Balay 35e5c89e4eSSatish Balay PetscFunctionBegin; 36e5c89e4eSSatish Balay ierr = MPI_Comm_rank(comm,&rank);CHKERRQ(ierr); 37e5c89e4eSSatish Balay ierr = MPI_Comm_size(comm,&size);CHKERRQ(ierr); 38e5c89e4eSSatish Balay if (size == 1) PetscFunctionReturn(0); 39e5c89e4eSSatish Balay 40e5c89e4eSSatish Balay /* Send to the first process in the next group */ 41e5c89e4eSSatish Balay if ((rank % ng) == ng - 1 || rank == size - 1) { 42e5c89e4eSSatish Balay ierr = MPI_Send(0,0,MPI_INT,(rank + 1) % size,tag,comm);CHKERRQ(ierr); 43e5c89e4eSSatish Balay } 44e5c89e4eSSatish Balay if (!rank) { 45e5c89e4eSSatish Balay ierr = MPI_Recv(0,0,MPI_INT,size-1,tag,comm,&status);CHKERRQ(ierr); 46e5c89e4eSSatish Balay } 47e5c89e4eSSatish Balay PetscFunctionReturn(0); 48e5c89e4eSSatish Balay } 49e5c89e4eSSatish Balay 50e5c89e4eSSatish Balay /* ---------------------------------------------------------------------*/ 51e5c89e4eSSatish Balay /* 52e5c89e4eSSatish Balay The variable Petsc_Seq_keyval is used to indicate an MPI attribute that 53e5c89e4eSSatish Balay is attached to a communicator that manages the sequential phase code below. 54e5c89e4eSSatish Balay */ 552fe4ef4aSBarry Smith PetscMPIInt Petsc_Seq_keyval = MPI_KEYVAL_INVALID; 56e5c89e4eSSatish Balay 57e30d2299SSatish Balay /*@ 58e5c89e4eSSatish Balay PetscSequentialPhaseBegin - Begins a sequential section of code. 59e5c89e4eSSatish Balay 60e5c89e4eSSatish Balay Collective on MPI_Comm 61e5c89e4eSSatish Balay 62e5c89e4eSSatish Balay Input Parameters: 63e5c89e4eSSatish Balay + comm - Communicator to sequentialize. 64e5c89e4eSSatish Balay - ng - Number in processor group. This many processes are allowed to execute 65e5c89e4eSSatish Balay at the same time (usually 1) 66e5c89e4eSSatish Balay 67e5c89e4eSSatish Balay Level: intermediate 68e5c89e4eSSatish Balay 69e5c89e4eSSatish Balay Notes: 70e5c89e4eSSatish Balay PetscSequentialPhaseBegin() and PetscSequentialPhaseEnd() provide a 71e5c89e4eSSatish Balay way to force a section of code to be executed by the processes in 72e5c89e4eSSatish Balay rank order. Typically, this is done with 73e5c89e4eSSatish Balay .vb 74e5c89e4eSSatish Balay PetscSequentialPhaseBegin(comm, 1); 75e5c89e4eSSatish Balay <code to be executed sequentially> 76e5c89e4eSSatish Balay PetscSequentialPhaseEnd(comm, 1); 77e5c89e4eSSatish Balay .ve 78e5c89e4eSSatish Balay 79e5c89e4eSSatish Balay Often, the sequential code contains output statements (e.g., printf) to 80e5c89e4eSSatish Balay be executed. Note that you may need to flush the I/O buffers before 81e5c89e4eSSatish Balay calling PetscSequentialPhaseEnd(). Also, note that some systems do 82e5c89e4eSSatish Balay not propagate I/O in any order to the controling terminal (in other words, 83e5c89e4eSSatish Balay even if you flush the output, you may not get the data in the order 84e5c89e4eSSatish Balay that you want). 85e5c89e4eSSatish Balay 86e5c89e4eSSatish Balay .seealso: PetscSequentialPhaseEnd() 87e5c89e4eSSatish Balay 88e5c89e4eSSatish Balay Concepts: sequential stage 89e5c89e4eSSatish Balay 90e5c89e4eSSatish Balay @*/ 917087cfbeSBarry Smith PetscErrorCode PetscSequentialPhaseBegin(MPI_Comm comm,int ng) 92e5c89e4eSSatish Balay { 93e5c89e4eSSatish Balay PetscErrorCode ierr; 94e5c89e4eSSatish Balay PetscMPIInt size; 95e5c89e4eSSatish Balay MPI_Comm local_comm,*addr_local_comm; 96e5c89e4eSSatish Balay 97e5c89e4eSSatish Balay PetscFunctionBegin; 98a70dbfdaSBarry Smith ierr = PetscSysInitializePackage();CHKERRQ(ierr); 99e5c89e4eSSatish Balay ierr = MPI_Comm_size(comm,&size);CHKERRQ(ierr); 100e5c89e4eSSatish Balay if (size == 1) PetscFunctionReturn(0); 101e5c89e4eSSatish Balay 102e5c89e4eSSatish Balay /* Get the private communicator for the sequential operations */ 103e5c89e4eSSatish Balay if (Petsc_Seq_keyval == MPI_KEYVAL_INVALID) { 104e5c89e4eSSatish Balay ierr = MPI_Keyval_create(MPI_NULL_COPY_FN,MPI_NULL_DELETE_FN,&Petsc_Seq_keyval,0);CHKERRQ(ierr); 105e5c89e4eSSatish Balay } 106e5c89e4eSSatish Balay 107e5c89e4eSSatish Balay ierr = MPI_Comm_dup(comm,&local_comm);CHKERRQ(ierr); 108854ce69bSBarry Smith ierr = PetscMalloc1(1,&addr_local_comm);CHKERRQ(ierr); 109a297a907SKarl Rupp 110e5c89e4eSSatish Balay *addr_local_comm = local_comm; 111a297a907SKarl Rupp 112e5c89e4eSSatish Balay ierr = MPI_Attr_put(comm,Petsc_Seq_keyval,(void*)addr_local_comm);CHKERRQ(ierr); 113e5c89e4eSSatish Balay ierr = PetscSequentialPhaseBegin_Private(local_comm,ng);CHKERRQ(ierr); 114e5c89e4eSSatish Balay PetscFunctionReturn(0); 115e5c89e4eSSatish Balay } 116e5c89e4eSSatish Balay 117e30d2299SSatish Balay /*@ 118e5c89e4eSSatish Balay PetscSequentialPhaseEnd - Ends a sequential section of code. 119e5c89e4eSSatish Balay 120e5c89e4eSSatish Balay Collective on MPI_Comm 121e5c89e4eSSatish Balay 122e5c89e4eSSatish Balay Input Parameters: 123e5c89e4eSSatish Balay + comm - Communicator to sequentialize. 124e5c89e4eSSatish Balay - ng - Number in processor group. This many processes are allowed to execute 125e5c89e4eSSatish Balay at the same time (usually 1) 126e5c89e4eSSatish Balay 127e5c89e4eSSatish Balay Level: intermediate 128e5c89e4eSSatish Balay 129e5c89e4eSSatish Balay Notes: 130e5c89e4eSSatish Balay See PetscSequentialPhaseBegin() for more details. 131e5c89e4eSSatish Balay 132e5c89e4eSSatish Balay .seealso: PetscSequentialPhaseBegin() 133e5c89e4eSSatish Balay 134e5c89e4eSSatish Balay Concepts: sequential stage 135e5c89e4eSSatish Balay 136e5c89e4eSSatish Balay @*/ 1377087cfbeSBarry Smith PetscErrorCode PetscSequentialPhaseEnd(MPI_Comm comm,int ng) 138e5c89e4eSSatish Balay { 139e5c89e4eSSatish Balay PetscErrorCode ierr; 140e5c89e4eSSatish Balay PetscMPIInt size,flag; 141e5c89e4eSSatish Balay MPI_Comm local_comm,*addr_local_comm; 142e5c89e4eSSatish Balay 143e5c89e4eSSatish Balay PetscFunctionBegin; 144e5c89e4eSSatish Balay ierr = MPI_Comm_size(comm,&size);CHKERRQ(ierr); 145e5c89e4eSSatish Balay if (size == 1) PetscFunctionReturn(0); 146e5c89e4eSSatish Balay 147e5c89e4eSSatish Balay ierr = MPI_Attr_get(comm,Petsc_Seq_keyval,(void**)&addr_local_comm,&flag);CHKERRQ(ierr); 14817186662SBarry Smith if (!flag) SETERRQ(PETSC_COMM_SELF,PETSC_ERR_ARG_INCOMP,"Wrong MPI communicator; must pass in one used with PetscSequentialPhaseBegin()"); 149e5c89e4eSSatish Balay local_comm = *addr_local_comm; 150e5c89e4eSSatish Balay 151e5c89e4eSSatish Balay ierr = PetscSequentialPhaseEnd_Private(local_comm,ng);CHKERRQ(ierr); 152e5c89e4eSSatish Balay 153e5c89e4eSSatish Balay ierr = PetscFree(addr_local_comm);CHKERRQ(ierr); 154e5c89e4eSSatish Balay ierr = MPI_Comm_free(&local_comm);CHKERRQ(ierr); 155e5c89e4eSSatish Balay ierr = MPI_Attr_delete(comm,Petsc_Seq_keyval);CHKERRQ(ierr); 156e5c89e4eSSatish Balay PetscFunctionReturn(0); 157e5c89e4eSSatish Balay } 158*6ba4152bSMatthew G. Knepley 159*6ba4152bSMatthew G. Knepley /*@C 160*6ba4152bSMatthew G. Knepley PetscGlobalMinMaxInt - Get the global min/max from local min/max input 161*6ba4152bSMatthew G. Knepley 162*6ba4152bSMatthew G. Knepley Collective on comm 163*6ba4152bSMatthew G. Knepley 164*6ba4152bSMatthew G. Knepley Input Parameter: 165*6ba4152bSMatthew G. Knepley . minMaxVal - An array with the local min and max 166*6ba4152bSMatthew G. Knepley 167*6ba4152bSMatthew G. Knepley Output Parameter: 168*6ba4152bSMatthew G. Knepley . minMaxValGlobal - An array with the global min and max 169*6ba4152bSMatthew G. Knepley 170*6ba4152bSMatthew G. Knepley Level: beginner 171*6ba4152bSMatthew G. Knepley 172*6ba4152bSMatthew G. Knepley .keywords: minimum, maximum 173*6ba4152bSMatthew G. Knepley .seealso: PetscSplitOwnership() 174*6ba4152bSMatthew G. Knepley @*/ 175*6ba4152bSMatthew G. Knepley PetscErrorCode PetscGlobalMinMaxInt(MPI_Comm comm, PetscInt minMaxVal[2], PetscInt minMaxValGlobal[2]) 176*6ba4152bSMatthew G. Knepley { 177*6ba4152bSMatthew G. Knepley PetscErrorCode ierr; 178*6ba4152bSMatthew G. Knepley 179*6ba4152bSMatthew G. Knepley PetscFunctionBegin; 180*6ba4152bSMatthew G. Knepley minMaxVal[1] = -minMaxVal[1]; 181*6ba4152bSMatthew G. Knepley ierr = MPI_Allreduce(minMaxVal, minMaxValGlobal, 2, MPIU_INT, MPI_MIN, comm);CHKERRQ(ierr); 182*6ba4152bSMatthew G. Knepley minMaxValGlobal[1] = -minMaxValGlobal[1]; 183*6ba4152bSMatthew G. Knepley PetscFunctionReturn(0); 184*6ba4152bSMatthew G. Knepley } 185*6ba4152bSMatthew G. Knepley 186*6ba4152bSMatthew G. Knepley /*@C 187*6ba4152bSMatthew G. Knepley PetscGlobalMinMaxReal - Get the global min/max from local min/max input 188*6ba4152bSMatthew G. Knepley 189*6ba4152bSMatthew G. Knepley Collective on comm 190*6ba4152bSMatthew G. Knepley 191*6ba4152bSMatthew G. Knepley Input Parameter: 192*6ba4152bSMatthew G. Knepley . minMaxVal - An array with the local min and max 193*6ba4152bSMatthew G. Knepley 194*6ba4152bSMatthew G. Knepley Output Parameter: 195*6ba4152bSMatthew G. Knepley . minMaxValGlobal - An array with the global min and max 196*6ba4152bSMatthew G. Knepley 197*6ba4152bSMatthew G. Knepley Level: beginner 198*6ba4152bSMatthew G. Knepley 199*6ba4152bSMatthew G. Knepley .keywords: minimum, maximum 200*6ba4152bSMatthew G. Knepley .seealso: PetscSplitOwnership() 201*6ba4152bSMatthew G. Knepley @*/ 202*6ba4152bSMatthew G. Knepley PetscErrorCode PetscGlobalMinMaxReal(MPI_Comm comm, PetscReal minMaxVal[2], PetscReal minMaxValGlobal[2]) 203*6ba4152bSMatthew G. Knepley { 204*6ba4152bSMatthew G. Knepley PetscErrorCode ierr; 205*6ba4152bSMatthew G. Knepley 206*6ba4152bSMatthew G. Knepley PetscFunctionBegin; 207*6ba4152bSMatthew G. Knepley minMaxVal[1] = -minMaxVal[1]; 208*6ba4152bSMatthew G. Knepley ierr = MPI_Allreduce(minMaxVal, minMaxValGlobal, 2, MPIU_REAL, MPI_MIN, comm);CHKERRQ(ierr); 209*6ba4152bSMatthew G. Knepley minMaxValGlobal[1] = -minMaxValGlobal[1]; 210*6ba4152bSMatthew G. Knepley PetscFunctionReturn(0); 211*6ba4152bSMatthew G. Knepley } 212