1c4762a1bSJed Brown 2c4762a1bSJed Brown static char help[] = "Solves a time-dependent nonlinear PDE with lower and upper bounds on the interior grid points. Uses implicit\n\ 3c4762a1bSJed Brown timestepping. Runtime options include:\n\ 4c4762a1bSJed Brown -M <xg>, where <xg> = number of grid points\n\ 5c4762a1bSJed Brown -debug : Activate debugging printouts\n\ 6c4762a1bSJed Brown -nox : Deactivate x-window graphics\n\ 7c4762a1bSJed Brown -ul : lower bound\n\ 8c4762a1bSJed Brown -uh : upper bound\n\n"; 9c4762a1bSJed Brown 10c4762a1bSJed Brown /* ------------------------------------------------------------------------ 11c4762a1bSJed Brown 12c4762a1bSJed Brown This is a variation of ex2.c to solve the PDE 13c4762a1bSJed Brown 14c4762a1bSJed Brown u * u_xx 15c4762a1bSJed Brown u_t = --------- 16c4762a1bSJed Brown 2*(t+1)^2 17c4762a1bSJed Brown 18c4762a1bSJed Brown with box constraints on the interior grid points 19c4762a1bSJed Brown ul <= u(t,x) <= uh with x != 0,1 20c4762a1bSJed Brown on the domain 0 <= x <= 1, with boundary conditions 21c4762a1bSJed Brown u(t,0) = t + 1, u(t,1) = 2*t + 2, 22c4762a1bSJed Brown and initial condition 23c4762a1bSJed Brown u(0,x) = 1 + x*x. 24c4762a1bSJed Brown 25c4762a1bSJed Brown The exact solution is: 26c4762a1bSJed Brown u(t,x) = (1 + x*x) * (1 + t) 27c4762a1bSJed Brown 28c4762a1bSJed Brown We use by default the backward Euler method. 29c4762a1bSJed Brown 30c4762a1bSJed Brown ------------------------------------------------------------------------- */ 31c4762a1bSJed Brown 32c4762a1bSJed Brown /* 33c4762a1bSJed Brown Include "petscts.h" to use the PETSc timestepping routines. Note that 34c4762a1bSJed Brown this file automatically includes "petscsys.h" and other lower-level 35c4762a1bSJed Brown PETSc include files. 36c4762a1bSJed Brown 37c4762a1bSJed Brown Include the "petscdmda.h" to allow us to use the distributed array data 38c4762a1bSJed Brown structures to manage the parallel grid. 39c4762a1bSJed Brown */ 40c4762a1bSJed Brown #include <petscts.h> 41c4762a1bSJed Brown #include <petscdm.h> 42c4762a1bSJed Brown #include <petscdmda.h> 43c4762a1bSJed Brown #include <petscdraw.h> 44c4762a1bSJed Brown 45c4762a1bSJed Brown /* 46c4762a1bSJed Brown User-defined application context - contains data needed by the 47c4762a1bSJed Brown application-provided callback routines. 48c4762a1bSJed Brown */ 49c4762a1bSJed Brown typedef struct { 50c4762a1bSJed Brown MPI_Comm comm; /* communicator */ 51c4762a1bSJed Brown DM da; /* distributed array data structure */ 52c4762a1bSJed Brown Vec localwork; /* local ghosted work vector */ 53c4762a1bSJed Brown Vec u_local; /* local ghosted approximate solution vector */ 54c4762a1bSJed Brown Vec solution; /* global exact solution vector */ 55c4762a1bSJed Brown PetscInt m; /* total number of grid points */ 56c4762a1bSJed Brown PetscReal h; /* mesh width: h = 1/(m-1) */ 57c4762a1bSJed Brown PetscBool debug; /* flag (1 indicates activation of debugging printouts) */ 58c4762a1bSJed Brown } AppCtx; 59c4762a1bSJed Brown 60c4762a1bSJed Brown /* 61c4762a1bSJed Brown User-defined routines, provided below. 62c4762a1bSJed Brown */ 63c4762a1bSJed Brown extern PetscErrorCode InitialConditions(Vec, AppCtx *); 64c4762a1bSJed Brown extern PetscErrorCode RHSFunction(TS, PetscReal, Vec, Vec, void *); 65c4762a1bSJed Brown extern PetscErrorCode RHSJacobian(TS, PetscReal, Vec, Mat, Mat, void *); 66c4762a1bSJed Brown extern PetscErrorCode Monitor(TS, PetscInt, PetscReal, Vec, void *); 67c4762a1bSJed Brown extern PetscErrorCode ExactSolution(PetscReal, Vec, AppCtx *); 68c4762a1bSJed Brown extern PetscErrorCode SetBounds(Vec, Vec, PetscScalar, PetscScalar, AppCtx *); 69c4762a1bSJed Brown 70*d71ae5a4SJacob Faibussowitsch int main(int argc, char **argv) 71*d71ae5a4SJacob Faibussowitsch { 72c4762a1bSJed Brown AppCtx appctx; /* user-defined application context */ 73c4762a1bSJed Brown TS ts; /* timestepping context */ 74c4762a1bSJed Brown Mat A; /* Jacobian matrix data structure */ 75c4762a1bSJed Brown Vec u; /* approximate solution vector */ 76c4762a1bSJed Brown Vec r; /* residual vector */ 77c4762a1bSJed Brown PetscInt time_steps_max = 1000; /* default max timesteps */ 78c4762a1bSJed Brown PetscReal dt; 79c4762a1bSJed Brown PetscReal time_total_max = 100.0; /* default max total time */ 80c4762a1bSJed Brown Vec xl, xu; /* Lower and upper bounds on variables */ 81c4762a1bSJed Brown PetscScalar ul = 0.0, uh = 3.0; 82c4762a1bSJed Brown PetscBool mymonitor; 83c4762a1bSJed Brown PetscReal bounds[] = {1.0, 3.3}; 84c4762a1bSJed Brown 85c4762a1bSJed Brown /* - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - 86c4762a1bSJed Brown Initialize program and set problem parameters 87c4762a1bSJed Brown - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - */ 88c4762a1bSJed Brown 89327415f7SBarry Smith PetscFunctionBeginUser; 909566063dSJacob Faibussowitsch PetscCall(PetscInitialize(&argc, &argv, (char *)0, help)); 919566063dSJacob Faibussowitsch PetscCall(PetscViewerDrawSetBounds(PETSC_VIEWER_DRAW_(PETSC_COMM_WORLD), 1, bounds)); 92c4762a1bSJed Brown 93c4762a1bSJed Brown appctx.comm = PETSC_COMM_WORLD; 94c4762a1bSJed Brown appctx.m = 60; 959566063dSJacob Faibussowitsch PetscCall(PetscOptionsGetInt(NULL, NULL, "-M", &appctx.m, NULL)); 969566063dSJacob Faibussowitsch PetscCall(PetscOptionsGetScalar(NULL, NULL, "-ul", &ul, NULL)); 979566063dSJacob Faibussowitsch PetscCall(PetscOptionsGetScalar(NULL, NULL, "-uh", &uh, NULL)); 989566063dSJacob Faibussowitsch PetscCall(PetscOptionsHasName(NULL, NULL, "-debug", &appctx.debug)); 999566063dSJacob Faibussowitsch PetscCall(PetscOptionsHasName(NULL, NULL, "-mymonitor", &mymonitor)); 100c4762a1bSJed Brown appctx.h = 1.0 / (appctx.m - 1.0); 101c4762a1bSJed Brown 102c4762a1bSJed Brown /* - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - 103c4762a1bSJed Brown Create vector data structures 104c4762a1bSJed Brown - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - */ 105c4762a1bSJed Brown 106c4762a1bSJed Brown /* 107c4762a1bSJed Brown Create distributed array (DMDA) to manage parallel grid and vectors 108c4762a1bSJed Brown and to set up the ghost point communication pattern. There are M 109c4762a1bSJed Brown total grid values spread equally among all the processors. 110c4762a1bSJed Brown */ 1119566063dSJacob Faibussowitsch PetscCall(DMDACreate1d(PETSC_COMM_WORLD, DM_BOUNDARY_NONE, appctx.m, 1, 1, NULL, &appctx.da)); 1129566063dSJacob Faibussowitsch PetscCall(DMSetFromOptions(appctx.da)); 1139566063dSJacob Faibussowitsch PetscCall(DMSetUp(appctx.da)); 114c4762a1bSJed Brown 115c4762a1bSJed Brown /* 116c4762a1bSJed Brown Extract global and local vectors from DMDA; we use these to store the 117c4762a1bSJed Brown approximate solution. Then duplicate these for remaining vectors that 118c4762a1bSJed Brown have the same types. 119c4762a1bSJed Brown */ 1209566063dSJacob Faibussowitsch PetscCall(DMCreateGlobalVector(appctx.da, &u)); 1219566063dSJacob Faibussowitsch PetscCall(DMCreateLocalVector(appctx.da, &appctx.u_local)); 122c4762a1bSJed Brown 123c4762a1bSJed Brown /* 124c4762a1bSJed Brown Create local work vector for use in evaluating right-hand-side function; 125c4762a1bSJed Brown create global work vector for storing exact solution. 126c4762a1bSJed Brown */ 1279566063dSJacob Faibussowitsch PetscCall(VecDuplicate(appctx.u_local, &appctx.localwork)); 1289566063dSJacob Faibussowitsch PetscCall(VecDuplicate(u, &appctx.solution)); 129c4762a1bSJed Brown 130c4762a1bSJed Brown /* Create residual vector */ 1319566063dSJacob Faibussowitsch PetscCall(VecDuplicate(u, &r)); 132c4762a1bSJed Brown /* Create lower and upper bound vectors */ 1339566063dSJacob Faibussowitsch PetscCall(VecDuplicate(u, &xl)); 1349566063dSJacob Faibussowitsch PetscCall(VecDuplicate(u, &xu)); 1359566063dSJacob Faibussowitsch PetscCall(SetBounds(xl, xu, ul, uh, &appctx)); 136c4762a1bSJed Brown 137c4762a1bSJed Brown /* - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - 138c4762a1bSJed Brown Create timestepping solver context; set callback routine for 139c4762a1bSJed Brown right-hand-side function evaluation. 140c4762a1bSJed Brown - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - */ 141c4762a1bSJed Brown 1429566063dSJacob Faibussowitsch PetscCall(TSCreate(PETSC_COMM_WORLD, &ts)); 1439566063dSJacob Faibussowitsch PetscCall(TSSetProblemType(ts, TS_NONLINEAR)); 1449566063dSJacob Faibussowitsch PetscCall(TSSetRHSFunction(ts, r, RHSFunction, &appctx)); 145c4762a1bSJed Brown 146c4762a1bSJed Brown /* - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - 147c4762a1bSJed Brown Set optional user-defined monitoring routine 148c4762a1bSJed Brown - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - */ 149c4762a1bSJed Brown 15048a46eb9SPierre Jolivet if (mymonitor) PetscCall(TSMonitorSet(ts, Monitor, &appctx, NULL)); 151c4762a1bSJed Brown 152c4762a1bSJed Brown /* - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - 153c4762a1bSJed Brown For nonlinear problems, the user can provide a Jacobian evaluation 154c4762a1bSJed Brown routine (or use a finite differencing approximation). 155c4762a1bSJed Brown 156c4762a1bSJed Brown Create matrix data structure; set Jacobian evaluation routine. 157c4762a1bSJed Brown - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - */ 158c4762a1bSJed Brown 1599566063dSJacob Faibussowitsch PetscCall(MatCreate(PETSC_COMM_WORLD, &A)); 1609566063dSJacob Faibussowitsch PetscCall(MatSetSizes(A, PETSC_DECIDE, PETSC_DECIDE, appctx.m, appctx.m)); 1619566063dSJacob Faibussowitsch PetscCall(MatSetFromOptions(A)); 1629566063dSJacob Faibussowitsch PetscCall(MatSetUp(A)); 1639566063dSJacob Faibussowitsch PetscCall(TSSetRHSJacobian(ts, A, A, RHSJacobian, &appctx)); 164c4762a1bSJed Brown 165c4762a1bSJed Brown /* - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - 166c4762a1bSJed Brown Set solution vector and initial timestep 167c4762a1bSJed Brown - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - */ 168c4762a1bSJed Brown 169c4762a1bSJed Brown dt = appctx.h / 2.0; 1709566063dSJacob Faibussowitsch PetscCall(TSSetTimeStep(ts, dt)); 171c4762a1bSJed Brown 172c4762a1bSJed Brown /* - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - 173c4762a1bSJed Brown Customize timestepping solver: 174c4762a1bSJed Brown - Set the solution method to be the Backward Euler method. 175c4762a1bSJed Brown - Set timestepping duration info 176c4762a1bSJed Brown Then set runtime options, which can override these defaults. 177c4762a1bSJed Brown For example, 178c4762a1bSJed Brown -ts_max_steps <maxsteps> -ts_max_time <maxtime> 179c4762a1bSJed Brown to override the defaults set by TSSetMaxSteps()/TSSetMaxTime(). 180c4762a1bSJed Brown - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - */ 181c4762a1bSJed Brown 1829566063dSJacob Faibussowitsch PetscCall(TSSetType(ts, TSBEULER)); 1839566063dSJacob Faibussowitsch PetscCall(TSSetMaxSteps(ts, time_steps_max)); 1849566063dSJacob Faibussowitsch PetscCall(TSSetMaxTime(ts, time_total_max)); 1859566063dSJacob Faibussowitsch PetscCall(TSSetExactFinalTime(ts, TS_EXACTFINALTIME_STEPOVER)); 186c4762a1bSJed Brown /* Set lower and upper bound on the solution vector for each time step */ 1879566063dSJacob Faibussowitsch PetscCall(TSVISetVariableBounds(ts, xl, xu)); 1889566063dSJacob Faibussowitsch PetscCall(TSSetFromOptions(ts)); 189c4762a1bSJed Brown 190c4762a1bSJed Brown /* - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - 191c4762a1bSJed Brown Solve the problem 192c4762a1bSJed Brown - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - */ 193c4762a1bSJed Brown 194c4762a1bSJed Brown /* 195c4762a1bSJed Brown Evaluate initial conditions 196c4762a1bSJed Brown */ 1979566063dSJacob Faibussowitsch PetscCall(InitialConditions(u, &appctx)); 198c4762a1bSJed Brown 199c4762a1bSJed Brown /* 200c4762a1bSJed Brown Run the timestepping solver 201c4762a1bSJed Brown */ 2029566063dSJacob Faibussowitsch PetscCall(TSSolve(ts, u)); 203c4762a1bSJed Brown 204c4762a1bSJed Brown /* - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - 205c4762a1bSJed Brown Free work space. All PETSc objects should be destroyed when they 206c4762a1bSJed Brown are no longer needed. 207c4762a1bSJed Brown - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - */ 208c4762a1bSJed Brown 2099566063dSJacob Faibussowitsch PetscCall(VecDestroy(&r)); 2109566063dSJacob Faibussowitsch PetscCall(VecDestroy(&xl)); 2119566063dSJacob Faibussowitsch PetscCall(VecDestroy(&xu)); 2129566063dSJacob Faibussowitsch PetscCall(TSDestroy(&ts)); 2139566063dSJacob Faibussowitsch PetscCall(VecDestroy(&u)); 2149566063dSJacob Faibussowitsch PetscCall(MatDestroy(&A)); 2159566063dSJacob Faibussowitsch PetscCall(DMDestroy(&appctx.da)); 2169566063dSJacob Faibussowitsch PetscCall(VecDestroy(&appctx.localwork)); 2179566063dSJacob Faibussowitsch PetscCall(VecDestroy(&appctx.solution)); 2189566063dSJacob Faibussowitsch PetscCall(VecDestroy(&appctx.u_local)); 219c4762a1bSJed Brown 220c4762a1bSJed Brown /* 221c4762a1bSJed Brown Always call PetscFinalize() before exiting a program. This routine 222c4762a1bSJed Brown - finalizes the PETSc libraries as well as MPI 223c4762a1bSJed Brown - provides summary and diagnostic information if certain runtime 224c4762a1bSJed Brown options are chosen (e.g., -log_view). 225c4762a1bSJed Brown */ 2269566063dSJacob Faibussowitsch PetscCall(PetscFinalize()); 227b122ec5aSJacob Faibussowitsch return 0; 228c4762a1bSJed Brown } 229c4762a1bSJed Brown /* --------------------------------------------------------------------- */ 230c4762a1bSJed Brown /* 231c4762a1bSJed Brown InitialConditions - Computes the solution at the initial time. 232c4762a1bSJed Brown 233c4762a1bSJed Brown Input Parameters: 234c4762a1bSJed Brown u - uninitialized solution vector (global) 235c4762a1bSJed Brown appctx - user-defined application context 236c4762a1bSJed Brown 237c4762a1bSJed Brown Output Parameter: 238c4762a1bSJed Brown u - vector with solution at initial time (global) 239c4762a1bSJed Brown */ 240*d71ae5a4SJacob Faibussowitsch PetscErrorCode InitialConditions(Vec u, AppCtx *appctx) 241*d71ae5a4SJacob Faibussowitsch { 242c4762a1bSJed Brown PetscScalar *u_localptr, h = appctx->h, x; 243c4762a1bSJed Brown PetscInt i, mybase, myend; 244c4762a1bSJed Brown 245c4762a1bSJed Brown /* 246c4762a1bSJed Brown Determine starting point of each processor's range of 247c4762a1bSJed Brown grid values. 248c4762a1bSJed Brown */ 2499566063dSJacob Faibussowitsch PetscCall(VecGetOwnershipRange(u, &mybase, &myend)); 250c4762a1bSJed Brown 251c4762a1bSJed Brown /* 252c4762a1bSJed Brown Get a pointer to vector data. 253c4762a1bSJed Brown - For default PETSc vectors, VecGetArray() returns a pointer to 254c4762a1bSJed Brown the data array. Otherwise, the routine is implementation dependent. 255c4762a1bSJed Brown - You MUST call VecRestoreArray() when you no longer need access to 256c4762a1bSJed Brown the array. 257c4762a1bSJed Brown - Note that the Fortran interface to VecGetArray() differs from the 258c4762a1bSJed Brown C version. See the users manual for details. 259c4762a1bSJed Brown */ 2609566063dSJacob Faibussowitsch PetscCall(VecGetArray(u, &u_localptr)); 261c4762a1bSJed Brown 262c4762a1bSJed Brown /* 263c4762a1bSJed Brown We initialize the solution array by simply writing the solution 264c4762a1bSJed Brown directly into the array locations. Alternatively, we could use 265c4762a1bSJed Brown VecSetValues() or VecSetValuesLocal(). 266c4762a1bSJed Brown */ 267c4762a1bSJed Brown for (i = mybase; i < myend; i++) { 268c4762a1bSJed Brown x = h * (PetscReal)i; /* current location in global grid */ 269c4762a1bSJed Brown u_localptr[i - mybase] = 1.0 + x * x; 270c4762a1bSJed Brown } 271c4762a1bSJed Brown 272c4762a1bSJed Brown /* 273c4762a1bSJed Brown Restore vector 274c4762a1bSJed Brown */ 2759566063dSJacob Faibussowitsch PetscCall(VecRestoreArray(u, &u_localptr)); 276c4762a1bSJed Brown 277c4762a1bSJed Brown /* 278c4762a1bSJed Brown Print debugging information if desired 279c4762a1bSJed Brown */ 280c4762a1bSJed Brown if (appctx->debug) { 2819566063dSJacob Faibussowitsch PetscCall(PetscPrintf(appctx->comm, "initial guess vector\n")); 2829566063dSJacob Faibussowitsch PetscCall(VecView(u, PETSC_VIEWER_STDOUT_WORLD)); 283c4762a1bSJed Brown } 284c4762a1bSJed Brown 285c4762a1bSJed Brown return 0; 286c4762a1bSJed Brown } 287c4762a1bSJed Brown 288c4762a1bSJed Brown /* --------------------------------------------------------------------- */ 289c4762a1bSJed Brown /* 290c4762a1bSJed Brown SetBounds - Sets the lower and uper bounds on the interior points 291c4762a1bSJed Brown 292c4762a1bSJed Brown Input parameters: 293c4762a1bSJed Brown xl - vector of lower bounds 294c4762a1bSJed Brown xu - vector of upper bounds 295c4762a1bSJed Brown ul - constant lower bound for all variables 296c4762a1bSJed Brown uh - constant upper bound for all variables 297c4762a1bSJed Brown appctx - Application context 298c4762a1bSJed Brown */ 299*d71ae5a4SJacob Faibussowitsch PetscErrorCode SetBounds(Vec xl, Vec xu, PetscScalar ul, PetscScalar uh, AppCtx *appctx) 300*d71ae5a4SJacob Faibussowitsch { 301c4762a1bSJed Brown PetscScalar *l, *u; 302c4762a1bSJed Brown PetscMPIInt rank, size; 303c4762a1bSJed Brown PetscInt localsize; 304c4762a1bSJed Brown 305c4762a1bSJed Brown PetscFunctionBeginUser; 3069566063dSJacob Faibussowitsch PetscCall(VecSet(xl, ul)); 3079566063dSJacob Faibussowitsch PetscCall(VecSet(xu, uh)); 3089566063dSJacob Faibussowitsch PetscCall(VecGetLocalSize(xl, &localsize)); 3099566063dSJacob Faibussowitsch PetscCall(VecGetArray(xl, &l)); 3109566063dSJacob Faibussowitsch PetscCall(VecGetArray(xu, &u)); 311c4762a1bSJed Brown 3129566063dSJacob Faibussowitsch PetscCallMPI(MPI_Comm_rank(appctx->comm, &rank)); 3139566063dSJacob Faibussowitsch PetscCallMPI(MPI_Comm_size(appctx->comm, &size)); 314dd400576SPatrick Sanan if (rank == 0) { 315c4762a1bSJed Brown l[0] = -PETSC_INFINITY; 316c4762a1bSJed Brown u[0] = PETSC_INFINITY; 317c4762a1bSJed Brown } 318c4762a1bSJed Brown if (rank == size - 1) { 319c4762a1bSJed Brown l[localsize - 1] = -PETSC_INFINITY; 320c4762a1bSJed Brown u[localsize - 1] = PETSC_INFINITY; 321c4762a1bSJed Brown } 3229566063dSJacob Faibussowitsch PetscCall(VecRestoreArray(xl, &l)); 3239566063dSJacob Faibussowitsch PetscCall(VecRestoreArray(xu, &u)); 324c4762a1bSJed Brown PetscFunctionReturn(0); 325c4762a1bSJed Brown } 326c4762a1bSJed Brown 327c4762a1bSJed Brown /* --------------------------------------------------------------------- */ 328c4762a1bSJed Brown /* 329c4762a1bSJed Brown ExactSolution - Computes the exact solution at a given time. 330c4762a1bSJed Brown 331c4762a1bSJed Brown Input Parameters: 332c4762a1bSJed Brown t - current time 333c4762a1bSJed Brown solution - vector in which exact solution will be computed 334c4762a1bSJed Brown appctx - user-defined application context 335c4762a1bSJed Brown 336c4762a1bSJed Brown Output Parameter: 337c4762a1bSJed Brown solution - vector with the newly computed exact solution 338c4762a1bSJed Brown */ 339*d71ae5a4SJacob Faibussowitsch PetscErrorCode ExactSolution(PetscReal t, Vec solution, AppCtx *appctx) 340*d71ae5a4SJacob Faibussowitsch { 341c4762a1bSJed Brown PetscScalar *s_localptr, h = appctx->h, x; 342c4762a1bSJed Brown PetscInt i, mybase, myend; 343c4762a1bSJed Brown 344c4762a1bSJed Brown /* 345c4762a1bSJed Brown Determine starting and ending points of each processor's 346c4762a1bSJed Brown range of grid values 347c4762a1bSJed Brown */ 3489566063dSJacob Faibussowitsch PetscCall(VecGetOwnershipRange(solution, &mybase, &myend)); 349c4762a1bSJed Brown 350c4762a1bSJed Brown /* 351c4762a1bSJed Brown Get a pointer to vector data. 352c4762a1bSJed Brown */ 3539566063dSJacob Faibussowitsch PetscCall(VecGetArray(solution, &s_localptr)); 354c4762a1bSJed Brown 355c4762a1bSJed Brown /* 356c4762a1bSJed Brown Simply write the solution directly into the array locations. 357c4762a1bSJed Brown Alternatively, we could use VecSetValues() or VecSetValuesLocal(). 358c4762a1bSJed Brown */ 359c4762a1bSJed Brown for (i = mybase; i < myend; i++) { 360c4762a1bSJed Brown x = h * (PetscReal)i; 361c4762a1bSJed Brown s_localptr[i - mybase] = (t + 1.0) * (1.0 + x * x); 362c4762a1bSJed Brown } 363c4762a1bSJed Brown 364c4762a1bSJed Brown /* 365c4762a1bSJed Brown Restore vector 366c4762a1bSJed Brown */ 3679566063dSJacob Faibussowitsch PetscCall(VecRestoreArray(solution, &s_localptr)); 368c4762a1bSJed Brown return 0; 369c4762a1bSJed Brown } 370c4762a1bSJed Brown /* --------------------------------------------------------------------- */ 371c4762a1bSJed Brown /* 372c4762a1bSJed Brown Monitor - User-provided routine to monitor the solution computed at 373c4762a1bSJed Brown each timestep. This example plots the solution and computes the 374c4762a1bSJed Brown error in two different norms. 375c4762a1bSJed Brown 376c4762a1bSJed Brown Input Parameters: 377c4762a1bSJed Brown ts - the timestep context 378c4762a1bSJed Brown step - the count of the current step (with 0 meaning the 379c4762a1bSJed Brown initial condition) 380c4762a1bSJed Brown time - the current time 381c4762a1bSJed Brown u - the solution at this timestep 382c4762a1bSJed Brown ctx - the user-provided context for this monitoring routine. 383c4762a1bSJed Brown In this case we use the application context which contains 384c4762a1bSJed Brown information about the problem size, workspace and the exact 385c4762a1bSJed Brown solution. 386c4762a1bSJed Brown */ 387*d71ae5a4SJacob Faibussowitsch PetscErrorCode Monitor(TS ts, PetscInt step, PetscReal time, Vec u, void *ctx) 388*d71ae5a4SJacob Faibussowitsch { 389c4762a1bSJed Brown AppCtx *appctx = (AppCtx *)ctx; /* user-defined application context */ 390c4762a1bSJed Brown PetscReal en2, en2s, enmax; 391c4762a1bSJed Brown PetscDraw draw; 392c4762a1bSJed Brown 393c4762a1bSJed Brown /* 394c4762a1bSJed Brown We use the default X windows viewer 395c4762a1bSJed Brown PETSC_VIEWER_DRAW_(appctx->comm) 396c4762a1bSJed Brown that is associated with the current communicator. This saves 397c4762a1bSJed Brown the effort of calling PetscViewerDrawOpen() to create the window. 398c4762a1bSJed Brown Note that if we wished to plot several items in separate windows we 399c4762a1bSJed Brown would create each viewer with PetscViewerDrawOpen() and store them in 400c4762a1bSJed Brown the application context, appctx. 401c4762a1bSJed Brown 402c4762a1bSJed Brown PetscReal buffering makes graphics look better. 403c4762a1bSJed Brown */ 4049566063dSJacob Faibussowitsch PetscCall(PetscViewerDrawGetDraw(PETSC_VIEWER_DRAW_(appctx->comm), 0, &draw)); 4059566063dSJacob Faibussowitsch PetscCall(PetscDrawSetDoubleBuffer(draw)); 4069566063dSJacob Faibussowitsch PetscCall(VecView(u, PETSC_VIEWER_DRAW_(appctx->comm))); 407c4762a1bSJed Brown 408c4762a1bSJed Brown /* 409c4762a1bSJed Brown Compute the exact solution at this timestep 410c4762a1bSJed Brown */ 4119566063dSJacob Faibussowitsch PetscCall(ExactSolution(time, appctx->solution, appctx)); 412c4762a1bSJed Brown 413c4762a1bSJed Brown /* 414c4762a1bSJed Brown Print debugging information if desired 415c4762a1bSJed Brown */ 416c4762a1bSJed Brown if (appctx->debug) { 4179566063dSJacob Faibussowitsch PetscCall(PetscPrintf(appctx->comm, "Computed solution vector\n")); 4189566063dSJacob Faibussowitsch PetscCall(VecView(u, PETSC_VIEWER_STDOUT_WORLD)); 4199566063dSJacob Faibussowitsch PetscCall(PetscPrintf(appctx->comm, "Exact solution vector\n")); 4209566063dSJacob Faibussowitsch PetscCall(VecView(appctx->solution, PETSC_VIEWER_STDOUT_WORLD)); 421c4762a1bSJed Brown } 422c4762a1bSJed Brown 423c4762a1bSJed Brown /* 424c4762a1bSJed Brown Compute the 2-norm and max-norm of the error 425c4762a1bSJed Brown */ 4269566063dSJacob Faibussowitsch PetscCall(VecAXPY(appctx->solution, -1.0, u)); 4279566063dSJacob Faibussowitsch PetscCall(VecNorm(appctx->solution, NORM_2, &en2)); 428c4762a1bSJed Brown en2s = PetscSqrtReal(appctx->h) * en2; /* scale the 2-norm by the grid spacing */ 4299566063dSJacob Faibussowitsch PetscCall(VecNorm(appctx->solution, NORM_MAX, &enmax)); 430c4762a1bSJed Brown 431c4762a1bSJed Brown /* 432c4762a1bSJed Brown PetscPrintf() causes only the first processor in this 433c4762a1bSJed Brown communicator to print the timestep information. 434c4762a1bSJed Brown */ 43563a3b9bcSJacob Faibussowitsch PetscCall(PetscPrintf(appctx->comm, "Timestep %" PetscInt_FMT ": time = %g,2-norm error = %g, max norm error = %g\n", step, (double)time, (double)en2s, (double)enmax)); 436c4762a1bSJed Brown 437c4762a1bSJed Brown /* 438c4762a1bSJed Brown Print debugging information if desired 439c4762a1bSJed Brown */ 440c4762a1bSJed Brown /* if (appctx->debug) { 4419566063dSJacob Faibussowitsch PetscCall(PetscPrintf(appctx->comm,"Error vector\n")); 4429566063dSJacob Faibussowitsch PetscCall(VecView(appctx->solution,PETSC_VIEWER_STDOUT_WORLD)); 443c4762a1bSJed Brown } */ 444c4762a1bSJed Brown return 0; 445c4762a1bSJed Brown } 446c4762a1bSJed Brown /* --------------------------------------------------------------------- */ 447c4762a1bSJed Brown /* 448c4762a1bSJed Brown RHSFunction - User-provided routine that evalues the right-hand-side 449c4762a1bSJed Brown function of the ODE. This routine is set in the main program by 450c4762a1bSJed Brown calling TSSetRHSFunction(). We compute: 451c4762a1bSJed Brown global_out = F(global_in) 452c4762a1bSJed Brown 453c4762a1bSJed Brown Input Parameters: 454c4762a1bSJed Brown ts - timesteping context 455c4762a1bSJed Brown t - current time 456c4762a1bSJed Brown global_in - vector containing the current iterate 457c4762a1bSJed Brown ctx - (optional) user-provided context for function evaluation. 458c4762a1bSJed Brown In this case we use the appctx defined above. 459c4762a1bSJed Brown 460c4762a1bSJed Brown Output Parameter: 461c4762a1bSJed Brown global_out - vector containing the newly evaluated function 462c4762a1bSJed Brown */ 463*d71ae5a4SJacob Faibussowitsch PetscErrorCode RHSFunction(TS ts, PetscReal t, Vec global_in, Vec global_out, void *ctx) 464*d71ae5a4SJacob Faibussowitsch { 465c4762a1bSJed Brown AppCtx *appctx = (AppCtx *)ctx; /* user-defined application context */ 466c4762a1bSJed Brown DM da = appctx->da; /* distributed array */ 467c4762a1bSJed Brown Vec local_in = appctx->u_local; /* local ghosted input vector */ 468c4762a1bSJed Brown Vec localwork = appctx->localwork; /* local ghosted work vector */ 469c4762a1bSJed Brown PetscInt i, localsize; 470c4762a1bSJed Brown PetscMPIInt rank, size; 471c4762a1bSJed Brown PetscScalar *copyptr, sc; 472c4762a1bSJed Brown const PetscScalar *localptr; 473c4762a1bSJed Brown 474c4762a1bSJed Brown /* - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - 475c4762a1bSJed Brown Get ready for local function computations 476c4762a1bSJed Brown - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - */ 477c4762a1bSJed Brown /* 478c4762a1bSJed Brown Scatter ghost points to local vector, using the 2-step process 479c4762a1bSJed Brown DMGlobalToLocalBegin(), DMGlobalToLocalEnd(). 480c4762a1bSJed Brown By placing code between these two statements, computations can be 481c4762a1bSJed Brown done while messages are in transition. 482c4762a1bSJed Brown */ 4839566063dSJacob Faibussowitsch PetscCall(DMGlobalToLocalBegin(da, global_in, INSERT_VALUES, local_in)); 4849566063dSJacob Faibussowitsch PetscCall(DMGlobalToLocalEnd(da, global_in, INSERT_VALUES, local_in)); 485c4762a1bSJed Brown 486c4762a1bSJed Brown /* 487c4762a1bSJed Brown Access directly the values in our local INPUT work array 488c4762a1bSJed Brown */ 4899566063dSJacob Faibussowitsch PetscCall(VecGetArrayRead(local_in, &localptr)); 490c4762a1bSJed Brown 491c4762a1bSJed Brown /* 492c4762a1bSJed Brown Access directly the values in our local OUTPUT work array 493c4762a1bSJed Brown */ 4949566063dSJacob Faibussowitsch PetscCall(VecGetArray(localwork, ©ptr)); 495c4762a1bSJed Brown 496c4762a1bSJed Brown sc = 1.0 / (appctx->h * appctx->h * 2.0 * (1.0 + t) * (1.0 + t)); 497c4762a1bSJed Brown 498c4762a1bSJed Brown /* 499c4762a1bSJed Brown Evaluate our function on the nodes owned by this processor 500c4762a1bSJed Brown */ 5019566063dSJacob Faibussowitsch PetscCall(VecGetLocalSize(local_in, &localsize)); 502c4762a1bSJed Brown 503c4762a1bSJed Brown /* - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - 504c4762a1bSJed Brown Compute entries for the locally owned part 505c4762a1bSJed Brown - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - */ 506c4762a1bSJed Brown 507c4762a1bSJed Brown /* 508c4762a1bSJed Brown Handle boundary conditions: This is done by using the boundary condition 509c4762a1bSJed Brown u(t,boundary) = g(t,boundary) 510c4762a1bSJed Brown for some function g. Now take the derivative with respect to t to obtain 511c4762a1bSJed Brown u_{t}(t,boundary) = g_{t}(t,boundary) 512c4762a1bSJed Brown 513c4762a1bSJed Brown In our case, u(t,0) = t + 1, so that u_{t}(t,0) = 1 514c4762a1bSJed Brown and u(t,1) = 2t+ 2, so that u_{t}(t,1) = 2 515c4762a1bSJed Brown */ 5169566063dSJacob Faibussowitsch PetscCallMPI(MPI_Comm_rank(appctx->comm, &rank)); 5179566063dSJacob Faibussowitsch PetscCallMPI(MPI_Comm_size(appctx->comm, &size)); 518dd400576SPatrick Sanan if (rank == 0) copyptr[0] = 1.0; 519c4762a1bSJed Brown if (rank == size - 1) copyptr[localsize - 1] = (t < .5) ? 2.0 : 0.0; 520c4762a1bSJed Brown 521c4762a1bSJed Brown /* 522c4762a1bSJed Brown Handle the interior nodes where the PDE is replace by finite 523c4762a1bSJed Brown difference operators. 524c4762a1bSJed Brown */ 525c4762a1bSJed Brown for (i = 1; i < localsize - 1; i++) copyptr[i] = localptr[i] * sc * (localptr[i + 1] + localptr[i - 1] - 2.0 * localptr[i]); 526c4762a1bSJed Brown 527c4762a1bSJed Brown /* 528c4762a1bSJed Brown Restore vectors 529c4762a1bSJed Brown */ 5309566063dSJacob Faibussowitsch PetscCall(VecRestoreArrayRead(local_in, &localptr)); 5319566063dSJacob Faibussowitsch PetscCall(VecRestoreArray(localwork, ©ptr)); 532c4762a1bSJed Brown 533c4762a1bSJed Brown /* 534c4762a1bSJed Brown Insert values from the local OUTPUT vector into the global 535c4762a1bSJed Brown output vector 536c4762a1bSJed Brown */ 5379566063dSJacob Faibussowitsch PetscCall(DMLocalToGlobalBegin(da, localwork, INSERT_VALUES, global_out)); 5389566063dSJacob Faibussowitsch PetscCall(DMLocalToGlobalEnd(da, localwork, INSERT_VALUES, global_out)); 539c4762a1bSJed Brown 540c4762a1bSJed Brown /* Print debugging information if desired */ 541c4762a1bSJed Brown /* if (appctx->debug) { 5429566063dSJacob Faibussowitsch PetscCall(PetscPrintf(appctx->comm,"RHS function vector\n")); 5439566063dSJacob Faibussowitsch PetscCall(VecView(global_out,PETSC_VIEWER_STDOUT_WORLD)); 544c4762a1bSJed Brown } */ 545c4762a1bSJed Brown 546c4762a1bSJed Brown return 0; 547c4762a1bSJed Brown } 548c4762a1bSJed Brown /* --------------------------------------------------------------------- */ 549c4762a1bSJed Brown /* 550c4762a1bSJed Brown RHSJacobian - User-provided routine to compute the Jacobian of 551c4762a1bSJed Brown the nonlinear right-hand-side function of the ODE. 552c4762a1bSJed Brown 553c4762a1bSJed Brown Input Parameters: 554c4762a1bSJed Brown ts - the TS context 555c4762a1bSJed Brown t - current time 556c4762a1bSJed Brown global_in - global input vector 557c4762a1bSJed Brown dummy - optional user-defined context, as set by TSetRHSJacobian() 558c4762a1bSJed Brown 559c4762a1bSJed Brown Output Parameters: 560c4762a1bSJed Brown AA - Jacobian matrix 561c4762a1bSJed Brown BB - optionally different preconditioning matrix 562c4762a1bSJed Brown str - flag indicating matrix structure 563c4762a1bSJed Brown 564c4762a1bSJed Brown Notes: 565c4762a1bSJed Brown RHSJacobian computes entries for the locally owned part of the Jacobian. 566c4762a1bSJed Brown - Currently, all PETSc parallel matrix formats are partitioned by 567c4762a1bSJed Brown contiguous chunks of rows across the processors. 568c4762a1bSJed Brown - Each processor needs to insert only elements that it owns 569c4762a1bSJed Brown locally (but any non-local elements will be sent to the 570c4762a1bSJed Brown appropriate processor during matrix assembly). 571c4762a1bSJed Brown - Always specify global row and columns of matrix entries when 572c4762a1bSJed Brown using MatSetValues(). 573c4762a1bSJed Brown - Here, we set all entries for a particular row at once. 574c4762a1bSJed Brown - Note that MatSetValues() uses 0-based row and column numbers 575c4762a1bSJed Brown in Fortran as well as in C. 576c4762a1bSJed Brown */ 577*d71ae5a4SJacob Faibussowitsch PetscErrorCode RHSJacobian(TS ts, PetscReal t, Vec global_in, Mat AA, Mat B, void *ctx) 578*d71ae5a4SJacob Faibussowitsch { 579c4762a1bSJed Brown AppCtx *appctx = (AppCtx *)ctx; /* user-defined application context */ 580c4762a1bSJed Brown Vec local_in = appctx->u_local; /* local ghosted input vector */ 581c4762a1bSJed Brown DM da = appctx->da; /* distributed array */ 582c4762a1bSJed Brown PetscScalar v[3], sc; 583c4762a1bSJed Brown const PetscScalar *localptr; 584c4762a1bSJed Brown PetscInt i, mstart, mend, mstarts, mends, idx[3], is; 585c4762a1bSJed Brown 586c4762a1bSJed Brown /* - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - 587c4762a1bSJed Brown Get ready for local Jacobian computations 588c4762a1bSJed Brown - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - */ 589c4762a1bSJed Brown /* 590c4762a1bSJed Brown Scatter ghost points to local vector, using the 2-step process 591c4762a1bSJed Brown DMGlobalToLocalBegin(), DMGlobalToLocalEnd(). 592c4762a1bSJed Brown By placing code between these two statements, computations can be 593c4762a1bSJed Brown done while messages are in transition. 594c4762a1bSJed Brown */ 5959566063dSJacob Faibussowitsch PetscCall(DMGlobalToLocalBegin(da, global_in, INSERT_VALUES, local_in)); 5969566063dSJacob Faibussowitsch PetscCall(DMGlobalToLocalEnd(da, global_in, INSERT_VALUES, local_in)); 597c4762a1bSJed Brown 598c4762a1bSJed Brown /* 599c4762a1bSJed Brown Get pointer to vector data 600c4762a1bSJed Brown */ 6019566063dSJacob Faibussowitsch PetscCall(VecGetArrayRead(local_in, &localptr)); 602c4762a1bSJed Brown 603c4762a1bSJed Brown /* 604c4762a1bSJed Brown Get starting and ending locally owned rows of the matrix 605c4762a1bSJed Brown */ 6069566063dSJacob Faibussowitsch PetscCall(MatGetOwnershipRange(B, &mstarts, &mends)); 6079371c9d4SSatish Balay mstart = mstarts; 6089371c9d4SSatish Balay mend = mends; 609c4762a1bSJed Brown 610c4762a1bSJed Brown /* - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - 611c4762a1bSJed Brown Compute entries for the locally owned part of the Jacobian. 612c4762a1bSJed Brown - Currently, all PETSc parallel matrix formats are partitioned by 613c4762a1bSJed Brown contiguous chunks of rows across the processors. 614c4762a1bSJed Brown - Each processor needs to insert only elements that it owns 615c4762a1bSJed Brown locally (but any non-local elements will be sent to the 616c4762a1bSJed Brown appropriate processor during matrix assembly). 617c4762a1bSJed Brown - Here, we set all entries for a particular row at once. 618c4762a1bSJed Brown - We can set matrix entries either using either 619c4762a1bSJed Brown MatSetValuesLocal() or MatSetValues(). 620c4762a1bSJed Brown - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - */ 621c4762a1bSJed Brown 622c4762a1bSJed Brown /* 623c4762a1bSJed Brown Set matrix rows corresponding to boundary data 624c4762a1bSJed Brown */ 625c4762a1bSJed Brown if (mstart == 0) { 626c4762a1bSJed Brown v[0] = 0.0; 6279566063dSJacob Faibussowitsch PetscCall(MatSetValues(B, 1, &mstart, 1, &mstart, v, INSERT_VALUES)); 628c4762a1bSJed Brown mstart++; 629c4762a1bSJed Brown } 630c4762a1bSJed Brown if (mend == appctx->m) { 631c4762a1bSJed Brown mend--; 632c4762a1bSJed Brown v[0] = 0.0; 6339566063dSJacob Faibussowitsch PetscCall(MatSetValues(B, 1, &mend, 1, &mend, v, INSERT_VALUES)); 634c4762a1bSJed Brown } 635c4762a1bSJed Brown 636c4762a1bSJed Brown /* 637c4762a1bSJed Brown Set matrix rows corresponding to interior data. We construct the 638c4762a1bSJed Brown matrix one row at a time. 639c4762a1bSJed Brown */ 640c4762a1bSJed Brown sc = 1.0 / (appctx->h * appctx->h * 2.0 * (1.0 + t) * (1.0 + t)); 641c4762a1bSJed Brown for (i = mstart; i < mend; i++) { 6429371c9d4SSatish Balay idx[0] = i - 1; 6439371c9d4SSatish Balay idx[1] = i; 6449371c9d4SSatish Balay idx[2] = i + 1; 645c4762a1bSJed Brown is = i - mstart + 1; 646c4762a1bSJed Brown v[0] = sc * localptr[is]; 647c4762a1bSJed Brown v[1] = sc * (localptr[is + 1] + localptr[is - 1] - 4.0 * localptr[is]); 648c4762a1bSJed Brown v[2] = sc * localptr[is]; 6499566063dSJacob Faibussowitsch PetscCall(MatSetValues(B, 1, &i, 3, idx, v, INSERT_VALUES)); 650c4762a1bSJed Brown } 651c4762a1bSJed Brown 652c4762a1bSJed Brown /* 653c4762a1bSJed Brown Restore vector 654c4762a1bSJed Brown */ 6559566063dSJacob Faibussowitsch PetscCall(VecRestoreArrayRead(local_in, &localptr)); 656c4762a1bSJed Brown 657c4762a1bSJed Brown /* - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - 658c4762a1bSJed Brown Complete the matrix assembly process and set some options 659c4762a1bSJed Brown - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - */ 660c4762a1bSJed Brown /* 661c4762a1bSJed Brown Assemble matrix, using the 2-step process: 662c4762a1bSJed Brown MatAssemblyBegin(), MatAssemblyEnd() 663c4762a1bSJed Brown Computations can be done while messages are in transition 664c4762a1bSJed Brown by placing code between these two statements. 665c4762a1bSJed Brown */ 6669566063dSJacob Faibussowitsch PetscCall(MatAssemblyBegin(B, MAT_FINAL_ASSEMBLY)); 6679566063dSJacob Faibussowitsch PetscCall(MatAssemblyEnd(B, MAT_FINAL_ASSEMBLY)); 668c4762a1bSJed Brown 669c4762a1bSJed Brown /* 670c4762a1bSJed Brown Set and option to indicate that we will never add a new nonzero location 671c4762a1bSJed Brown to the matrix. If we do, it will generate an error. 672c4762a1bSJed Brown */ 6739566063dSJacob Faibussowitsch PetscCall(MatSetOption(B, MAT_NEW_NONZERO_LOCATION_ERR, PETSC_TRUE)); 674c4762a1bSJed Brown 675c4762a1bSJed Brown return 0; 676c4762a1bSJed Brown } 677c4762a1bSJed Brown 678c4762a1bSJed Brown /*TEST 679c4762a1bSJed Brown 680c4762a1bSJed Brown test: 681c4762a1bSJed Brown args: -snes_type vinewtonrsls -ts_type glee -mymonitor -ts_max_steps 10 -nox 682c4762a1bSJed Brown requires: !single 683c4762a1bSJed Brown 684c4762a1bSJed Brown TEST*/ 685