Actual source code: ex10.c

petsc-3.10.5 2019-03-28
Report Typos and Errors

  2: /*
  3:    Include "petscsnes.h" so that we can use SNES solvers.  Note that this
  4:    file automatically includes:
  5:      petscsys.h       - base PETSc routines   petscvec.h - vectors
  6:      petscmat.h - matrices
  7:      petscis.h     - index sets            petscksp.h - Krylov subspace methods
  8:      petscviewer.h - viewers               petscpc.h  - preconditioners
  9:      petscksp.h   - linear solvers
 10: */
 11: #include <petscsnes.h>
 12: #include <petscao.h>

 14: static char help[] = "An Unstructured Grid Example.\n\
 15: This example demonstrates how to solve a nonlinear system in parallel\n\
 16: with SNES for an unstructured mesh. The mesh and partitioning information\n\
 17: is read in an application defined ordering,which is later transformed\n\
 18: into another convenient ordering (called the local ordering). The local\n\
 19: ordering, apart from being efficient on cpu cycles and memory, allows\n\
 20: the use of the SPMD model of parallel programming. After partitioning\n\
 21: is done, scatters are created between local (sequential)and global\n\
 22: (distributed) vectors. Finally, we set up the nonlinear solver context\n\
 23: in the usual way as a structured grid  (see\n\
 24: petsc/src/snes/examples/tutorials/ex5.c).\n\
 25: This example also illustrates the use of parallel matrix coloring.\n\
 26: The command line options include:\n\
 27:   -vert <Nv>, where Nv is the global number of nodes\n\
 28:   -elem <Ne>, where Ne is the global number of elements\n\
 29:   -nl_par <lambda>, where lambda is the multiplier for the non linear term (u*u) term\n\
 30:   -lin_par <alpha>, where alpha is the multiplier for the linear term (u)\n\
 31:   -fd_jacobian_coloring -mat_coloring_type lf\n";

 33: /*T
 34:    Concepts: SNES^unstructured grid
 35:    Concepts: AO^application to PETSc ordering or vice versa;
 36:    Concepts: VecScatter^using vector scatter operations;
 37:    Processors: n
 38: T*/



 42: /* ------------------------------------------------------------------------

 44:    PDE Solved : L(u) + lambda*u*u + alpha*u = 0 where L(u) is the Laplacian.

 46:    The Laplacian is approximated in the following way: each edge is given a weight
 47:    of one meaning that the diagonal term will have the weight equal to the degree
 48:    of a node. The off diagonal terms will get a weight of -1.

 50:    -----------------------------------------------------------------------*/


 53: #define MAX_ELEM      500  /* Maximum number of elements */
 54: #define MAX_VERT      100  /* Maximum number of vertices */
 55: #define MAX_VERT_ELEM   3  /* Vertices per element       */

 57: /*
 58:   Application-defined context for problem specific data
 59: */
 60: typedef struct {
 61:   PetscInt   Nvglobal,Nvlocal;              /* global and local number of vertices */
 62:   PetscInt   Neglobal,Nelocal;              /* global and local number of vertices */
 63:   PetscInt   AdjM[MAX_VERT][50];            /* adjacency list of a vertex */
 64:   PetscInt   itot[MAX_VERT];                /* total number of neighbors for a vertex */
 65:   PetscInt   icv[MAX_ELEM][MAX_VERT_ELEM];  /* vertices belonging to an element */
 66:   PetscInt   v2p[MAX_VERT];                 /* processor number for a vertex */
 67:   PetscInt   *locInd,*gloInd;               /* local and global orderings for a node */
 68:   Vec        localX,localF;                 /* local solution (u) and f(u) vectors */
 69:   PetscReal  non_lin_param;                 /* nonlinear parameter for the PDE */
 70:   PetscReal  lin_param;                     /* linear parameter for the PDE */
 71:   VecScatter scatter;                       /* scatter context for the local and
 72:                                                distributed vectors */
 73: } AppCtx;

 75: /*
 76:   User-defined routines
 77: */
 78: PetscErrorCode FormJacobian(SNES,Vec,Mat,Mat,void*);
 79: PetscErrorCode FormFunction(SNES,Vec,Vec,void*);
 80: PetscErrorCode FormInitialGuess(AppCtx*,Vec);

 82: int main(int argc,char **argv)
 83: {
 84:   SNES                   snes;                 /* SNES context */
 85:   SNESType               type = SNESNEWTONLS;  /* default nonlinear solution method */
 86:   Vec                    x,r;                  /* solution, residual vectors */
 87:   Mat                    Jac;                  /* Jacobian matrix */
 88:   AppCtx                 user;                 /* user-defined application context */
 89:   AO                     ao;                   /* Application Ordering object */
 90:   IS                     isglobal,islocal;     /* global and local index sets */
 91:   PetscMPIInt            rank,size;            /* rank of a process, number of processors */
 92:   PetscInt               rstart;               /* starting index of PETSc ordering for a processor */
 93:   PetscInt               nfails;               /* number of unsuccessful Newton steps */
 94:   PetscInt               bs = 1;               /* block size for multicomponent systems */
 95:   PetscInt               nvertices;            /* number of local plus ghost nodes of a processor */
 96:   PetscInt               *pordering;           /* PETSc ordering */
 97:   PetscInt               *vertices;            /* list of all vertices (incl. ghost ones) on a processor */
 98:   PetscInt               *verticesmask;
 99:   PetscInt               *tmp;
100:   PetscInt               i,j,jstart,inode,nb,nbrs,Nvneighborstotal = 0;
101:   PetscErrorCode         ierr;
102:   PetscInt               its,N;
103:   PetscScalar            *xx;
104:   char                   str[256],form[256],part_name[256];
105:   FILE                   *fptr,*fptr1;
106:   ISLocalToGlobalMapping isl2g;
107:   int                    dtmp;
108: #if defined(UNUSED_VARIABLES)
109:   PetscDraw              draw;                 /* drawing context */
110:   PetscScalar            *ff,*gg;
111:   PetscReal              tiny = 1.0e-10,zero = 0.0,one = 1.0,big = 1.0e+10;
112:   PetscInt               *tmp1,*tmp2;
113: #endif
114:   MatFDColoring          matfdcoloring = 0;
115:   PetscBool              fd_jacobian_coloring = PETSC_FALSE;

117:   /* - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -
118:      Initialize program
119:      - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - */

121:    PetscInitialize(&argc,&argv,"options.inf",help);if (ierr) return ierr;
122:   MPI_Comm_rank(MPI_COMM_WORLD,&rank);
123:   MPI_Comm_size(MPI_COMM_WORLD,&size);

125:   /* The current input file options.inf is for 2 proc run only */
126:   if (size != 2) SETERRQ(PETSC_COMM_SELF,1,"This Example currently runs on 2 procs only.");

128:   /*
129:      Initialize problem parameters
130:   */
131:   user.Nvglobal = 16;      /*Global # of vertices  */
132:   user.Neglobal = 18;      /*Global # of elements  */

134:   PetscOptionsGetInt(NULL,NULL,"-vert",&user.Nvglobal,NULL);
135:   PetscOptionsGetInt(NULL,NULL,"-elem",&user.Neglobal,NULL);

137:   user.non_lin_param = 0.06;

139:   PetscOptionsGetReal(NULL,NULL,"-nl_par",&user.non_lin_param,NULL);

141:   user.lin_param = -1.0;

143:   PetscOptionsGetReal(NULL,NULL,"-lin_par",&user.lin_param,NULL);

145:   user.Nvlocal = 0;
146:   user.Nelocal = 0;

148:   /* - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -
149:       Read the mesh and partitioning information
150:      - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - */

152:   /*
153:      Read the mesh and partitioning information from 'adj.in'.
154:      The file format is as follows.
155:      For each line the first entry is the processor rank where the
156:      current node belongs. The second entry is the number of
157:      neighbors of a node. The rest of the line is the adjacency
158:      list of a node. Currently this file is set up to work on two
159:      processors.

161:      This is not a very good example of reading input. In the future,
162:      we will put an example that shows the style that should be
163:      used in a real application, where partitioning will be done
164:      dynamically by calling partitioning routines (at present, we have
165:      a  ready interface to ParMeTiS).
166:    */
167:   fptr = fopen("adj.in","r");
168:   if (!fptr) SETERRQ(PETSC_COMM_SELF,0,"Could not open adj.in");

170:   /*
171:      Each processor writes to the file output.<rank> where rank is the
172:      processor's rank.
173:   */
174:   sprintf(part_name,"output.%d",rank);
175:   fptr1 = fopen(part_name,"w");
176:   if (!fptr1) SETERRQ(PETSC_COMM_SELF,0,"Could no open output file");
177:   PetscMalloc1(user.Nvglobal,&user.gloInd);
178:   PetscFPrintf(PETSC_COMM_SELF,fptr1,"Rank is %D\n",rank);
179:   for (inode = 0; inode < user.Nvglobal; inode++) {
180:     if (!fgets(str,256,fptr)) SETERRQ(PETSC_COMM_SELF,1,"fgets read failed");
181:     sscanf(str,"%d",&dtmp);user.v2p[inode] = dtmp;
182:     if (user.v2p[inode] == rank) {
183:       PetscFPrintf(PETSC_COMM_SELF,fptr1,"Node %D belongs to processor %D\n",inode,user.v2p[inode]);

185:       user.gloInd[user.Nvlocal] = inode;
186:       sscanf(str,"%*d %d",&dtmp);
187:       nbrs = dtmp;
188:       PetscFPrintf(PETSC_COMM_SELF,fptr1,"Number of neighbors for the vertex %D is %D\n",inode,nbrs);

190:       user.itot[user.Nvlocal] = nbrs;
191:       Nvneighborstotal       += nbrs;
192:       for (i = 0; i < user.itot[user.Nvlocal]; i++) {
193:         form[0]='\0';
194:         for (j=0; j < i+2; j++) {
195:           PetscStrlcat(form,"%*d ",sizeof(form));
196:         }
197:         PetscStrlcat(form,"%d",sizeof(form));

199:         sscanf(str,form,&dtmp);
200:         user.AdjM[user.Nvlocal][i] = dtmp;

202:         PetscFPrintf(PETSC_COMM_SELF,fptr1,"%D ",user.AdjM[user.Nvlocal][i]);
203:       }
204:       PetscFPrintf(PETSC_COMM_SELF,fptr1,"\n");
205:       user.Nvlocal++;
206:     }
207:   }
208:   PetscFPrintf(PETSC_COMM_SELF,fptr1,"Total # of Local Vertices is %D \n",user.Nvlocal);

210:   /* - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -
211:      Create different orderings
212:      - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - */

214:   /*
215:     Create the local ordering list for vertices. First a list using the PETSc global
216:     ordering is created. Then we use the AO object to get the PETSc-to-application and
217:     application-to-PETSc mappings. Each vertex also gets a local index (stored in the
218:     locInd array).
219:   */
220:   MPI_Scan(&user.Nvlocal,&rstart,1,MPIU_INT,MPI_SUM,PETSC_COMM_WORLD);
221:   rstart -= user.Nvlocal;
222:   PetscMalloc1(user.Nvlocal,&pordering);

224:   for (i=0; i < user.Nvlocal; i++) pordering[i] = rstart + i;

226:   /*
227:     Create the AO object
228:   */
229:   AOCreateBasic(MPI_COMM_WORLD,user.Nvlocal,user.gloInd,pordering,&ao);
230:   PetscFree(pordering);

232:   /*
233:     Keep the global indices for later use
234:   */
235:   PetscMalloc1(user.Nvlocal,&user.locInd);
236:   PetscMalloc1(Nvneighborstotal,&tmp);

238:   /* - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -
239:     Demonstrate the use of AO functionality
240:      - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - */

242:   PetscFPrintf(PETSC_COMM_SELF,fptr1,"Before AOApplicationToPetsc, local indices are : \n");
243:   for (i=0; i < user.Nvlocal; i++) {
244:     PetscFPrintf(PETSC_COMM_SELF,fptr1," %D ",user.gloInd[i]);

246:     user.locInd[i] = user.gloInd[i];
247:   }
248:   PetscFPrintf(PETSC_COMM_SELF,fptr1,"\n");
249:   jstart = 0;
250:   for (i=0; i < user.Nvlocal; i++) {
251:     PetscFPrintf(PETSC_COMM_SELF,fptr1,"Neghbors of local vertex %D are : ",user.gloInd[i]);
252:     for (j=0; j < user.itot[i]; j++) {
253:       PetscFPrintf(PETSC_COMM_SELF,fptr1,"%D ",user.AdjM[i][j]);

255:       tmp[j + jstart] = user.AdjM[i][j];
256:     }
257:     jstart += user.itot[i];
258:     PetscFPrintf(PETSC_COMM_SELF,fptr1,"\n");
259:   }

261:   /*
262:     Now map the vlocal and neighbor lists to the PETSc ordering
263:   */
264:   AOApplicationToPetsc(ao,user.Nvlocal,user.locInd);
265:   AOApplicationToPetsc(ao,Nvneighborstotal,tmp);
266:   AODestroy(&ao);

268:   PetscFPrintf(PETSC_COMM_SELF,fptr1,"After AOApplicationToPetsc, local indices are : \n");
269:   for (i=0; i < user.Nvlocal; i++) {
270:     PetscFPrintf(PETSC_COMM_SELF,fptr1," %D ",user.locInd[i]);
271:   }
272:   PetscFPrintf(PETSC_COMM_SELF,fptr1,"\n");

274:   jstart = 0;
275:   for (i=0; i < user.Nvlocal; i++) {
276:     PetscFPrintf(PETSC_COMM_SELF,fptr1,"Neghbors of local vertex %D are : ",user.locInd[i]);
277:     for (j=0; j < user.itot[i]; j++) {
278:       user.AdjM[i][j] = tmp[j+jstart];

280:       PetscFPrintf(PETSC_COMM_SELF,fptr1,"%D ",user.AdjM[i][j]);
281:     }
282:     jstart += user.itot[i];
283:     PetscFPrintf(PETSC_COMM_SELF,fptr1,"\n");
284:   }

286:   /* - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -
287:      Extract the ghost vertex information for each processor
288:      - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - */
289:   /*
290:    Next, we need to generate a list of vertices required for this processor
291:    and a local numbering scheme for all vertices required on this processor.
292:       vertices - integer array of all vertices needed on this processor in PETSc
293:                  global numbering; this list consists of first the "locally owned"
294:                  vertices followed by the ghost vertices.
295:       verticesmask - integer array that for each global vertex lists its local
296:                      vertex number (in vertices) + 1. If the global vertex is not
297:                      represented on this processor, then the corresponding
298:                      entry in verticesmask is zero

300:       Note: vertices and verticesmask are both Nvglobal in length; this may
301:     sound terribly non-scalable, but in fact is not so bad for a reasonable
302:     number of processors. Importantly, it allows us to use NO SEARCHING
303:     in setting up the data structures.
304:   */
305:   PetscMalloc1(user.Nvglobal,&vertices);
306:   PetscMalloc1(user.Nvglobal,&verticesmask);
307:   PetscMemzero(verticesmask,user.Nvglobal*sizeof(PetscInt));
308:   nvertices = 0;

310:   /*
311:     First load "owned vertices" into list
312:   */
313:   for (i=0; i < user.Nvlocal; i++) {
314:     vertices[nvertices++]        = user.locInd[i];
315:     verticesmask[user.locInd[i]] = nvertices;
316:   }

318:   /*
319:     Now load ghost vertices into list
320:   */
321:   for (i=0; i < user.Nvlocal; i++) {
322:     for (j=0; j < user.itot[i]; j++) {
323:       nb = user.AdjM[i][j];
324:       if (!verticesmask[nb]) {
325:         vertices[nvertices++] = nb;
326:         verticesmask[nb]      = nvertices;
327:       }
328:     }
329:   }

331:   PetscFPrintf(PETSC_COMM_SELF,fptr1,"\n");
332:   PetscFPrintf(PETSC_COMM_SELF,fptr1,"The array vertices is :\n");
333:   for (i=0; i < nvertices; i++) {
334:     PetscFPrintf(PETSC_COMM_SELF,fptr1,"%D ",vertices[i]);
335:   }
336:   PetscFPrintf(PETSC_COMM_SELF,fptr1,"\n");

338:   /*
339:      Map the vertices listed in the neighbors to the local numbering from
340:     the global ordering that they contained initially.
341:   */
342:   PetscFPrintf(PETSC_COMM_SELF,fptr1,"\n");
343:   PetscFPrintf(PETSC_COMM_SELF,fptr1,"After mapping neighbors in the local contiguous ordering\n");
344:   for (i=0; i<user.Nvlocal; i++) {
345:     PetscFPrintf(PETSC_COMM_SELF,fptr1,"Neghbors of local vertex %D are :\n",i);
346:     for (j = 0; j < user.itot[i]; j++) {
347:       nb              = user.AdjM[i][j];
348:       user.AdjM[i][j] = verticesmask[nb] - 1;

350:       PetscFPrintf(PETSC_COMM_SELF,fptr1,"%D ",user.AdjM[i][j]);
351:     }
352:     PetscFPrintf(PETSC_COMM_SELF,fptr1,"\n");
353:   }

355:   N = user.Nvglobal;

357:   /* - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -
358:      Create vector and matrix data structures
359:      - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - */

361:   /*
362:     Create vector data structures
363:   */
364:   VecCreate(MPI_COMM_WORLD,&x);
365:   VecSetSizes(x,user.Nvlocal,N);
366:   VecSetFromOptions(x);
367:   VecDuplicate(x,&r);
368:   VecCreateSeq(MPI_COMM_SELF,bs*nvertices,&user.localX);
369:   VecDuplicate(user.localX,&user.localF);

371:   /*
372:     Create the scatter between the global representation and the
373:     local representation
374:   */
375:   ISCreateStride(MPI_COMM_SELF,bs*nvertices,0,1,&islocal);
376:   ISCreateBlock(MPI_COMM_SELF,bs,nvertices,vertices,PETSC_COPY_VALUES,&isglobal);
377:   VecScatterCreate(x,isglobal,user.localX,islocal,&user.scatter);
378:   ISDestroy(&isglobal);
379:   ISDestroy(&islocal);

381:   /*
382:      Create matrix data structure; Just to keep the example simple, we have not done any
383:      preallocation of memory for the matrix. In real application code with big matrices,
384:      preallocation should always be done to expedite the matrix creation.
385:   */
386:   MatCreate(MPI_COMM_WORLD,&Jac);
387:   MatSetSizes(Jac,PETSC_DECIDE,PETSC_DECIDE,N,N);
388:   MatSetFromOptions(Jac);
389:   MatSetUp(Jac);

391:   /*
392:     The following routine allows us to set the matrix values in local ordering
393:   */
394:   ISLocalToGlobalMappingCreate(MPI_COMM_SELF,bs,nvertices,vertices,PETSC_COPY_VALUES,&isl2g);
395:   MatSetLocalToGlobalMapping(Jac,isl2g,isl2g);

397:   /* - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -
398:      Create nonlinear solver context
399:      - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - */

401:   SNESCreate(MPI_COMM_WORLD,&snes);
402:   SNESSetType(snes,type);

404:   /* - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -
405:     Set routines for function and Jacobian evaluation
406:      - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - */
407:   SNESSetFunction(snes,r,FormFunction,(void*)&user);

409:   PetscOptionsGetBool(NULL,NULL,"-fd_jacobian_coloring",&fd_jacobian_coloring,0);
410:   if (!fd_jacobian_coloring) {
411:     SNESSetJacobian(snes,Jac,Jac,FormJacobian,(void*)&user);
412:   } else {  /* Use matfdcoloring */
413:     ISColoring   iscoloring;
414:     MatColoring  mc;

416:     /* Get the data structure of Jac */
417:     FormJacobian(snes,x,Jac,Jac,&user);
418:     /* Create coloring context */
419:     MatColoringCreate(Jac,&mc);
420:     MatColoringSetType(mc,MATCOLORINGSL);
421:     MatColoringSetFromOptions(mc);
422:     MatColoringApply(mc,&iscoloring);
423:     MatColoringDestroy(&mc);
424:     MatFDColoringCreate(Jac,iscoloring,&matfdcoloring);
425:     MatFDColoringSetFunction(matfdcoloring,(PetscErrorCode (*)(void))FormFunction,&user);
426:     MatFDColoringSetFromOptions(matfdcoloring);
427:     MatFDColoringSetUp(Jac,iscoloring,matfdcoloring);
428:     /* MatFDColoringView(matfdcoloring,PETSC_VIEWER_STDOUT_WORLD); */
429:     SNESSetJacobian(snes,Jac,Jac,SNESComputeJacobianDefaultColor,matfdcoloring);
430:     ISColoringDestroy(&iscoloring);
431:   }

433:   /* - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -
434:      Customize nonlinear solver; set runtime options
435:    - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - */

437:   SNESSetFromOptions(snes);

439:   /* - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -
440:      Evaluate initial guess; then solve nonlinear system
441:      - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - */

443:   /*
444:      Note: The user should initialize the vector, x, with the initial guess
445:      for the nonlinear solver prior to calling SNESSolve().  In particular,
446:      to employ an initial guess of zero, the user should explicitly set
447:      this vector to zero by calling VecSet().
448:   */
449:   FormInitialGuess(&user,x);

451:   /*
452:     Print the initial guess
453:   */
454:   VecGetArray(x,&xx);
455:   for (inode = 0; inode < user.Nvlocal; inode++) {
456:     PetscFPrintf(PETSC_COMM_SELF,fptr1,"Initial Solution at node %D is %f \n",inode,xx[inode]);
457:   }
458:   VecRestoreArray(x,&xx);

460:   /* - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -
461:      Now solve the nonlinear system
462:    - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - */

464:   SNESSolve(snes,NULL,x);
465:   SNESGetIterationNumber(snes,&its);
466:   SNESGetNonlinearStepFailures(snes,&nfails);

468:   /* - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -
469:     Print the output : solution vector and other information
470:      Each processor writes to the file output.<rank> where rank is the
471:      processor's rank.
472:    - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - */

474:   VecGetArray(x,&xx);
475:   for (inode = 0; inode < user.Nvlocal; inode++) {
476:     PetscFPrintf(PETSC_COMM_SELF,fptr1,"Solution at node %D is %f \n",inode,xx[inode]);
477:   }
478:   VecRestoreArray(x,&xx);
479:   fclose(fptr1);
480:   PetscPrintf(MPI_COMM_WORLD,"number of SNES iterations = %D, ",its);
481:   PetscPrintf(MPI_COMM_WORLD,"number of unsuccessful steps = %D\n",nfails);

483:   /* - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -
484:      Free work space.  All PETSc objects should be destroyed when they
485:      are no longer needed.
486:    - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - */
487:   PetscFree(user.gloInd);
488:   PetscFree(user.locInd);
489:   PetscFree(vertices);
490:   PetscFree(verticesmask);
491:   PetscFree(tmp);
492:   VecScatterDestroy(&user.scatter);
493:   ISLocalToGlobalMappingDestroy(&isl2g);
494:   VecDestroy(&x);
495:   VecDestroy(&r);
496:   VecDestroy(&user.localX);
497:   VecDestroy(&user.localF);
498:   MatDestroy(&Jac);  SNESDestroy(&snes);
499:   /*PetscDrawDestroy(draw);*/
500:   if (fd_jacobian_coloring) {
501:     MatFDColoringDestroy(&matfdcoloring);
502:   }
503:   PetscFinalize();
504:   return ierr;
505: }
506: /* --------------------  Form initial approximation ----------------- */

508: /*
509:    FormInitialGuess - Forms initial approximation.

511:    Input Parameters:
512:    user - user-defined application context
513:    X - vector

515:    Output Parameter:
516:    X - vector
517:  */
518: PetscErrorCode FormInitialGuess(AppCtx *user,Vec X)
519: {
520:   PetscInt    i,Nvlocal,ierr;
521:   PetscInt    *gloInd;
522:   PetscScalar *x;
523: #if defined(UNUSED_VARIABLES)
524:   PetscReal temp1,temp,hx,hy,hxdhy,hydhx,sc;
525:   PetscInt  Neglobal,Nvglobal,j,row;
526:   PetscReal alpha,lambda;

528:   Nvglobal = user->Nvglobal;
529:   Neglobal = user->Neglobal;
530:   lambda   = user->non_lin_param;
531:   alpha    = user->lin_param;
532: #endif

534:   Nvlocal = user->Nvlocal;
535:   gloInd  = user->gloInd;

537:   /*
538:      Get a pointer to vector data.
539:        - For default PETSc vectors, VecGetArray() returns a pointer to
540:          the data array.  Otherwise, the routine is implementation dependent.
541:        - You MUST call VecRestoreArray() when you no longer need access to
542:          the array.
543:   */
544:   VecGetArray(X,&x);

546:   /*
547:      Compute initial guess over the locally owned part of the grid
548:   */
549:   for (i=0; i < Nvlocal; i++) x[i] = (PetscReal)gloInd[i];

551:   /*
552:      Restore vector
553:   */
554:   VecRestoreArray(X,&x);
555:   return 0;
556: }
557: /* --------------------  Evaluate Function F(x) --------------------- */
558: /*
559:    FormFunction - Evaluates nonlinear function, F(x).

561:    Input Parameters:
562: .  snes - the SNES context
563: .  X - input vector
564: .  ptr - optional user-defined context, as set by SNESSetFunction()

566:    Output Parameter:
567: .  F - function vector
568:  */
569: PetscErrorCode FormFunction(SNES snes,Vec X,Vec F,void *ptr)
570: {
571:   AppCtx         *user = (AppCtx*)ptr;
573:   PetscInt       i,j,Nvlocal;
574:   PetscReal      alpha,lambda;
575:   PetscScalar    *x,*f;
576:   VecScatter     scatter;
577:   Vec            localX = user->localX;
578: #if defined(UNUSED_VARIABLES)
579:   PetscScalar ut,ub,ul,ur,u,*g,sc,uyy,uxx;
580:   PetscReal   hx,hy,hxdhy,hydhx;
581:   PetscReal   two = 2.0,one = 1.0;
582:   PetscInt    Nvglobal,Neglobal,row;
583:   PetscInt    *gloInd;

585:   Nvglobal = user->Nvglobal;
586:   Neglobal = user->Neglobal;
587:   gloInd   = user->gloInd;
588: #endif

590:   Nvlocal = user->Nvlocal;
591:   lambda  = user->non_lin_param;
592:   alpha   = user->lin_param;
593:   scatter = user->scatter;

595:   /*
596:      PDE : L(u) + lambda*u*u +alpha*u = 0 where L(u) is the approximate Laplacian as
597:      described in the beginning of this code

599:      First scatter the distributed vector X into local vector localX (that includes
600:      values for ghost nodes. If we wish,we can put some other work between
601:      VecScatterBegin() and VecScatterEnd() to overlap the communication with
602:      computation.
603:  */
604:   VecScatterBegin(scatter,X,localX,INSERT_VALUES,SCATTER_FORWARD);
605:   VecScatterEnd(scatter,X,localX,INSERT_VALUES,SCATTER_FORWARD);

607:   /*
608:      Get pointers to vector data
609:   */
610:   VecGetArray(localX,&x);
611:   VecGetArray(F,&f);

613:   /*
614:     Now compute the f(x). As mentioned earlier, the computed Laplacian is just an
615:     approximate one chosen for illustrative purpose only. Another point to notice
616:     is that this is a local (completly parallel) calculation. In practical application
617:     codes, function calculation time is a dominat portion of the overall execution time.
618:   */
619:   for (i=0; i < Nvlocal; i++) {
620:     f[i] = (user->itot[i] - alpha)*x[i] - lambda*x[i]*x[i];
621:     for (j = 0; j < user->itot[i]; j++) f[i] -= x[user->AdjM[i][j]];
622:   }

624:   /*
625:      Restore vectors
626:   */
627:   VecRestoreArray(localX,&x);
628:   VecRestoreArray(F,&f);
629:   /*VecView(F,PETSC_VIEWER_STDOUT_WORLD);*/

631:   return 0;
632: }

634: /* --------------------  Evaluate Jacobian F'(x) -------------------- */
635: /*
636:    FormJacobian - Evaluates Jacobian matrix.

638:    Input Parameters:
639: .  snes - the SNES context
640: .  X - input vector
641: .  ptr - optional user-defined context, as set by SNESSetJacobian()

643:    Output Parameters:
644: .  A - Jacobian matrix
645: .  B - optionally different preconditioning matrix
646: .  flag - flag indicating matrix structure

648: */
649: PetscErrorCode FormJacobian(SNES snes,Vec X,Mat J,Mat jac,void *ptr)
650: {
651:   AppCtx      *user = (AppCtx*)ptr;
652:   PetscInt    i,j,Nvlocal,col[50],ierr;
653:   PetscScalar alpha,lambda,value[50];
654:   Vec         localX = user->localX;
655:   VecScatter  scatter;
656:   PetscScalar *x;
657: #if defined(UNUSED_VARIABLES)
658:   PetscScalar two = 2.0,one = 1.0;
659:   PetscInt    row,Nvglobal,Neglobal;
660:   PetscInt    *gloInd;

662:   Nvglobal = user->Nvglobal;
663:   Neglobal = user->Neglobal;
664:   gloInd   = user->gloInd;
665: #endif

667:   /*printf("Entering into FormJacobian \n");*/
668:   Nvlocal = user->Nvlocal;
669:   lambda  = user->non_lin_param;
670:   alpha   =  user->lin_param;
671:   scatter = user->scatter;

673:   /*
674:      PDE : L(u) + lambda*u*u +alpha*u = 0 where L(u) is the approximate Laplacian as
675:      described in the beginning of this code

677:      First scatter the distributed vector X into local vector localX (that includes
678:      values for ghost nodes. If we wish, we can put some other work between
679:      VecScatterBegin() and VecScatterEnd() to overlap the communication with
680:      computation.
681:   */
682:   VecScatterBegin(scatter,X,localX,INSERT_VALUES,SCATTER_FORWARD);
683:   VecScatterEnd(scatter,X,localX,INSERT_VALUES,SCATTER_FORWARD);

685:   /*
686:      Get pointer to vector data
687:   */
688:   VecGetArray(localX,&x);

690:   for (i=0; i < Nvlocal; i++) {
691:     col[0]   = i;
692:     value[0] = user->itot[i] - 2.0*lambda*x[i] - alpha;
693:     for (j = 0; j < user->itot[i]; j++) {
694:       col[j+1]   = user->AdjM[i][j];
695:       value[j+1] = -1.0;
696:     }

698:     /*
699:       Set the matrix values in the local ordering. Note that in order to use this
700:       feature we must call the routine MatSetLocalToGlobalMapping() after the
701:       matrix has been created.
702:     */
703:     MatSetValuesLocal(jac,1,&i,1+user->itot[i],col,value,INSERT_VALUES);
704:   }

706:   /*
707:      Assemble matrix, using the 2-step process:
708:        MatAssemblyBegin(), MatAssemblyEnd().
709:      Between these two calls, the pointer to vector data has been restored to
710:      demonstrate the use of overlapping communicationn with computation.
711:   */
712:   MatAssemblyBegin(jac,MAT_FINAL_ASSEMBLY);
713:   VecRestoreArray(localX,&x);
714:   MatAssemblyEnd(jac,MAT_FINAL_ASSEMBLY);

716:   /*
717:      Tell the matrix we will never add a new nonzero location to the
718:      matrix. If we do, it will generate an error.
719:   */
720:   MatSetOption(jac,MAT_NEW_NONZERO_LOCATION_ERR,PETSC_TRUE);
721:   /* MatView(jac,PETSC_VIEWER_STDOUT_SELF); */
722:   return 0;
723: }



727: /*TEST

729:    build:
730:       requires: !complex

732:    test:
733:       nsize: 2
734:       args: -snes_monitor_short
735:       localrunfiles: options.inf adj.in

737:    test:
738:       suffix: 2
739:       nsize: 2
740:       args: -snes_monitor_short -fd_jacobian_coloring
741:       localrunfiles: options.inf adj.in

743: TEST*/