Actual source code: mpibaij.h

petsc-3.11.4 2019-09-28
Report Typos and Errors

  4:  #include <../src/mat/impls/baij/seq/baij.h>
  5:  #include <../src/mat/impls/aij/mpi/mpiaij.h>
  6:  #include <petscctable.h>

  8: #if defined(PETSC_USE_CTABLE)
  9: #define PETSCTABLE PetscTable
 10: #else
 11: #define PETSCTABLE PetscInt*
 12: #endif

 14: #define MPIBAIJHEADER \
 15:   PetscInt    *rangebs;                /* rmap->range/bs */                                    \
 16:   PetscInt    rstartbs,rendbs,cstartbs,cendbs;    /* map values / bs  */                       \
 17:   Mat         A,B;                     /* local submatrices: A (diag part), B (off-diag part) */ \
 18:   PetscMPIInt size;                    /* size of communicator */                              \
 19:   PetscMPIInt rank;                    /* rank of proc in communicator */                      \
 20:   PetscInt    bs2;                     /* block size, bs2 = bs*bs */                           \
 21:   PetscInt    Mbs,Nbs;                 /* number block rows/cols in matrix; M/bs, N/bs */      \
 22:   PetscInt    mbs,nbs;                 /* number block rows/cols on processor; m/bs, n/bs */   \
 23:                                                                                                \
 24:   /* The following variables are used for matrix assembly */                                   \
 25:                                                                                                \
 26:   PetscBool   donotstash;               /* if 1, off processor entries dropped */              \
 27:   PetscBool   subset_off_proc_entries;  /* PETSC_TRUE if assembly will always communicate a subset of the entries communicated the first time */ \
 28:   MPI_Request *send_waits;              /* array of send requests */                           \
 29:   MPI_Request *recv_waits;              /* array of receive requests */                        \
 30:   PetscInt    nsends,nrecvs;           /* numbers of sends and receives */                     \
 31:   MatScalar   *svalues,*rvalues;       /* sending and receiving data */                        \
 32:   PetscInt    rmax;                     /* maximum message length */                           \
 33:   PETSCTABLE  colmap;                   /* local col number of off-diag col */                 \
 34:                                                                                                \
 35:   PetscInt *garray;                    /* work array */                                        \
 36:                                                                                                \
 37:   /* The following variable is used by blocked matrix assembly */                              \
 38:   MatScalar *barray;                    /* Block array of size bs2 */                          \
 39:                                                                                                \
 40:   /* The following variables are used for matrix-vector products */                            \
 41:                                                                                                \
 42:   Vec        lvec;                 /* local vector */                                          \
 43:   VecScatter Mvctx;                /* scatter context for vector */                            \
 44:   PetscBool  roworiented;          /* if true, row-oriented input, default true */             \
 45:                                                                                                \
 46:   /* The following variables are for MatGetRow() */                                            \
 47:                                                                                                \
 48:   PetscInt    *rowindices;         /* column indices for row */                                \
 49:   PetscScalar *rowvalues;          /* nonzero values in row */                                 \
 50:   PetscBool   getrowactive;        /* indicates MatGetRow(), not restored */                   \
 51:                                                                                                \
 52:   /* Some variables to make MatSetValues and others more efficient */                          \
 53:   PetscInt  rstart_bs,rend_bs;                                                                 \
 54:   PetscInt  cstart_bs,cend_bs;                                                                 \
 55:   PetscInt  *ht;                          /* Hash table to speed up matrix assembly */         \
 56:   MatScalar **hd;                         /* Hash table data */                                \
 57:   PetscInt  ht_size;                                                                           \
 58:   PetscInt  ht_total_ct,ht_insert_ct;     /* Hash table statistics */                          \
 59:   PetscBool ht_flag;                      /* Flag to indicate if hash tables are used */       \
 60:   double    ht_fact;                      /* Factor to determine the HT size */                \
 61:                                                                                                \
 62:   PetscInt  setvalueslen;       /* only used for single precision computations */              \
 63:   MatScalar *setvaluescopy;     /* area double precision values in MatSetValuesXXX() are copied*/ \
 64:                                 /* before calling MatSetValuesXXX_MPIBAIJ_MatScalar() */       \
 65:   PetscBool ijonly             /* used in  MatCreateSubMatrices_MPIBAIJ_local() for getting ij structure only */

 67: typedef struct {
 68:   MPIBAIJHEADER;
 69: } Mat_MPIBAIJ;

 71: PETSC_INTERN PetscErrorCode MatLoad_MPIBAIJ(Mat,PetscViewer);
 72: PETSC_INTERN PetscErrorCode MatCreateColmap_MPIBAIJ_Private(Mat);
 73: PETSC_INTERN PetscErrorCode MatCreateSubMatrices_MPIBAIJ(Mat,PetscInt,const IS[],const IS[],MatReuse,Mat*[]);
 74: PETSC_INTERN PetscErrorCode MatCreateSubMatrices_MPIBAIJ_local(Mat,PetscInt,const IS[],const IS[],MatReuse,Mat*);
 75: PETSC_INTERN PetscErrorCode MatCreateSubMatrix_MPIBAIJ_Private(Mat,IS,IS,PetscInt,MatReuse,Mat*);
 76: PETSC_INTERN PetscErrorCode MatGetMultiProcBlock_MPIBAIJ(Mat,MPI_Comm,MatReuse,Mat*);
 77: PETSC_INTERN PetscErrorCode MatIncreaseOverlap_MPIBAIJ(Mat,PetscInt,IS[],PetscInt);
 78: PETSC_INTERN PetscErrorCode MatIncreaseOverlap_MPIBAIJ_Once(Mat,PetscInt,IS*);
 79: PETSC_INTERN PetscErrorCode MatMPIBAIJSetPreallocation_MPIBAIJ(Mat B,PetscInt bs,PetscInt d_nz,const PetscInt *d_nnz,PetscInt o_nz,const PetscInt *o_nnz);
 80: PETSC_INTERN PetscErrorCode MatAXPYGetPreallocation_MPIBAIJ(Mat,const PetscInt *,Mat,const PetscInt*,PetscInt*);
 81: #endif