xref: /petsc/src/mat/impls/sbaij/mpi/mpisbaij.c (revision 67f7d742d70e61a8a21ffc5235261ec57ea3c5cd)
1 #include <../src/mat/impls/baij/mpi/mpibaij.h> /*I "petscmat.h" I*/
2 #include <../src/mat/impls/sbaij/mpi/mpisbaij.h>
3 #include <../src/mat/impls/sbaij/seq/sbaij.h>
4 #include <petscblaslapack.h>
5 
6 static PetscErrorCode MatDestroy_MPISBAIJ(Mat mat)
7 {
8   Mat_MPISBAIJ *baij = (Mat_MPISBAIJ *)mat->data;
9 
10   PetscFunctionBegin;
11   PetscCall(PetscLogObjectState((PetscObject)mat, "Rows=%" PetscInt_FMT ",Cols=%" PetscInt_FMT, mat->rmap->N, mat->cmap->N));
12   PetscCall(MatStashDestroy_Private(&mat->stash));
13   PetscCall(MatStashDestroy_Private(&mat->bstash));
14   PetscCall(MatDestroy(&baij->A));
15   PetscCall(MatDestroy(&baij->B));
16 #if defined(PETSC_USE_CTABLE)
17   PetscCall(PetscHMapIDestroy(&baij->colmap));
18 #else
19   PetscCall(PetscFree(baij->colmap));
20 #endif
21   PetscCall(PetscFree(baij->garray));
22   PetscCall(VecDestroy(&baij->lvec));
23   PetscCall(VecScatterDestroy(&baij->Mvctx));
24   PetscCall(VecDestroy(&baij->slvec0));
25   PetscCall(VecDestroy(&baij->slvec0b));
26   PetscCall(VecDestroy(&baij->slvec1));
27   PetscCall(VecDestroy(&baij->slvec1a));
28   PetscCall(VecDestroy(&baij->slvec1b));
29   PetscCall(VecScatterDestroy(&baij->sMvctx));
30   PetscCall(PetscFree2(baij->rowvalues, baij->rowindices));
31   PetscCall(PetscFree(baij->barray));
32   PetscCall(PetscFree(baij->hd));
33   PetscCall(VecDestroy(&baij->diag));
34   PetscCall(VecDestroy(&baij->bb1));
35   PetscCall(VecDestroy(&baij->xx1));
36 #if defined(PETSC_USE_REAL_MAT_SINGLE)
37   PetscCall(PetscFree(baij->setvaluescopy));
38 #endif
39   PetscCall(PetscFree(baij->in_loc));
40   PetscCall(PetscFree(baij->v_loc));
41   PetscCall(PetscFree(baij->rangebs));
42   PetscCall(PetscFree(mat->data));
43 
44   PetscCall(PetscObjectChangeTypeName((PetscObject)mat, NULL));
45   PetscCall(PetscObjectComposeFunction((PetscObject)mat, "MatStoreValues_C", NULL));
46   PetscCall(PetscObjectComposeFunction((PetscObject)mat, "MatRetrieveValues_C", NULL));
47   PetscCall(PetscObjectComposeFunction((PetscObject)mat, "MatMPISBAIJSetPreallocation_C", NULL));
48   PetscCall(PetscObjectComposeFunction((PetscObject)mat, "MatMPISBAIJSetPreallocationCSR_C", NULL));
49 #if defined(PETSC_HAVE_ELEMENTAL)
50   PetscCall(PetscObjectComposeFunction((PetscObject)mat, "MatConvert_mpisbaij_elemental_C", NULL));
51 #endif
52 #if defined(PETSC_HAVE_SCALAPACK)
53   PetscCall(PetscObjectComposeFunction((PetscObject)mat, "MatConvert_mpisbaij_scalapack_C", NULL));
54 #endif
55   PetscCall(PetscObjectComposeFunction((PetscObject)mat, "MatConvert_mpisbaij_mpiaij_C", NULL));
56   PetscCall(PetscObjectComposeFunction((PetscObject)mat, "MatConvert_mpisbaij_mpibaij_C", NULL));
57   PetscFunctionReturn(PETSC_SUCCESS);
58 }
59 
60 /* defines MatSetValues_MPI_Hash(), MatAssemblyBegin_MPI_Hash(), MatAssemblyEnd_MPI_Hash(), MatSetUp_MPI_Hash() */
61 #define TYPE SBAIJ
62 #define TYPE_SBAIJ
63 #include "../src/mat/impls/aij/mpi/mpihashmat.h"
64 #undef TYPE
65 #undef TYPE_SBAIJ
66 
67 #if defined(PETSC_HAVE_ELEMENTAL)
68 PETSC_INTERN PetscErrorCode MatConvert_MPISBAIJ_Elemental(Mat, MatType, MatReuse, Mat *);
69 #endif
70 #if defined(PETSC_HAVE_SCALAPACK)
71 PETSC_INTERN PetscErrorCode MatConvert_SBAIJ_ScaLAPACK(Mat, MatType, MatReuse, Mat *);
72 #endif
73 
74 /* This could be moved to matimpl.h */
75 static PetscErrorCode MatPreallocateWithMats_Private(Mat B, PetscInt nm, Mat X[], PetscBool symm[], PetscBool fill)
76 {
77   Mat       preallocator;
78   PetscInt  r, rstart, rend;
79   PetscInt  bs, i, m, n, M, N;
80   PetscBool cong = PETSC_TRUE;
81 
82   PetscFunctionBegin;
83   PetscValidHeaderSpecific(B, MAT_CLASSID, 1);
84   PetscValidLogicalCollectiveInt(B, nm, 2);
85   for (i = 0; i < nm; i++) {
86     PetscValidHeaderSpecific(X[i], MAT_CLASSID, 3);
87     PetscCall(PetscLayoutCompare(B->rmap, X[i]->rmap, &cong));
88     PetscCheck(cong, PetscObjectComm((PetscObject)B), PETSC_ERR_SUP, "Not for different layouts");
89   }
90   PetscValidLogicalCollectiveBool(B, fill, 5);
91   PetscCall(MatGetBlockSize(B, &bs));
92   PetscCall(MatGetSize(B, &M, &N));
93   PetscCall(MatGetLocalSize(B, &m, &n));
94   PetscCall(MatCreate(PetscObjectComm((PetscObject)B), &preallocator));
95   PetscCall(MatSetType(preallocator, MATPREALLOCATOR));
96   PetscCall(MatSetBlockSize(preallocator, bs));
97   PetscCall(MatSetSizes(preallocator, m, n, M, N));
98   PetscCall(MatSetUp(preallocator));
99   PetscCall(MatGetOwnershipRange(preallocator, &rstart, &rend));
100   for (r = rstart; r < rend; ++r) {
101     PetscInt           ncols;
102     const PetscInt    *row;
103     const PetscScalar *vals;
104 
105     for (i = 0; i < nm; i++) {
106       PetscCall(MatGetRow(X[i], r, &ncols, &row, &vals));
107       PetscCall(MatSetValues(preallocator, 1, &r, ncols, row, vals, INSERT_VALUES));
108       if (symm && symm[i]) PetscCall(MatSetValues(preallocator, ncols, row, 1, &r, vals, INSERT_VALUES));
109       PetscCall(MatRestoreRow(X[i], r, &ncols, &row, &vals));
110     }
111   }
112   PetscCall(MatAssemblyBegin(preallocator, MAT_FINAL_ASSEMBLY));
113   PetscCall(MatAssemblyEnd(preallocator, MAT_FINAL_ASSEMBLY));
114   PetscCall(MatPreallocatorPreallocate(preallocator, fill, B));
115   PetscCall(MatDestroy(&preallocator));
116   PetscFunctionReturn(PETSC_SUCCESS);
117 }
118 
119 PETSC_INTERN PetscErrorCode MatConvert_MPISBAIJ_Basic(Mat A, MatType newtype, MatReuse reuse, Mat *newmat)
120 {
121   Mat      B;
122   PetscInt r;
123 
124   PetscFunctionBegin;
125   if (reuse != MAT_REUSE_MATRIX) {
126     PetscBool symm = PETSC_TRUE, isdense;
127     PetscInt  bs;
128 
129     PetscCall(MatCreate(PetscObjectComm((PetscObject)A), &B));
130     PetscCall(MatSetSizes(B, A->rmap->n, A->cmap->n, A->rmap->N, A->cmap->N));
131     PetscCall(MatSetType(B, newtype));
132     PetscCall(MatGetBlockSize(A, &bs));
133     PetscCall(MatSetBlockSize(B, bs));
134     PetscCall(PetscLayoutSetUp(B->rmap));
135     PetscCall(PetscLayoutSetUp(B->cmap));
136     PetscCall(PetscObjectTypeCompareAny((PetscObject)B, &isdense, MATSEQDENSE, MATMPIDENSE, MATSEQDENSECUDA, ""));
137     if (!isdense) {
138       PetscCall(MatGetRowUpperTriangular(A));
139       PetscCall(MatPreallocateWithMats_Private(B, 1, &A, &symm, PETSC_TRUE));
140       PetscCall(MatRestoreRowUpperTriangular(A));
141     } else {
142       PetscCall(MatSetUp(B));
143     }
144   } else {
145     B = *newmat;
146     PetscCall(MatZeroEntries(B));
147   }
148 
149   PetscCall(MatGetRowUpperTriangular(A));
150   for (r = A->rmap->rstart; r < A->rmap->rend; r++) {
151     PetscInt           ncols;
152     const PetscInt    *row;
153     const PetscScalar *vals;
154 
155     PetscCall(MatGetRow(A, r, &ncols, &row, &vals));
156     PetscCall(MatSetValues(B, 1, &r, ncols, row, vals, INSERT_VALUES));
157 #if defined(PETSC_USE_COMPLEX)
158     if (A->hermitian == PETSC_BOOL3_TRUE) {
159       PetscInt i;
160       for (i = 0; i < ncols; i++) PetscCall(MatSetValue(B, row[i], r, PetscConj(vals[i]), INSERT_VALUES));
161     } else {
162       PetscCall(MatSetValues(B, ncols, row, 1, &r, vals, INSERT_VALUES));
163     }
164 #else
165     PetscCall(MatSetValues(B, ncols, row, 1, &r, vals, INSERT_VALUES));
166 #endif
167     PetscCall(MatRestoreRow(A, r, &ncols, &row, &vals));
168   }
169   PetscCall(MatRestoreRowUpperTriangular(A));
170   PetscCall(MatAssemblyBegin(B, MAT_FINAL_ASSEMBLY));
171   PetscCall(MatAssemblyEnd(B, MAT_FINAL_ASSEMBLY));
172 
173   if (reuse == MAT_INPLACE_MATRIX) {
174     PetscCall(MatHeaderReplace(A, &B));
175   } else {
176     *newmat = B;
177   }
178   PetscFunctionReturn(PETSC_SUCCESS);
179 }
180 
181 static PetscErrorCode MatStoreValues_MPISBAIJ(Mat mat)
182 {
183   Mat_MPISBAIJ *aij = (Mat_MPISBAIJ *)mat->data;
184 
185   PetscFunctionBegin;
186   PetscCall(MatStoreValues(aij->A));
187   PetscCall(MatStoreValues(aij->B));
188   PetscFunctionReturn(PETSC_SUCCESS);
189 }
190 
191 static PetscErrorCode MatRetrieveValues_MPISBAIJ(Mat mat)
192 {
193   Mat_MPISBAIJ *aij = (Mat_MPISBAIJ *)mat->data;
194 
195   PetscFunctionBegin;
196   PetscCall(MatRetrieveValues(aij->A));
197   PetscCall(MatRetrieveValues(aij->B));
198   PetscFunctionReturn(PETSC_SUCCESS);
199 }
200 
201 #define MatSetValues_SeqSBAIJ_A_Private(row, col, value, addv, orow, ocol) \
202   do { \
203     brow = row / bs; \
204     rp   = aj + ai[brow]; \
205     ap   = aa + bs2 * ai[brow]; \
206     rmax = aimax[brow]; \
207     nrow = ailen[brow]; \
208     bcol = col / bs; \
209     ridx = row % bs; \
210     cidx = col % bs; \
211     low  = 0; \
212     high = nrow; \
213     while (high - low > 3) { \
214       t = (low + high) / 2; \
215       if (rp[t] > bcol) high = t; \
216       else low = t; \
217     } \
218     for (_i = low; _i < high; _i++) { \
219       if (rp[_i] > bcol) break; \
220       if (rp[_i] == bcol) { \
221         bap = ap + bs2 * _i + bs * cidx + ridx; \
222         if (addv == ADD_VALUES) *bap += value; \
223         else *bap = value; \
224         goto a_noinsert; \
225       } \
226     } \
227     if (a->nonew == 1) goto a_noinsert; \
228     PetscCheck(a->nonew != -1, PETSC_COMM_SELF, PETSC_ERR_ARG_OUTOFRANGE, "Inserting a new nonzero at global row/column (%" PetscInt_FMT ", %" PetscInt_FMT ") into matrix", orow, ocol); \
229     MatSeqXAIJReallocateAIJ(A, a->mbs, bs2, nrow, brow, bcol, rmax, aa, ai, aj, rp, ap, aimax, a->nonew, MatScalar); \
230     N = nrow++ - 1; \
231     /* shift up all the later entries in this row */ \
232     PetscCall(PetscArraymove(rp + _i + 1, rp + _i, N - _i + 1)); \
233     PetscCall(PetscArraymove(ap + bs2 * (_i + 1), ap + bs2 * _i, bs2 * (N - _i + 1))); \
234     PetscCall(PetscArrayzero(ap + bs2 * _i, bs2)); \
235     rp[_i]                          = bcol; \
236     ap[bs2 * _i + bs * cidx + ridx] = value; \
237   a_noinsert:; \
238     ailen[brow] = nrow; \
239   } while (0)
240 
241 #define MatSetValues_SeqSBAIJ_B_Private(row, col, value, addv, orow, ocol) \
242   do { \
243     brow = row / bs; \
244     rp   = bj + bi[brow]; \
245     ap   = ba + bs2 * bi[brow]; \
246     rmax = bimax[brow]; \
247     nrow = bilen[brow]; \
248     bcol = col / bs; \
249     ridx = row % bs; \
250     cidx = col % bs; \
251     low  = 0; \
252     high = nrow; \
253     while (high - low > 3) { \
254       t = (low + high) / 2; \
255       if (rp[t] > bcol) high = t; \
256       else low = t; \
257     } \
258     for (_i = low; _i < high; _i++) { \
259       if (rp[_i] > bcol) break; \
260       if (rp[_i] == bcol) { \
261         bap = ap + bs2 * _i + bs * cidx + ridx; \
262         if (addv == ADD_VALUES) *bap += value; \
263         else *bap = value; \
264         goto b_noinsert; \
265       } \
266     } \
267     if (b->nonew == 1) goto b_noinsert; \
268     PetscCheck(b->nonew != -1, PETSC_COMM_SELF, PETSC_ERR_ARG_OUTOFRANGE, "Inserting a new nonzero at global row/column (%" PetscInt_FMT ", %" PetscInt_FMT ") into matrix", orow, ocol); \
269     MatSeqXAIJReallocateAIJ(B, b->mbs, bs2, nrow, brow, bcol, rmax, ba, bi, bj, rp, ap, bimax, b->nonew, MatScalar); \
270     N = nrow++ - 1; \
271     /* shift up all the later entries in this row */ \
272     PetscCall(PetscArraymove(rp + _i + 1, rp + _i, N - _i + 1)); \
273     PetscCall(PetscArraymove(ap + bs2 * (_i + 1), ap + bs2 * _i, bs2 * (N - _i + 1))); \
274     PetscCall(PetscArrayzero(ap + bs2 * _i, bs2)); \
275     rp[_i]                          = bcol; \
276     ap[bs2 * _i + bs * cidx + ridx] = value; \
277   b_noinsert:; \
278     bilen[brow] = nrow; \
279   } while (0)
280 
281 /* Only add/insert a(i,j) with i<=j (blocks).
282    Any a(i,j) with i>j input by user is ignored or generates an error
283 */
284 static PetscErrorCode MatSetValues_MPISBAIJ(Mat mat, PetscInt m, const PetscInt im[], PetscInt n, const PetscInt in[], const PetscScalar v[], InsertMode addv)
285 {
286   Mat_MPISBAIJ *baij = (Mat_MPISBAIJ *)mat->data;
287   MatScalar     value;
288   PetscBool     roworiented = baij->roworiented;
289   PetscInt      i, j, row, col;
290   PetscInt      rstart_orig = mat->rmap->rstart;
291   PetscInt      rend_orig = mat->rmap->rend, cstart_orig = mat->cmap->rstart;
292   PetscInt      cend_orig = mat->cmap->rend, bs = mat->rmap->bs;
293 
294   /* Some Variables required in the macro */
295   Mat           A     = baij->A;
296   Mat_SeqSBAIJ *a     = (Mat_SeqSBAIJ *)A->data;
297   PetscInt     *aimax = a->imax, *ai = a->i, *ailen = a->ilen, *aj = a->j;
298   MatScalar    *aa = a->a;
299 
300   Mat          B     = baij->B;
301   Mat_SeqBAIJ *b     = (Mat_SeqBAIJ *)B->data;
302   PetscInt    *bimax = b->imax, *bi = b->i, *bilen = b->ilen, *bj = b->j;
303   MatScalar   *ba = b->a;
304 
305   PetscInt  *rp, ii, nrow, _i, rmax, N, brow, bcol;
306   PetscInt   low, high, t, ridx, cidx, bs2 = a->bs2;
307   MatScalar *ap, *bap;
308 
309   /* for stash */
310   PetscInt   n_loc, *in_loc = NULL;
311   MatScalar *v_loc = NULL;
312 
313   PetscFunctionBegin;
314   if (!baij->donotstash) {
315     if (n > baij->n_loc) {
316       PetscCall(PetscFree(baij->in_loc));
317       PetscCall(PetscFree(baij->v_loc));
318       PetscCall(PetscMalloc1(n, &baij->in_loc));
319       PetscCall(PetscMalloc1(n, &baij->v_loc));
320 
321       baij->n_loc = n;
322     }
323     in_loc = baij->in_loc;
324     v_loc  = baij->v_loc;
325   }
326 
327   for (i = 0; i < m; i++) {
328     if (im[i] < 0) continue;
329     PetscCheck(im[i] < mat->rmap->N, PETSC_COMM_SELF, PETSC_ERR_ARG_OUTOFRANGE, "Row too large: row %" PetscInt_FMT " max %" PetscInt_FMT, im[i], mat->rmap->N - 1);
330     if (im[i] >= rstart_orig && im[i] < rend_orig) { /* this processor entry */
331       row = im[i] - rstart_orig;                     /* local row index */
332       for (j = 0; j < n; j++) {
333         if (im[i] / bs > in[j] / bs) {
334           if (a->ignore_ltriangular) {
335             continue; /* ignore lower triangular blocks */
336           } else SETERRQ(PETSC_COMM_SELF, PETSC_ERR_USER, "Lower triangular value cannot be set for sbaij format. Ignoring these values, run with -mat_ignore_lower_triangular or call MatSetOption(mat,MAT_IGNORE_LOWER_TRIANGULAR,PETSC_TRUE)");
337         }
338         if (in[j] >= cstart_orig && in[j] < cend_orig) { /* diag entry (A) */
339           col  = in[j] - cstart_orig;                    /* local col index */
340           brow = row / bs;
341           bcol = col / bs;
342           if (brow > bcol) continue; /* ignore lower triangular blocks of A */
343           if (roworiented) value = v[i * n + j];
344           else value = v[i + j * m];
345           MatSetValues_SeqSBAIJ_A_Private(row, col, value, addv, im[i], in[j]);
346           /* PetscCall(MatSetValues_SeqBAIJ(baij->A,1,&row,1,&col,&value,addv)); */
347         } else if (in[j] < 0) {
348           continue;
349         } else {
350           PetscCheck(in[j] < mat->cmap->N, PETSC_COMM_SELF, PETSC_ERR_ARG_OUTOFRANGE, "Column too large: col %" PetscInt_FMT " max %" PetscInt_FMT, in[j], mat->cmap->N - 1);
351           /* off-diag entry (B) */
352           if (mat->was_assembled) {
353             if (!baij->colmap) PetscCall(MatCreateColmap_MPIBAIJ_Private(mat));
354 #if defined(PETSC_USE_CTABLE)
355             PetscCall(PetscHMapIGetWithDefault(baij->colmap, in[j] / bs + 1, 0, &col));
356             col = col - 1;
357 #else
358             col = baij->colmap[in[j] / bs] - 1;
359 #endif
360             if (col < 0 && !((Mat_SeqSBAIJ *)baij->A->data)->nonew) {
361               PetscCall(MatDisAssemble_MPISBAIJ(mat));
362               col = in[j];
363               /* Reinitialize the variables required by MatSetValues_SeqBAIJ_B_Private() */
364               B     = baij->B;
365               b     = (Mat_SeqBAIJ *)B->data;
366               bimax = b->imax;
367               bi    = b->i;
368               bilen = b->ilen;
369               bj    = b->j;
370               ba    = b->a;
371             } else col += in[j] % bs;
372           } else col = in[j];
373           if (roworiented) value = v[i * n + j];
374           else value = v[i + j * m];
375           MatSetValues_SeqSBAIJ_B_Private(row, col, value, addv, im[i], in[j]);
376           /* PetscCall(MatSetValues_SeqBAIJ(baij->B,1,&row,1,&col,&value,addv)); */
377         }
378       }
379     } else { /* off processor entry */
380       PetscCheck(!mat->nooffprocentries, PETSC_COMM_SELF, PETSC_ERR_ARG_WRONG, "Setting off process row %" PetscInt_FMT " even though MatSetOption(,MAT_NO_OFF_PROC_ENTRIES,PETSC_TRUE) was set", im[i]);
381       if (!baij->donotstash) {
382         mat->assembled = PETSC_FALSE;
383         n_loc          = 0;
384         for (j = 0; j < n; j++) {
385           if (im[i] / bs > in[j] / bs) continue; /* ignore lower triangular blocks */
386           in_loc[n_loc] = in[j];
387           if (roworiented) {
388             v_loc[n_loc] = v[i * n + j];
389           } else {
390             v_loc[n_loc] = v[j * m + i];
391           }
392           n_loc++;
393         }
394         PetscCall(MatStashValuesRow_Private(&mat->stash, im[i], n_loc, in_loc, v_loc, PETSC_FALSE));
395       }
396     }
397   }
398   PetscFunctionReturn(PETSC_SUCCESS);
399 }
400 
401 static inline PetscErrorCode MatSetValuesBlocked_SeqSBAIJ_Inlined(Mat A, PetscInt row, PetscInt col, const PetscScalar v[], InsertMode is, PetscInt orow, PetscInt ocol)
402 {
403   Mat_SeqSBAIJ      *a = (Mat_SeqSBAIJ *)A->data;
404   PetscInt          *rp, low, high, t, ii, jj, nrow, i, rmax, N;
405   PetscInt          *imax = a->imax, *ai = a->i, *ailen = a->ilen;
406   PetscInt          *aj = a->j, nonew = a->nonew, bs2 = a->bs2, bs = A->rmap->bs;
407   PetscBool          roworiented = a->roworiented;
408   const PetscScalar *value       = v;
409   MatScalar         *ap, *aa = a->a, *bap;
410 
411   PetscFunctionBegin;
412   if (col < row) {
413     if (a->ignore_ltriangular) PetscFunctionReturn(PETSC_SUCCESS); /* ignore lower triangular block */
414     else SETERRQ(PETSC_COMM_SELF, PETSC_ERR_USER, "Lower triangular value cannot be set for sbaij format. Ignoring these values, run with -mat_ignore_lower_triangular or call MatSetOption(mat,MAT_IGNORE_LOWER_TRIANGULAR,PETSC_TRUE)");
415   }
416   rp    = aj + ai[row];
417   ap    = aa + bs2 * ai[row];
418   rmax  = imax[row];
419   nrow  = ailen[row];
420   value = v;
421   low   = 0;
422   high  = nrow;
423 
424   while (high - low > 7) {
425     t = (low + high) / 2;
426     if (rp[t] > col) high = t;
427     else low = t;
428   }
429   for (i = low; i < high; i++) {
430     if (rp[i] > col) break;
431     if (rp[i] == col) {
432       bap = ap + bs2 * i;
433       if (roworiented) {
434         if (is == ADD_VALUES) {
435           for (ii = 0; ii < bs; ii++) {
436             for (jj = ii; jj < bs2; jj += bs) bap[jj] += *value++;
437           }
438         } else {
439           for (ii = 0; ii < bs; ii++) {
440             for (jj = ii; jj < bs2; jj += bs) bap[jj] = *value++;
441           }
442         }
443       } else {
444         if (is == ADD_VALUES) {
445           for (ii = 0; ii < bs; ii++) {
446             for (jj = 0; jj < bs; jj++) *bap++ += *value++;
447           }
448         } else {
449           for (ii = 0; ii < bs; ii++) {
450             for (jj = 0; jj < bs; jj++) *bap++ = *value++;
451           }
452         }
453       }
454       goto noinsert2;
455     }
456   }
457   if (nonew == 1) goto noinsert2;
458   PetscCheck(nonew != -1, PETSC_COMM_SELF, PETSC_ERR_ARG_OUTOFRANGE, "Inserting a new block index nonzero block (%" PetscInt_FMT ", %" PetscInt_FMT ") in the matrix", orow, ocol);
459   MatSeqXAIJReallocateAIJ(A, a->mbs, bs2, nrow, row, col, rmax, aa, ai, aj, rp, ap, imax, nonew, MatScalar);
460   N = nrow++ - 1;
461   high++;
462   /* shift up all the later entries in this row */
463   PetscCall(PetscArraymove(rp + i + 1, rp + i, N - i + 1));
464   PetscCall(PetscArraymove(ap + bs2 * (i + 1), ap + bs2 * i, bs2 * (N - i + 1)));
465   rp[i] = col;
466   bap   = ap + bs2 * i;
467   if (roworiented) {
468     for (ii = 0; ii < bs; ii++) {
469       for (jj = ii; jj < bs2; jj += bs) bap[jj] = *value++;
470     }
471   } else {
472     for (ii = 0; ii < bs; ii++) {
473       for (jj = 0; jj < bs; jj++) *bap++ = *value++;
474     }
475   }
476 noinsert2:;
477   ailen[row] = nrow;
478   PetscFunctionReturn(PETSC_SUCCESS);
479 }
480 
481 /*
482    This routine is exactly duplicated in mpibaij.c
483 */
484 static inline PetscErrorCode MatSetValuesBlocked_SeqBAIJ_Inlined(Mat A, PetscInt row, PetscInt col, const PetscScalar v[], InsertMode is, PetscInt orow, PetscInt ocol)
485 {
486   Mat_SeqBAIJ       *a = (Mat_SeqBAIJ *)A->data;
487   PetscInt          *rp, low, high, t, ii, jj, nrow, i, rmax, N;
488   PetscInt          *imax = a->imax, *ai = a->i, *ailen = a->ilen;
489   PetscInt          *aj = a->j, nonew = a->nonew, bs2 = a->bs2, bs = A->rmap->bs;
490   PetscBool          roworiented = a->roworiented;
491   const PetscScalar *value       = v;
492   MatScalar         *ap, *aa = a->a, *bap;
493 
494   PetscFunctionBegin;
495   rp    = aj + ai[row];
496   ap    = aa + bs2 * ai[row];
497   rmax  = imax[row];
498   nrow  = ailen[row];
499   low   = 0;
500   high  = nrow;
501   value = v;
502   while (high - low > 7) {
503     t = (low + high) / 2;
504     if (rp[t] > col) high = t;
505     else low = t;
506   }
507   for (i = low; i < high; i++) {
508     if (rp[i] > col) break;
509     if (rp[i] == col) {
510       bap = ap + bs2 * i;
511       if (roworiented) {
512         if (is == ADD_VALUES) {
513           for (ii = 0; ii < bs; ii++) {
514             for (jj = ii; jj < bs2; jj += bs) bap[jj] += *value++;
515           }
516         } else {
517           for (ii = 0; ii < bs; ii++) {
518             for (jj = ii; jj < bs2; jj += bs) bap[jj] = *value++;
519           }
520         }
521       } else {
522         if (is == ADD_VALUES) {
523           for (ii = 0; ii < bs; ii++, value += bs) {
524             for (jj = 0; jj < bs; jj++) bap[jj] += value[jj];
525             bap += bs;
526           }
527         } else {
528           for (ii = 0; ii < bs; ii++, value += bs) {
529             for (jj = 0; jj < bs; jj++) bap[jj] = value[jj];
530             bap += bs;
531           }
532         }
533       }
534       goto noinsert2;
535     }
536   }
537   if (nonew == 1) goto noinsert2;
538   PetscCheck(nonew != -1, PETSC_COMM_SELF, PETSC_ERR_ARG_OUTOFRANGE, "Inserting a new global block indexed nonzero block (%" PetscInt_FMT ", %" PetscInt_FMT ") in the matrix", orow, ocol);
539   MatSeqXAIJReallocateAIJ(A, a->mbs, bs2, nrow, row, col, rmax, aa, ai, aj, rp, ap, imax, nonew, MatScalar);
540   N = nrow++ - 1;
541   high++;
542   /* shift up all the later entries in this row */
543   PetscCall(PetscArraymove(rp + i + 1, rp + i, N - i + 1));
544   PetscCall(PetscArraymove(ap + bs2 * (i + 1), ap + bs2 * i, bs2 * (N - i + 1)));
545   rp[i] = col;
546   bap   = ap + bs2 * i;
547   if (roworiented) {
548     for (ii = 0; ii < bs; ii++) {
549       for (jj = ii; jj < bs2; jj += bs) bap[jj] = *value++;
550     }
551   } else {
552     for (ii = 0; ii < bs; ii++) {
553       for (jj = 0; jj < bs; jj++) *bap++ = *value++;
554     }
555   }
556 noinsert2:;
557   ailen[row] = nrow;
558   PetscFunctionReturn(PETSC_SUCCESS);
559 }
560 
561 /*
562     This routine could be optimized by removing the need for the block copy below and passing stride information
563   to the above inline routines; similarly in MatSetValuesBlocked_MPIBAIJ()
564 */
565 static PetscErrorCode MatSetValuesBlocked_MPISBAIJ(Mat mat, PetscInt m, const PetscInt im[], PetscInt n, const PetscInt in[], const MatScalar v[], InsertMode addv)
566 {
567   Mat_MPISBAIJ    *baij = (Mat_MPISBAIJ *)mat->data;
568   const MatScalar *value;
569   MatScalar       *barray      = baij->barray;
570   PetscBool        roworiented = baij->roworiented, ignore_ltriangular = ((Mat_SeqSBAIJ *)baij->A->data)->ignore_ltriangular;
571   PetscInt         i, j, ii, jj, row, col, rstart = baij->rstartbs;
572   PetscInt         rend = baij->rendbs, cstart = baij->cstartbs, stepval;
573   PetscInt         cend = baij->cendbs, bs = mat->rmap->bs, bs2 = baij->bs2;
574 
575   PetscFunctionBegin;
576   if (!barray) {
577     PetscCall(PetscMalloc1(bs2, &barray));
578     baij->barray = barray;
579   }
580 
581   if (roworiented) {
582     stepval = (n - 1) * bs;
583   } else {
584     stepval = (m - 1) * bs;
585   }
586   for (i = 0; i < m; i++) {
587     if (im[i] < 0) continue;
588     PetscCheck(im[i] < baij->Mbs, PETSC_COMM_SELF, PETSC_ERR_ARG_OUTOFRANGE, "Block indexed row too large %" PetscInt_FMT " max %" PetscInt_FMT, im[i], baij->Mbs - 1);
589     if (im[i] >= rstart && im[i] < rend) {
590       row = im[i] - rstart;
591       for (j = 0; j < n; j++) {
592         if (im[i] > in[j]) {
593           if (ignore_ltriangular) continue; /* ignore lower triangular blocks */
594           else SETERRQ(PETSC_COMM_SELF, PETSC_ERR_USER, "Lower triangular value cannot be set for sbaij format. Ignoring these values, run with -mat_ignore_lower_triangular or call MatSetOption(mat,MAT_IGNORE_LOWER_TRIANGULAR,PETSC_TRUE)");
595         }
596         /* If NumCol = 1 then a copy is not required */
597         if ((roworiented) && (n == 1)) {
598           barray = (MatScalar *)v + i * bs2;
599         } else if ((!roworiented) && (m == 1)) {
600           barray = (MatScalar *)v + j * bs2;
601         } else { /* Here a copy is required */
602           if (roworiented) {
603             value = v + i * (stepval + bs) * bs + j * bs;
604           } else {
605             value = v + j * (stepval + bs) * bs + i * bs;
606           }
607           for (ii = 0; ii < bs; ii++, value += stepval) {
608             for (jj = 0; jj < bs; jj++) *barray++ = *value++;
609           }
610           barray -= bs2;
611         }
612 
613         if (in[j] >= cstart && in[j] < cend) {
614           col = in[j] - cstart;
615           PetscCall(MatSetValuesBlocked_SeqSBAIJ_Inlined(baij->A, row, col, barray, addv, im[i], in[j]));
616         } else if (in[j] < 0) {
617           continue;
618         } else {
619           PetscCheck(in[j] < baij->Nbs, PETSC_COMM_SELF, PETSC_ERR_ARG_OUTOFRANGE, "Block indexed column too large %" PetscInt_FMT " max %" PetscInt_FMT, in[j], baij->Nbs - 1);
620           if (mat->was_assembled) {
621             if (!baij->colmap) PetscCall(MatCreateColmap_MPIBAIJ_Private(mat));
622 
623 #if defined(PETSC_USE_CTABLE)
624             PetscCall(PetscHMapIGetWithDefault(baij->colmap, in[j] + 1, 0, &col));
625             col = col < 1 ? -1 : (col - 1) / bs;
626 #else
627             col = baij->colmap[in[j]] < 1 ? -1 : (baij->colmap[in[j]] - 1) / bs;
628 #endif
629             if (col < 0 && !((Mat_SeqBAIJ *)baij->A->data)->nonew) {
630               PetscCall(MatDisAssemble_MPISBAIJ(mat));
631               col = in[j];
632             }
633           } else col = in[j];
634           PetscCall(MatSetValuesBlocked_SeqBAIJ_Inlined(baij->B, row, col, barray, addv, im[i], in[j]));
635         }
636       }
637     } else {
638       PetscCheck(!mat->nooffprocentries, PETSC_COMM_SELF, PETSC_ERR_ARG_WRONG, "Setting off process block indexed row %" PetscInt_FMT " even though MatSetOption(,MAT_NO_OFF_PROC_ENTRIES,PETSC_TRUE) was set", im[i]);
639       if (!baij->donotstash) {
640         if (roworiented) {
641           PetscCall(MatStashValuesRowBlocked_Private(&mat->bstash, im[i], n, in, v, m, n, i));
642         } else {
643           PetscCall(MatStashValuesColBlocked_Private(&mat->bstash, im[i], n, in, v, m, n, i));
644         }
645       }
646     }
647   }
648   PetscFunctionReturn(PETSC_SUCCESS);
649 }
650 
651 static PetscErrorCode MatGetValues_MPISBAIJ(Mat mat, PetscInt m, const PetscInt idxm[], PetscInt n, const PetscInt idxn[], PetscScalar v[])
652 {
653   Mat_MPISBAIJ *baij = (Mat_MPISBAIJ *)mat->data;
654   PetscInt      bs = mat->rmap->bs, i, j, bsrstart = mat->rmap->rstart, bsrend = mat->rmap->rend;
655   PetscInt      bscstart = mat->cmap->rstart, bscend = mat->cmap->rend, row, col, data;
656 
657   PetscFunctionBegin;
658   for (i = 0; i < m; i++) {
659     if (idxm[i] < 0) continue; /* negative row */
660     PetscCheck(idxm[i] < mat->rmap->N, PETSC_COMM_SELF, PETSC_ERR_ARG_OUTOFRANGE, "Row too large: row %" PetscInt_FMT " max %" PetscInt_FMT, idxm[i], mat->rmap->N - 1);
661     if (idxm[i] >= bsrstart && idxm[i] < bsrend) {
662       row = idxm[i] - bsrstart;
663       for (j = 0; j < n; j++) {
664         if (idxn[j] < 0) continue; /* negative column */
665         PetscCheck(idxn[j] < mat->cmap->N, PETSC_COMM_SELF, PETSC_ERR_ARG_OUTOFRANGE, "Column too large: col %" PetscInt_FMT " max %" PetscInt_FMT, idxn[j], mat->cmap->N - 1);
666         if (idxn[j] >= bscstart && idxn[j] < bscend) {
667           col = idxn[j] - bscstart;
668           PetscCall(MatGetValues_SeqSBAIJ(baij->A, 1, &row, 1, &col, v + i * n + j));
669         } else {
670           if (!baij->colmap) PetscCall(MatCreateColmap_MPIBAIJ_Private(mat));
671 #if defined(PETSC_USE_CTABLE)
672           PetscCall(PetscHMapIGetWithDefault(baij->colmap, idxn[j] / bs + 1, 0, &data));
673           data--;
674 #else
675           data = baij->colmap[idxn[j] / bs] - 1;
676 #endif
677           if ((data < 0) || (baij->garray[data / bs] != idxn[j] / bs)) *(v + i * n + j) = 0.0;
678           else {
679             col = data + idxn[j] % bs;
680             PetscCall(MatGetValues_SeqBAIJ(baij->B, 1, &row, 1, &col, v + i * n + j));
681           }
682         }
683       }
684     } else SETERRQ(PETSC_COMM_SELF, PETSC_ERR_SUP, "Only local values currently supported");
685   }
686   PetscFunctionReturn(PETSC_SUCCESS);
687 }
688 
689 static PetscErrorCode MatNorm_MPISBAIJ(Mat mat, NormType type, PetscReal *norm)
690 {
691   Mat_MPISBAIJ *baij = (Mat_MPISBAIJ *)mat->data;
692   PetscReal     sum[2], *lnorm2;
693 
694   PetscFunctionBegin;
695   if (baij->size == 1) {
696     PetscCall(MatNorm(baij->A, type, norm));
697   } else {
698     if (type == NORM_FROBENIUS) {
699       PetscCall(PetscMalloc1(2, &lnorm2));
700       PetscCall(MatNorm(baij->A, type, lnorm2));
701       *lnorm2 = (*lnorm2) * (*lnorm2);
702       lnorm2++; /* squar power of norm(A) */
703       PetscCall(MatNorm(baij->B, type, lnorm2));
704       *lnorm2 = (*lnorm2) * (*lnorm2);
705       lnorm2--; /* squar power of norm(B) */
706       PetscCallMPI(MPIU_Allreduce(lnorm2, sum, 2, MPIU_REAL, MPIU_SUM, PetscObjectComm((PetscObject)mat)));
707       *norm = PetscSqrtReal(sum[0] + 2 * sum[1]);
708       PetscCall(PetscFree(lnorm2));
709     } else if (type == NORM_INFINITY || type == NORM_1) { /* max row/column sum */
710       Mat_SeqSBAIJ *amat = (Mat_SeqSBAIJ *)baij->A->data;
711       Mat_SeqBAIJ  *bmat = (Mat_SeqBAIJ *)baij->B->data;
712       PetscReal    *rsum, *rsum2, vabs;
713       PetscInt     *jj, *garray = baij->garray, rstart = baij->rstartbs, nz;
714       PetscInt      brow, bcol, col, bs = baij->A->rmap->bs, row, grow, gcol, mbs = amat->mbs;
715       MatScalar    *v;
716       PetscMPIInt   iN;
717 
718       PetscCall(PetscMalloc2(mat->cmap->N, &rsum, mat->cmap->N, &rsum2));
719       PetscCall(PetscArrayzero(rsum, mat->cmap->N));
720       /* Amat */
721       v  = amat->a;
722       jj = amat->j;
723       for (brow = 0; brow < mbs; brow++) {
724         grow = bs * (rstart + brow);
725         nz   = amat->i[brow + 1] - amat->i[brow];
726         for (bcol = 0; bcol < nz; bcol++) {
727           gcol = bs * (rstart + *jj);
728           jj++;
729           for (col = 0; col < bs; col++) {
730             for (row = 0; row < bs; row++) {
731               vabs = PetscAbsScalar(*v);
732               v++;
733               rsum[gcol + col] += vabs;
734               /* non-diagonal block */
735               if (bcol > 0 && vabs > 0.0) rsum[grow + row] += vabs;
736             }
737           }
738         }
739         PetscCall(PetscLogFlops(nz * bs * bs));
740       }
741       /* Bmat */
742       v  = bmat->a;
743       jj = bmat->j;
744       for (brow = 0; brow < mbs; brow++) {
745         grow = bs * (rstart + brow);
746         nz   = bmat->i[brow + 1] - bmat->i[brow];
747         for (bcol = 0; bcol < nz; bcol++) {
748           gcol = bs * garray[*jj];
749           jj++;
750           for (col = 0; col < bs; col++) {
751             for (row = 0; row < bs; row++) {
752               vabs = PetscAbsScalar(*v);
753               v++;
754               rsum[gcol + col] += vabs;
755               rsum[grow + row] += vabs;
756             }
757           }
758         }
759         PetscCall(PetscLogFlops(nz * bs * bs));
760       }
761       PetscCall(PetscMPIIntCast(mat->cmap->N, &iN));
762       PetscCallMPI(MPIU_Allreduce(rsum, rsum2, iN, MPIU_REAL, MPIU_SUM, PetscObjectComm((PetscObject)mat)));
763       *norm = 0.0;
764       for (col = 0; col < mat->cmap->N; col++) {
765         if (rsum2[col] > *norm) *norm = rsum2[col];
766       }
767       PetscCall(PetscFree2(rsum, rsum2));
768     } else SETERRQ(PETSC_COMM_SELF, PETSC_ERR_SUP, "No support for this norm yet");
769   }
770   PetscFunctionReturn(PETSC_SUCCESS);
771 }
772 
773 static PetscErrorCode MatAssemblyBegin_MPISBAIJ(Mat mat, MatAssemblyType mode)
774 {
775   Mat_MPISBAIJ *baij = (Mat_MPISBAIJ *)mat->data;
776   PetscInt      nstash, reallocs;
777 
778   PetscFunctionBegin;
779   if (baij->donotstash || mat->nooffprocentries) PetscFunctionReturn(PETSC_SUCCESS);
780 
781   PetscCall(MatStashScatterBegin_Private(mat, &mat->stash, mat->rmap->range));
782   PetscCall(MatStashScatterBegin_Private(mat, &mat->bstash, baij->rangebs));
783   PetscCall(MatStashGetInfo_Private(&mat->stash, &nstash, &reallocs));
784   PetscCall(PetscInfo(mat, "Stash has %" PetscInt_FMT " entries,uses %" PetscInt_FMT " mallocs.\n", nstash, reallocs));
785   PetscCall(MatStashGetInfo_Private(&mat->stash, &nstash, &reallocs));
786   PetscCall(PetscInfo(mat, "Block-Stash has %" PetscInt_FMT " entries, uses %" PetscInt_FMT " mallocs.\n", nstash, reallocs));
787   PetscFunctionReturn(PETSC_SUCCESS);
788 }
789 
790 static PetscErrorCode MatAssemblyEnd_MPISBAIJ(Mat mat, MatAssemblyType mode)
791 {
792   Mat_MPISBAIJ *baij = (Mat_MPISBAIJ *)mat->data;
793   Mat_SeqSBAIJ *a    = (Mat_SeqSBAIJ *)baij->A->data;
794   PetscInt      i, j, rstart, ncols, flg, bs2 = baij->bs2;
795   PetscInt     *row, *col;
796   PetscBool     other_disassembled;
797   PetscMPIInt   n;
798   PetscBool     r1, r2, r3;
799   MatScalar    *val;
800 
801   /* do not use 'b=(Mat_SeqBAIJ*)baij->B->data' as B can be reset in disassembly */
802   PetscFunctionBegin;
803   if (!baij->donotstash && !mat->nooffprocentries) {
804     while (1) {
805       PetscCall(MatStashScatterGetMesg_Private(&mat->stash, &n, &row, &col, &val, &flg));
806       if (!flg) break;
807 
808       for (i = 0; i < n;) {
809         /* Now identify the consecutive vals belonging to the same row */
810         for (j = i, rstart = row[j]; j < n; j++) {
811           if (row[j] != rstart) break;
812         }
813         if (j < n) ncols = j - i;
814         else ncols = n - i;
815         /* Now assemble all these values with a single function call */
816         PetscCall(MatSetValues_MPISBAIJ(mat, 1, row + i, ncols, col + i, val + i, mat->insertmode));
817         i = j;
818       }
819     }
820     PetscCall(MatStashScatterEnd_Private(&mat->stash));
821     /* Now process the block-stash. Since the values are stashed column-oriented,
822        set the row-oriented flag to column-oriented, and after MatSetValues()
823        restore the original flags */
824     r1 = baij->roworiented;
825     r2 = a->roworiented;
826     r3 = ((Mat_SeqBAIJ *)baij->B->data)->roworiented;
827 
828     baij->roworiented = PETSC_FALSE;
829     a->roworiented    = PETSC_FALSE;
830 
831     ((Mat_SeqBAIJ *)baij->B->data)->roworiented = PETSC_FALSE; /* b->roworinted */
832     while (1) {
833       PetscCall(MatStashScatterGetMesg_Private(&mat->bstash, &n, &row, &col, &val, &flg));
834       if (!flg) break;
835 
836       for (i = 0; i < n;) {
837         /* Now identify the consecutive vals belonging to the same row */
838         for (j = i, rstart = row[j]; j < n; j++) {
839           if (row[j] != rstart) break;
840         }
841         if (j < n) ncols = j - i;
842         else ncols = n - i;
843         PetscCall(MatSetValuesBlocked_MPISBAIJ(mat, 1, row + i, ncols, col + i, val + i * bs2, mat->insertmode));
844         i = j;
845       }
846     }
847     PetscCall(MatStashScatterEnd_Private(&mat->bstash));
848 
849     baij->roworiented = r1;
850     a->roworiented    = r2;
851 
852     ((Mat_SeqBAIJ *)baij->B->data)->roworiented = r3; /* b->roworinted */
853   }
854 
855   PetscCall(MatAssemblyBegin(baij->A, mode));
856   PetscCall(MatAssemblyEnd(baij->A, mode));
857 
858   /* determine if any processor has disassembled, if so we must
859      also disassemble ourselves, in order that we may reassemble. */
860   /*
861      if nonzero structure of submatrix B cannot change then we know that
862      no processor disassembled thus we can skip this stuff
863   */
864   if (!((Mat_SeqBAIJ *)baij->B->data)->nonew) {
865     PetscCallMPI(MPIU_Allreduce(&mat->was_assembled, &other_disassembled, 1, MPIU_BOOL, MPI_LAND, PetscObjectComm((PetscObject)mat)));
866     if (mat->was_assembled && !other_disassembled) PetscCall(MatDisAssemble_MPISBAIJ(mat));
867   }
868 
869   if (!mat->was_assembled && mode == MAT_FINAL_ASSEMBLY) { PetscCall(MatSetUpMultiply_MPISBAIJ(mat)); /* setup Mvctx and sMvctx */ }
870   PetscCall(MatAssemblyBegin(baij->B, mode));
871   PetscCall(MatAssemblyEnd(baij->B, mode));
872 
873   PetscCall(PetscFree2(baij->rowvalues, baij->rowindices));
874 
875   baij->rowvalues = NULL;
876 
877   /* if no new nonzero locations are allowed in matrix then only set the matrix state the first time through */
878   if ((!mat->was_assembled && mode == MAT_FINAL_ASSEMBLY) || !((Mat_SeqBAIJ *)baij->A->data)->nonew) {
879     PetscObjectState state = baij->A->nonzerostate + baij->B->nonzerostate;
880     PetscCallMPI(MPIU_Allreduce(&state, &mat->nonzerostate, 1, MPIU_INT64, MPI_SUM, PetscObjectComm((PetscObject)mat)));
881   }
882   PetscFunctionReturn(PETSC_SUCCESS);
883 }
884 
885 extern PetscErrorCode MatSetValues_MPIBAIJ(Mat, PetscInt, const PetscInt[], PetscInt, const PetscInt[], const PetscScalar[], InsertMode);
886 #include <petscdraw.h>
887 static PetscErrorCode MatView_MPISBAIJ_ASCIIorDraworSocket(Mat mat, PetscViewer viewer)
888 {
889   Mat_MPISBAIJ     *baij = (Mat_MPISBAIJ *)mat->data;
890   PetscInt          bs   = mat->rmap->bs;
891   PetscMPIInt       rank = baij->rank;
892   PetscBool         iascii, isdraw;
893   PetscViewer       sviewer;
894   PetscViewerFormat format;
895 
896   PetscFunctionBegin;
897   PetscCall(PetscObjectTypeCompare((PetscObject)viewer, PETSCVIEWERASCII, &iascii));
898   PetscCall(PetscObjectTypeCompare((PetscObject)viewer, PETSCVIEWERDRAW, &isdraw));
899   if (iascii) {
900     PetscCall(PetscViewerGetFormat(viewer, &format));
901     if (format == PETSC_VIEWER_ASCII_INFO_DETAIL) {
902       MatInfo info;
903       PetscCallMPI(MPI_Comm_rank(PetscObjectComm((PetscObject)mat), &rank));
904       PetscCall(MatGetInfo(mat, MAT_LOCAL, &info));
905       PetscCall(PetscViewerASCIIPushSynchronized(viewer));
906       PetscCall(PetscViewerASCIISynchronizedPrintf(viewer, "[%d] Local rows %" PetscInt_FMT " nz %" PetscInt_FMT " nz alloced %" PetscInt_FMT " bs %" PetscInt_FMT " mem %g\n", rank, mat->rmap->n, (PetscInt)info.nz_used, (PetscInt)info.nz_allocated,
907                                                    mat->rmap->bs, info.memory));
908       PetscCall(MatGetInfo(baij->A, MAT_LOCAL, &info));
909       PetscCall(PetscViewerASCIISynchronizedPrintf(viewer, "[%d] on-diagonal part: nz %" PetscInt_FMT " \n", rank, (PetscInt)info.nz_used));
910       PetscCall(MatGetInfo(baij->B, MAT_LOCAL, &info));
911       PetscCall(PetscViewerASCIISynchronizedPrintf(viewer, "[%d] off-diagonal part: nz %" PetscInt_FMT " \n", rank, (PetscInt)info.nz_used));
912       PetscCall(PetscViewerFlush(viewer));
913       PetscCall(PetscViewerASCIIPopSynchronized(viewer));
914       PetscCall(PetscViewerASCIIPrintf(viewer, "Information on VecScatter used in matrix-vector product: \n"));
915       PetscCall(VecScatterView(baij->Mvctx, viewer));
916       PetscFunctionReturn(PETSC_SUCCESS);
917     } else if (format == PETSC_VIEWER_ASCII_INFO) {
918       PetscCall(PetscViewerASCIIPrintf(viewer, "  block size is %" PetscInt_FMT "\n", bs));
919       PetscFunctionReturn(PETSC_SUCCESS);
920     } else if (format == PETSC_VIEWER_ASCII_FACTOR_INFO) {
921       PetscFunctionReturn(PETSC_SUCCESS);
922     }
923   }
924 
925   if (isdraw) {
926     PetscDraw draw;
927     PetscBool isnull;
928     PetscCall(PetscViewerDrawGetDraw(viewer, 0, &draw));
929     PetscCall(PetscDrawIsNull(draw, &isnull));
930     if (isnull) PetscFunctionReturn(PETSC_SUCCESS);
931   }
932 
933   {
934     /* assemble the entire matrix onto first processor. */
935     Mat           A;
936     Mat_SeqSBAIJ *Aloc;
937     Mat_SeqBAIJ  *Bloc;
938     PetscInt      M = mat->rmap->N, N = mat->cmap->N, *ai, *aj, col, i, j, k, *rvals, mbs = baij->mbs;
939     MatScalar    *a;
940     const char   *matname;
941 
942     /* Should this be the same type as mat? */
943     PetscCall(MatCreate(PetscObjectComm((PetscObject)mat), &A));
944     if (rank == 0) {
945       PetscCall(MatSetSizes(A, M, N, M, N));
946     } else {
947       PetscCall(MatSetSizes(A, 0, 0, M, N));
948     }
949     PetscCall(MatSetType(A, MATMPISBAIJ));
950     PetscCall(MatMPISBAIJSetPreallocation(A, mat->rmap->bs, 0, NULL, 0, NULL));
951     PetscCall(MatSetOption(A, MAT_NEW_NONZERO_LOCATION_ERR, PETSC_FALSE));
952 
953     /* copy over the A part */
954     Aloc = (Mat_SeqSBAIJ *)baij->A->data;
955     ai   = Aloc->i;
956     aj   = Aloc->j;
957     a    = Aloc->a;
958     PetscCall(PetscMalloc1(bs, &rvals));
959 
960     for (i = 0; i < mbs; i++) {
961       rvals[0] = bs * (baij->rstartbs + i);
962       for (j = 1; j < bs; j++) rvals[j] = rvals[j - 1] + 1;
963       for (j = ai[i]; j < ai[i + 1]; j++) {
964         col = (baij->cstartbs + aj[j]) * bs;
965         for (k = 0; k < bs; k++) {
966           PetscCall(MatSetValues_MPISBAIJ(A, bs, rvals, 1, &col, a, INSERT_VALUES));
967           col++;
968           a += bs;
969         }
970       }
971     }
972     /* copy over the B part */
973     Bloc = (Mat_SeqBAIJ *)baij->B->data;
974     ai   = Bloc->i;
975     aj   = Bloc->j;
976     a    = Bloc->a;
977     for (i = 0; i < mbs; i++) {
978       rvals[0] = bs * (baij->rstartbs + i);
979       for (j = 1; j < bs; j++) rvals[j] = rvals[j - 1] + 1;
980       for (j = ai[i]; j < ai[i + 1]; j++) {
981         col = baij->garray[aj[j]] * bs;
982         for (k = 0; k < bs; k++) {
983           PetscCall(MatSetValues_MPIBAIJ(A, bs, rvals, 1, &col, a, INSERT_VALUES));
984           col++;
985           a += bs;
986         }
987       }
988     }
989     PetscCall(PetscFree(rvals));
990     PetscCall(MatAssemblyBegin(A, MAT_FINAL_ASSEMBLY));
991     PetscCall(MatAssemblyEnd(A, MAT_FINAL_ASSEMBLY));
992     /*
993        Everyone has to call to draw the matrix since the graphics waits are
994        synchronized across all processors that share the PetscDraw object
995     */
996     PetscCall(PetscViewerGetSubViewer(viewer, PETSC_COMM_SELF, &sviewer));
997     if (((PetscObject)mat)->name) PetscCall(PetscObjectGetName((PetscObject)mat, &matname));
998     if (rank == 0) {
999       if (((PetscObject)mat)->name) PetscCall(PetscObjectSetName((PetscObject)((Mat_MPISBAIJ *)A->data)->A, matname));
1000       PetscCall(MatView_SeqSBAIJ(((Mat_MPISBAIJ *)A->data)->A, sviewer));
1001     }
1002     PetscCall(PetscViewerRestoreSubViewer(viewer, PETSC_COMM_SELF, &sviewer));
1003     PetscCall(MatDestroy(&A));
1004   }
1005   PetscFunctionReturn(PETSC_SUCCESS);
1006 }
1007 
1008 /* Used for both MPIBAIJ and MPISBAIJ matrices */
1009 #define MatView_MPISBAIJ_Binary MatView_MPIBAIJ_Binary
1010 
1011 static PetscErrorCode MatView_MPISBAIJ(Mat mat, PetscViewer viewer)
1012 {
1013   PetscBool iascii, isdraw, issocket, isbinary;
1014 
1015   PetscFunctionBegin;
1016   PetscCall(PetscObjectTypeCompare((PetscObject)viewer, PETSCVIEWERASCII, &iascii));
1017   PetscCall(PetscObjectTypeCompare((PetscObject)viewer, PETSCVIEWERDRAW, &isdraw));
1018   PetscCall(PetscObjectTypeCompare((PetscObject)viewer, PETSCVIEWERSOCKET, &issocket));
1019   PetscCall(PetscObjectTypeCompare((PetscObject)viewer, PETSCVIEWERBINARY, &isbinary));
1020   if (iascii || isdraw || issocket) {
1021     PetscCall(MatView_MPISBAIJ_ASCIIorDraworSocket(mat, viewer));
1022   } else if (isbinary) PetscCall(MatView_MPISBAIJ_Binary(mat, viewer));
1023   PetscFunctionReturn(PETSC_SUCCESS);
1024 }
1025 
1026 #if defined(PETSC_USE_COMPLEX)
1027 static PetscErrorCode MatMult_MPISBAIJ_Hermitian(Mat A, Vec xx, Vec yy)
1028 {
1029   Mat_MPISBAIJ      *a   = (Mat_MPISBAIJ *)A->data;
1030   PetscInt           mbs = a->mbs, bs = A->rmap->bs;
1031   PetscScalar       *from;
1032   const PetscScalar *x;
1033 
1034   PetscFunctionBegin;
1035   /* diagonal part */
1036   PetscCall((*a->A->ops->mult)(a->A, xx, a->slvec1a));
1037   /* since a->slvec1b shares memory (dangerously) with a->slec1 changes to a->slec1 will affect it */
1038   PetscCall(PetscObjectStateIncrease((PetscObject)a->slvec1b));
1039   PetscCall(VecZeroEntries(a->slvec1b));
1040 
1041   /* subdiagonal part */
1042   PetscCheck(a->B->ops->multhermitiantranspose, PetscObjectComm((PetscObject)a->B), PETSC_ERR_SUP, "Not for type %s", ((PetscObject)a->B)->type_name);
1043   PetscCall((*a->B->ops->multhermitiantranspose)(a->B, xx, a->slvec0b));
1044 
1045   /* copy x into the vec slvec0 */
1046   PetscCall(VecGetArray(a->slvec0, &from));
1047   PetscCall(VecGetArrayRead(xx, &x));
1048 
1049   PetscCall(PetscArraycpy(from, x, bs * mbs));
1050   PetscCall(VecRestoreArray(a->slvec0, &from));
1051   PetscCall(VecRestoreArrayRead(xx, &x));
1052 
1053   PetscCall(VecScatterBegin(a->sMvctx, a->slvec0, a->slvec1, ADD_VALUES, SCATTER_FORWARD));
1054   PetscCall(VecScatterEnd(a->sMvctx, a->slvec0, a->slvec1, ADD_VALUES, SCATTER_FORWARD));
1055   /* supperdiagonal part */
1056   PetscCall((*a->B->ops->multadd)(a->B, a->slvec1b, a->slvec1a, yy));
1057   PetscFunctionReturn(PETSC_SUCCESS);
1058 }
1059 #endif
1060 
1061 static PetscErrorCode MatMult_MPISBAIJ(Mat A, Vec xx, Vec yy)
1062 {
1063   Mat_MPISBAIJ      *a   = (Mat_MPISBAIJ *)A->data;
1064   PetscInt           mbs = a->mbs, bs = A->rmap->bs;
1065   PetscScalar       *from;
1066   const PetscScalar *x;
1067 
1068   PetscFunctionBegin;
1069   /* diagonal part */
1070   PetscCall((*a->A->ops->mult)(a->A, xx, a->slvec1a));
1071   /* since a->slvec1b shares memory (dangerously) with a->slec1 changes to a->slec1 will affect it */
1072   PetscCall(PetscObjectStateIncrease((PetscObject)a->slvec1b));
1073   PetscCall(VecZeroEntries(a->slvec1b));
1074 
1075   /* subdiagonal part */
1076   PetscCall((*a->B->ops->multtranspose)(a->B, xx, a->slvec0b));
1077 
1078   /* copy x into the vec slvec0 */
1079   PetscCall(VecGetArray(a->slvec0, &from));
1080   PetscCall(VecGetArrayRead(xx, &x));
1081 
1082   PetscCall(PetscArraycpy(from, x, bs * mbs));
1083   PetscCall(VecRestoreArray(a->slvec0, &from));
1084   PetscCall(VecRestoreArrayRead(xx, &x));
1085 
1086   PetscCall(VecScatterBegin(a->sMvctx, a->slvec0, a->slvec1, ADD_VALUES, SCATTER_FORWARD));
1087   PetscCall(VecScatterEnd(a->sMvctx, a->slvec0, a->slvec1, ADD_VALUES, SCATTER_FORWARD));
1088   /* supperdiagonal part */
1089   PetscCall((*a->B->ops->multadd)(a->B, a->slvec1b, a->slvec1a, yy));
1090   PetscFunctionReturn(PETSC_SUCCESS);
1091 }
1092 
1093 #if PetscDefined(USE_COMPLEX)
1094 static PetscErrorCode MatMultAdd_MPISBAIJ_Hermitian(Mat A, Vec xx, Vec yy, Vec zz)
1095 {
1096   Mat_MPISBAIJ      *a   = (Mat_MPISBAIJ *)A->data;
1097   PetscInt           mbs = a->mbs, bs = A->rmap->bs;
1098   PetscScalar       *from;
1099   const PetscScalar *x;
1100 
1101   PetscFunctionBegin;
1102   /* diagonal part */
1103   PetscCall((*a->A->ops->multadd)(a->A, xx, yy, a->slvec1a));
1104   PetscCall(PetscObjectStateIncrease((PetscObject)a->slvec1b));
1105   PetscCall(VecZeroEntries(a->slvec1b));
1106 
1107   /* subdiagonal part */
1108   PetscCheck(a->B->ops->multhermitiantranspose, PetscObjectComm((PetscObject)a->B), PETSC_ERR_SUP, "Not for type %s", ((PetscObject)a->B)->type_name);
1109   PetscCall((*a->B->ops->multhermitiantranspose)(a->B, xx, a->slvec0b));
1110 
1111   /* copy x into the vec slvec0 */
1112   PetscCall(VecGetArray(a->slvec0, &from));
1113   PetscCall(VecGetArrayRead(xx, &x));
1114   PetscCall(PetscArraycpy(from, x, bs * mbs));
1115   PetscCall(VecRestoreArray(a->slvec0, &from));
1116 
1117   PetscCall(VecScatterBegin(a->sMvctx, a->slvec0, a->slvec1, ADD_VALUES, SCATTER_FORWARD));
1118   PetscCall(VecRestoreArrayRead(xx, &x));
1119   PetscCall(VecScatterEnd(a->sMvctx, a->slvec0, a->slvec1, ADD_VALUES, SCATTER_FORWARD));
1120 
1121   /* supperdiagonal part */
1122   PetscCall((*a->B->ops->multadd)(a->B, a->slvec1b, a->slvec1a, zz));
1123   PetscFunctionReturn(PETSC_SUCCESS);
1124 }
1125 #endif
1126 
1127 static PetscErrorCode MatMultAdd_MPISBAIJ(Mat A, Vec xx, Vec yy, Vec zz)
1128 {
1129   Mat_MPISBAIJ      *a   = (Mat_MPISBAIJ *)A->data;
1130   PetscInt           mbs = a->mbs, bs = A->rmap->bs;
1131   PetscScalar       *from;
1132   const PetscScalar *x;
1133 
1134   PetscFunctionBegin;
1135   /* diagonal part */
1136   PetscCall((*a->A->ops->multadd)(a->A, xx, yy, a->slvec1a));
1137   PetscCall(PetscObjectStateIncrease((PetscObject)a->slvec1b));
1138   PetscCall(VecZeroEntries(a->slvec1b));
1139 
1140   /* subdiagonal part */
1141   PetscCall((*a->B->ops->multtranspose)(a->B, xx, a->slvec0b));
1142 
1143   /* copy x into the vec slvec0 */
1144   PetscCall(VecGetArray(a->slvec0, &from));
1145   PetscCall(VecGetArrayRead(xx, &x));
1146   PetscCall(PetscArraycpy(from, x, bs * mbs));
1147   PetscCall(VecRestoreArray(a->slvec0, &from));
1148 
1149   PetscCall(VecScatterBegin(a->sMvctx, a->slvec0, a->slvec1, ADD_VALUES, SCATTER_FORWARD));
1150   PetscCall(VecRestoreArrayRead(xx, &x));
1151   PetscCall(VecScatterEnd(a->sMvctx, a->slvec0, a->slvec1, ADD_VALUES, SCATTER_FORWARD));
1152 
1153   /* supperdiagonal part */
1154   PetscCall((*a->B->ops->multadd)(a->B, a->slvec1b, a->slvec1a, zz));
1155   PetscFunctionReturn(PETSC_SUCCESS);
1156 }
1157 
1158 /*
1159   This only works correctly for square matrices where the subblock A->A is the
1160    diagonal block
1161 */
1162 static PetscErrorCode MatGetDiagonal_MPISBAIJ(Mat A, Vec v)
1163 {
1164   Mat_MPISBAIJ *a = (Mat_MPISBAIJ *)A->data;
1165 
1166   PetscFunctionBegin;
1167   /* PetscCheck(a->rmap->N == a->cmap->N,PETSC_COMM_SELF,PETSC_ERR_SUP,"Supports only square matrix where A->A is diag block"); */
1168   PetscCall(MatGetDiagonal(a->A, v));
1169   PetscFunctionReturn(PETSC_SUCCESS);
1170 }
1171 
1172 static PetscErrorCode MatScale_MPISBAIJ(Mat A, PetscScalar aa)
1173 {
1174   Mat_MPISBAIJ *a = (Mat_MPISBAIJ *)A->data;
1175 
1176   PetscFunctionBegin;
1177   PetscCall(MatScale(a->A, aa));
1178   PetscCall(MatScale(a->B, aa));
1179   PetscFunctionReturn(PETSC_SUCCESS);
1180 }
1181 
1182 static PetscErrorCode MatGetRow_MPISBAIJ(Mat matin, PetscInt row, PetscInt *nz, PetscInt **idx, PetscScalar **v)
1183 {
1184   Mat_MPISBAIJ *mat = (Mat_MPISBAIJ *)matin->data;
1185   PetscScalar  *vworkA, *vworkB, **pvA, **pvB, *v_p;
1186   PetscInt      bs = matin->rmap->bs, bs2 = mat->bs2, i, *cworkA, *cworkB, **pcA, **pcB;
1187   PetscInt      nztot, nzA, nzB, lrow, brstart = matin->rmap->rstart, brend = matin->rmap->rend;
1188   PetscInt     *cmap, *idx_p, cstart = mat->rstartbs;
1189 
1190   PetscFunctionBegin;
1191   PetscCheck(!mat->getrowactive, PETSC_COMM_SELF, PETSC_ERR_ARG_WRONGSTATE, "Already active");
1192   mat->getrowactive = PETSC_TRUE;
1193 
1194   if (!mat->rowvalues && (idx || v)) {
1195     /*
1196         allocate enough space to hold information from the longest row.
1197     */
1198     Mat_SeqSBAIJ *Aa  = (Mat_SeqSBAIJ *)mat->A->data;
1199     Mat_SeqBAIJ  *Ba  = (Mat_SeqBAIJ *)mat->B->data;
1200     PetscInt      max = 1, mbs = mat->mbs, tmp;
1201     for (i = 0; i < mbs; i++) {
1202       tmp = Aa->i[i + 1] - Aa->i[i] + Ba->i[i + 1] - Ba->i[i]; /* row length */
1203       if (max < tmp) max = tmp;
1204     }
1205     PetscCall(PetscMalloc2(max * bs2, &mat->rowvalues, max * bs2, &mat->rowindices));
1206   }
1207 
1208   PetscCheck(row >= brstart && row < brend, PETSC_COMM_SELF, PETSC_ERR_SUP, "Only local rows");
1209   lrow = row - brstart; /* local row index */
1210 
1211   pvA = &vworkA;
1212   pcA = &cworkA;
1213   pvB = &vworkB;
1214   pcB = &cworkB;
1215   if (!v) {
1216     pvA = NULL;
1217     pvB = NULL;
1218   }
1219   if (!idx) {
1220     pcA = NULL;
1221     if (!v) pcB = NULL;
1222   }
1223   PetscCall((*mat->A->ops->getrow)(mat->A, lrow, &nzA, pcA, pvA));
1224   PetscCall((*mat->B->ops->getrow)(mat->B, lrow, &nzB, pcB, pvB));
1225   nztot = nzA + nzB;
1226 
1227   cmap = mat->garray;
1228   if (v || idx) {
1229     if (nztot) {
1230       /* Sort by increasing column numbers, assuming A and B already sorted */
1231       PetscInt imark = -1;
1232       if (v) {
1233         *v = v_p = mat->rowvalues;
1234         for (i = 0; i < nzB; i++) {
1235           if (cmap[cworkB[i] / bs] < cstart) v_p[i] = vworkB[i];
1236           else break;
1237         }
1238         imark = i;
1239         for (i = 0; i < nzA; i++) v_p[imark + i] = vworkA[i];
1240         for (i = imark; i < nzB; i++) v_p[nzA + i] = vworkB[i];
1241       }
1242       if (idx) {
1243         *idx = idx_p = mat->rowindices;
1244         if (imark > -1) {
1245           for (i = 0; i < imark; i++) idx_p[i] = cmap[cworkB[i] / bs] * bs + cworkB[i] % bs;
1246         } else {
1247           for (i = 0; i < nzB; i++) {
1248             if (cmap[cworkB[i] / bs] < cstart) idx_p[i] = cmap[cworkB[i] / bs] * bs + cworkB[i] % bs;
1249             else break;
1250           }
1251           imark = i;
1252         }
1253         for (i = 0; i < nzA; i++) idx_p[imark + i] = cstart * bs + cworkA[i];
1254         for (i = imark; i < nzB; i++) idx_p[nzA + i] = cmap[cworkB[i] / bs] * bs + cworkB[i] % bs;
1255       }
1256     } else {
1257       if (idx) *idx = NULL;
1258       if (v) *v = NULL;
1259     }
1260   }
1261   *nz = nztot;
1262   PetscCall((*mat->A->ops->restorerow)(mat->A, lrow, &nzA, pcA, pvA));
1263   PetscCall((*mat->B->ops->restorerow)(mat->B, lrow, &nzB, pcB, pvB));
1264   PetscFunctionReturn(PETSC_SUCCESS);
1265 }
1266 
1267 static PetscErrorCode MatRestoreRow_MPISBAIJ(Mat mat, PetscInt row, PetscInt *nz, PetscInt **idx, PetscScalar **v)
1268 {
1269   Mat_MPISBAIJ *baij = (Mat_MPISBAIJ *)mat->data;
1270 
1271   PetscFunctionBegin;
1272   PetscCheck(baij->getrowactive, PETSC_COMM_SELF, PETSC_ERR_ARG_WRONGSTATE, "MatGetRow() must be called first");
1273   baij->getrowactive = PETSC_FALSE;
1274   PetscFunctionReturn(PETSC_SUCCESS);
1275 }
1276 
1277 static PetscErrorCode MatGetRowUpperTriangular_MPISBAIJ(Mat A)
1278 {
1279   Mat_MPISBAIJ *a  = (Mat_MPISBAIJ *)A->data;
1280   Mat_SeqSBAIJ *aA = (Mat_SeqSBAIJ *)a->A->data;
1281 
1282   PetscFunctionBegin;
1283   aA->getrow_utriangular = PETSC_TRUE;
1284   PetscFunctionReturn(PETSC_SUCCESS);
1285 }
1286 static PetscErrorCode MatRestoreRowUpperTriangular_MPISBAIJ(Mat A)
1287 {
1288   Mat_MPISBAIJ *a  = (Mat_MPISBAIJ *)A->data;
1289   Mat_SeqSBAIJ *aA = (Mat_SeqSBAIJ *)a->A->data;
1290 
1291   PetscFunctionBegin;
1292   aA->getrow_utriangular = PETSC_FALSE;
1293   PetscFunctionReturn(PETSC_SUCCESS);
1294 }
1295 
1296 static PetscErrorCode MatConjugate_MPISBAIJ(Mat mat)
1297 {
1298   PetscFunctionBegin;
1299   if (PetscDefined(USE_COMPLEX)) {
1300     Mat_MPISBAIJ *a = (Mat_MPISBAIJ *)mat->data;
1301 
1302     PetscCall(MatConjugate(a->A));
1303     PetscCall(MatConjugate(a->B));
1304   }
1305   PetscFunctionReturn(PETSC_SUCCESS);
1306 }
1307 
1308 static PetscErrorCode MatRealPart_MPISBAIJ(Mat A)
1309 {
1310   Mat_MPISBAIJ *a = (Mat_MPISBAIJ *)A->data;
1311 
1312   PetscFunctionBegin;
1313   PetscCall(MatRealPart(a->A));
1314   PetscCall(MatRealPart(a->B));
1315   PetscFunctionReturn(PETSC_SUCCESS);
1316 }
1317 
1318 static PetscErrorCode MatImaginaryPart_MPISBAIJ(Mat A)
1319 {
1320   Mat_MPISBAIJ *a = (Mat_MPISBAIJ *)A->data;
1321 
1322   PetscFunctionBegin;
1323   PetscCall(MatImaginaryPart(a->A));
1324   PetscCall(MatImaginaryPart(a->B));
1325   PetscFunctionReturn(PETSC_SUCCESS);
1326 }
1327 
1328 /* Check if isrow is a subset of iscol_local, called by MatCreateSubMatrix_MPISBAIJ()
1329    Input: isrow       - distributed(parallel),
1330           iscol_local - locally owned (seq)
1331 */
1332 static PetscErrorCode ISEqual_private(IS isrow, IS iscol_local, PetscBool *flg)
1333 {
1334   PetscInt        sz1, sz2, *a1, *a2, i, j, k, nmatch;
1335   const PetscInt *ptr1, *ptr2;
1336 
1337   PetscFunctionBegin;
1338   *flg = PETSC_FALSE;
1339   PetscCall(ISGetLocalSize(isrow, &sz1));
1340   PetscCall(ISGetLocalSize(iscol_local, &sz2));
1341   if (sz1 > sz2) PetscFunctionReturn(PETSC_SUCCESS);
1342 
1343   PetscCall(ISGetIndices(isrow, &ptr1));
1344   PetscCall(ISGetIndices(iscol_local, &ptr2));
1345 
1346   PetscCall(PetscMalloc1(sz1, &a1));
1347   PetscCall(PetscMalloc1(sz2, &a2));
1348   PetscCall(PetscArraycpy(a1, ptr1, sz1));
1349   PetscCall(PetscArraycpy(a2, ptr2, sz2));
1350   PetscCall(PetscSortInt(sz1, a1));
1351   PetscCall(PetscSortInt(sz2, a2));
1352 
1353   nmatch = 0;
1354   k      = 0;
1355   for (i = 0; i < sz1; i++) {
1356     for (j = k; j < sz2; j++) {
1357       if (a1[i] == a2[j]) {
1358         k = j;
1359         nmatch++;
1360         break;
1361       }
1362     }
1363   }
1364   PetscCall(ISRestoreIndices(isrow, &ptr1));
1365   PetscCall(ISRestoreIndices(iscol_local, &ptr2));
1366   PetscCall(PetscFree(a1));
1367   PetscCall(PetscFree(a2));
1368   if (nmatch < sz1) {
1369     *flg = PETSC_FALSE;
1370   } else {
1371     *flg = PETSC_TRUE;
1372   }
1373   PetscFunctionReturn(PETSC_SUCCESS);
1374 }
1375 
1376 static PetscErrorCode MatCreateSubMatrix_MPISBAIJ(Mat mat, IS isrow, IS iscol, MatReuse call, Mat *newmat)
1377 {
1378   Mat       C[2];
1379   IS        iscol_local, isrow_local;
1380   PetscInt  csize, csize_local, rsize;
1381   PetscBool isequal, issorted, isidentity = PETSC_FALSE;
1382 
1383   PetscFunctionBegin;
1384   PetscCall(ISGetLocalSize(iscol, &csize));
1385   PetscCall(ISGetLocalSize(isrow, &rsize));
1386   if (call == MAT_REUSE_MATRIX) {
1387     PetscCall(PetscObjectQuery((PetscObject)*newmat, "ISAllGather", (PetscObject *)&iscol_local));
1388     PetscCheck(iscol_local, PETSC_COMM_SELF, PETSC_ERR_ARG_WRONGSTATE, "Submatrix passed in was not used before, cannot reuse");
1389   } else {
1390     PetscCall(ISAllGather(iscol, &iscol_local));
1391     PetscCall(ISSorted(iscol_local, &issorted));
1392     PetscCheck(issorted, PETSC_COMM_SELF, PETSC_ERR_ARG_INCOMP, "For symmetric format, iscol must be sorted");
1393   }
1394   PetscCall(ISEqual_private(isrow, iscol_local, &isequal));
1395   if (!isequal) {
1396     PetscCall(ISGetLocalSize(iscol_local, &csize_local));
1397     isidentity = (PetscBool)(mat->cmap->N == csize_local);
1398     if (!isidentity) {
1399       if (call == MAT_REUSE_MATRIX) {
1400         PetscCall(PetscObjectQuery((PetscObject)*newmat, "ISAllGather_other", (PetscObject *)&isrow_local));
1401         PetscCheck(isrow_local, PETSC_COMM_SELF, PETSC_ERR_ARG_WRONGSTATE, "Submatrix passed in was not used before, cannot reuse");
1402       } else {
1403         PetscCall(ISAllGather(isrow, &isrow_local));
1404         PetscCall(ISSorted(isrow_local, &issorted));
1405         PetscCheck(issorted, PETSC_COMM_SELF, PETSC_ERR_ARG_INCOMP, "For symmetric format, isrow must be sorted");
1406       }
1407     }
1408   }
1409   /* now call MatCreateSubMatrix_MPIBAIJ() */
1410   PetscCall(MatCreateSubMatrix_MPIBAIJ_Private(mat, isrow, iscol_local, csize, isequal || isidentity ? call : MAT_INITIAL_MATRIX, isequal || isidentity ? newmat : C, (PetscBool)(isequal || isidentity)));
1411   if (!isequal && !isidentity) {
1412     if (call == MAT_INITIAL_MATRIX) {
1413       IS       intersect;
1414       PetscInt ni;
1415 
1416       PetscCall(ISIntersect(isrow_local, iscol_local, &intersect));
1417       PetscCall(ISGetLocalSize(intersect, &ni));
1418       PetscCall(ISDestroy(&intersect));
1419       PetscCheck(ni == 0, PETSC_COMM_SELF, PETSC_ERR_ARG_WRONG, "Cannot create such a submatrix: for symmetric format, when requesting an off-diagonal submatrix, isrow and iscol should have an empty intersection (number of common indices is %" PetscInt_FMT ")", ni);
1420     }
1421     PetscCall(MatCreateSubMatrix_MPIBAIJ_Private(mat, iscol, isrow_local, rsize, MAT_INITIAL_MATRIX, C + 1, PETSC_FALSE));
1422     PetscCall(MatTranspose(C[1], MAT_INPLACE_MATRIX, C + 1));
1423     PetscCall(MatAXPY(C[0], 1.0, C[1], DIFFERENT_NONZERO_PATTERN));
1424     if (call == MAT_REUSE_MATRIX) PetscCall(MatCopy(C[0], *newmat, SAME_NONZERO_PATTERN));
1425     else if (mat->rmap->bs == 1) PetscCall(MatConvert(C[0], MATAIJ, MAT_INITIAL_MATRIX, newmat));
1426     else PetscCall(MatCopy(C[0], *newmat, SAME_NONZERO_PATTERN));
1427     PetscCall(MatDestroy(C));
1428     PetscCall(MatDestroy(C + 1));
1429   }
1430   if (call == MAT_INITIAL_MATRIX) {
1431     if (!isequal && !isidentity) {
1432       PetscCall(PetscObjectCompose((PetscObject)*newmat, "ISAllGather_other", (PetscObject)isrow_local));
1433       PetscCall(ISDestroy(&isrow_local));
1434     }
1435     PetscCall(PetscObjectCompose((PetscObject)*newmat, "ISAllGather", (PetscObject)iscol_local));
1436     PetscCall(ISDestroy(&iscol_local));
1437   }
1438   PetscFunctionReturn(PETSC_SUCCESS);
1439 }
1440 
1441 static PetscErrorCode MatZeroEntries_MPISBAIJ(Mat A)
1442 {
1443   Mat_MPISBAIJ *l = (Mat_MPISBAIJ *)A->data;
1444 
1445   PetscFunctionBegin;
1446   PetscCall(MatZeroEntries(l->A));
1447   PetscCall(MatZeroEntries(l->B));
1448   PetscFunctionReturn(PETSC_SUCCESS);
1449 }
1450 
1451 static PetscErrorCode MatGetInfo_MPISBAIJ(Mat matin, MatInfoType flag, MatInfo *info)
1452 {
1453   Mat_MPISBAIJ  *a = (Mat_MPISBAIJ *)matin->data;
1454   Mat            A = a->A, B = a->B;
1455   PetscLogDouble isend[5], irecv[5];
1456 
1457   PetscFunctionBegin;
1458   info->block_size = (PetscReal)matin->rmap->bs;
1459 
1460   PetscCall(MatGetInfo(A, MAT_LOCAL, info));
1461 
1462   isend[0] = info->nz_used;
1463   isend[1] = info->nz_allocated;
1464   isend[2] = info->nz_unneeded;
1465   isend[3] = info->memory;
1466   isend[4] = info->mallocs;
1467 
1468   PetscCall(MatGetInfo(B, MAT_LOCAL, info));
1469 
1470   isend[0] += info->nz_used;
1471   isend[1] += info->nz_allocated;
1472   isend[2] += info->nz_unneeded;
1473   isend[3] += info->memory;
1474   isend[4] += info->mallocs;
1475   if (flag == MAT_LOCAL) {
1476     info->nz_used      = isend[0];
1477     info->nz_allocated = isend[1];
1478     info->nz_unneeded  = isend[2];
1479     info->memory       = isend[3];
1480     info->mallocs      = isend[4];
1481   } else if (flag == MAT_GLOBAL_MAX) {
1482     PetscCallMPI(MPIU_Allreduce(isend, irecv, 5, MPIU_PETSCLOGDOUBLE, MPI_MAX, PetscObjectComm((PetscObject)matin)));
1483 
1484     info->nz_used      = irecv[0];
1485     info->nz_allocated = irecv[1];
1486     info->nz_unneeded  = irecv[2];
1487     info->memory       = irecv[3];
1488     info->mallocs      = irecv[4];
1489   } else if (flag == MAT_GLOBAL_SUM) {
1490     PetscCallMPI(MPIU_Allreduce(isend, irecv, 5, MPIU_PETSCLOGDOUBLE, MPI_SUM, PetscObjectComm((PetscObject)matin)));
1491 
1492     info->nz_used      = irecv[0];
1493     info->nz_allocated = irecv[1];
1494     info->nz_unneeded  = irecv[2];
1495     info->memory       = irecv[3];
1496     info->mallocs      = irecv[4];
1497   } else SETERRQ(PETSC_COMM_SELF, PETSC_ERR_ARG_WRONG, "Unknown MatInfoType argument %d", (int)flag);
1498   info->fill_ratio_given  = 0; /* no parallel LU/ILU/Cholesky */
1499   info->fill_ratio_needed = 0;
1500   info->factor_mallocs    = 0;
1501   PetscFunctionReturn(PETSC_SUCCESS);
1502 }
1503 
1504 static PetscErrorCode MatSetOption_MPISBAIJ(Mat A, MatOption op, PetscBool flg)
1505 {
1506   Mat_MPISBAIJ *a  = (Mat_MPISBAIJ *)A->data;
1507   Mat_SeqSBAIJ *aA = (Mat_SeqSBAIJ *)a->A->data;
1508 
1509   PetscFunctionBegin;
1510   switch (op) {
1511   case MAT_NEW_NONZERO_LOCATIONS:
1512   case MAT_NEW_NONZERO_ALLOCATION_ERR:
1513   case MAT_UNUSED_NONZERO_LOCATION_ERR:
1514   case MAT_KEEP_NONZERO_PATTERN:
1515   case MAT_SUBMAT_SINGLEIS:
1516   case MAT_NEW_NONZERO_LOCATION_ERR:
1517     MatCheckPreallocated(A, 1);
1518     PetscCall(MatSetOption(a->A, op, flg));
1519     PetscCall(MatSetOption(a->B, op, flg));
1520     break;
1521   case MAT_ROW_ORIENTED:
1522     MatCheckPreallocated(A, 1);
1523     a->roworiented = flg;
1524 
1525     PetscCall(MatSetOption(a->A, op, flg));
1526     PetscCall(MatSetOption(a->B, op, flg));
1527     break;
1528   case MAT_FORCE_DIAGONAL_ENTRIES:
1529   case MAT_SORTED_FULL:
1530     PetscCall(PetscInfo(A, "Option %s ignored\n", MatOptions[op]));
1531     break;
1532   case MAT_IGNORE_OFF_PROC_ENTRIES:
1533     a->donotstash = flg;
1534     break;
1535   case MAT_USE_HASH_TABLE:
1536     a->ht_flag = flg;
1537     break;
1538   case MAT_HERMITIAN:
1539     MatCheckPreallocated(A, 1);
1540     PetscCall(MatSetOption(a->A, op, flg));
1541 #if defined(PETSC_USE_COMPLEX)
1542     if (flg) { /* need different mat-vec ops */
1543       A->ops->mult             = MatMult_MPISBAIJ_Hermitian;
1544       A->ops->multadd          = MatMultAdd_MPISBAIJ_Hermitian;
1545       A->ops->multtranspose    = NULL;
1546       A->ops->multtransposeadd = NULL;
1547       A->symmetric             = PETSC_BOOL3_FALSE;
1548     }
1549 #endif
1550     break;
1551   case MAT_SPD:
1552   case MAT_SYMMETRIC:
1553     MatCheckPreallocated(A, 1);
1554     PetscCall(MatSetOption(a->A, op, flg));
1555 #if defined(PETSC_USE_COMPLEX)
1556     if (flg) { /* restore to use default mat-vec ops */
1557       A->ops->mult             = MatMult_MPISBAIJ;
1558       A->ops->multadd          = MatMultAdd_MPISBAIJ;
1559       A->ops->multtranspose    = MatMult_MPISBAIJ;
1560       A->ops->multtransposeadd = MatMultAdd_MPISBAIJ;
1561     }
1562 #endif
1563     break;
1564   case MAT_STRUCTURALLY_SYMMETRIC:
1565     MatCheckPreallocated(A, 1);
1566     PetscCall(MatSetOption(a->A, op, flg));
1567     break;
1568   case MAT_SYMMETRY_ETERNAL:
1569   case MAT_STRUCTURAL_SYMMETRY_ETERNAL:
1570     PetscCheck(flg, PETSC_COMM_SELF, PETSC_ERR_SUP, "Matrix must be symmetric");
1571     PetscCall(PetscInfo(A, "Option %s ignored\n", MatOptions[op]));
1572     break;
1573   case MAT_SPD_ETERNAL:
1574     break;
1575   case MAT_IGNORE_LOWER_TRIANGULAR:
1576     aA->ignore_ltriangular = flg;
1577     break;
1578   case MAT_ERROR_LOWER_TRIANGULAR:
1579     aA->ignore_ltriangular = flg;
1580     break;
1581   case MAT_GETROW_UPPERTRIANGULAR:
1582     aA->getrow_utriangular = flg;
1583     break;
1584   default:
1585     SETERRQ(PETSC_COMM_SELF, PETSC_ERR_SUP, "unknown option %d", op);
1586   }
1587   PetscFunctionReturn(PETSC_SUCCESS);
1588 }
1589 
1590 static PetscErrorCode MatTranspose_MPISBAIJ(Mat A, MatReuse reuse, Mat *B)
1591 {
1592   PetscFunctionBegin;
1593   if (reuse == MAT_REUSE_MATRIX) PetscCall(MatTransposeCheckNonzeroState_Private(A, *B));
1594   if (reuse == MAT_INITIAL_MATRIX) {
1595     PetscCall(MatDuplicate(A, MAT_COPY_VALUES, B));
1596   } else if (reuse == MAT_REUSE_MATRIX) {
1597     PetscCall(MatCopy(A, *B, SAME_NONZERO_PATTERN));
1598   }
1599   PetscFunctionReturn(PETSC_SUCCESS);
1600 }
1601 
1602 static PetscErrorCode MatDiagonalScale_MPISBAIJ(Mat mat, Vec ll, Vec rr)
1603 {
1604   Mat_MPISBAIJ *baij = (Mat_MPISBAIJ *)mat->data;
1605   Mat           a = baij->A, b = baij->B;
1606   PetscInt      nv, m, n;
1607   PetscBool     flg;
1608 
1609   PetscFunctionBegin;
1610   if (ll != rr) {
1611     PetscCall(VecEqual(ll, rr, &flg));
1612     PetscCheck(flg, PETSC_COMM_SELF, PETSC_ERR_ARG_OUTOFRANGE, "For symmetric format, left and right scaling vectors must be same");
1613   }
1614   if (!ll) PetscFunctionReturn(PETSC_SUCCESS);
1615 
1616   PetscCall(MatGetLocalSize(mat, &m, &n));
1617   PetscCheck(m == n, PETSC_COMM_SELF, PETSC_ERR_ARG_SIZ, "For symmetric format, local size %" PetscInt_FMT " %" PetscInt_FMT " must be same", m, n);
1618 
1619   PetscCall(VecGetLocalSize(rr, &nv));
1620   PetscCheck(nv == n, PETSC_COMM_SELF, PETSC_ERR_ARG_SIZ, "Left and right vector non-conforming local size");
1621 
1622   PetscCall(VecScatterBegin(baij->Mvctx, rr, baij->lvec, INSERT_VALUES, SCATTER_FORWARD));
1623 
1624   /* left diagonalscale the off-diagonal part */
1625   PetscUseTypeMethod(b, diagonalscale, ll, NULL);
1626 
1627   /* scale the diagonal part */
1628   PetscUseTypeMethod(a, diagonalscale, ll, rr);
1629 
1630   /* right diagonalscale the off-diagonal part */
1631   PetscCall(VecScatterEnd(baij->Mvctx, rr, baij->lvec, INSERT_VALUES, SCATTER_FORWARD));
1632   PetscUseTypeMethod(b, diagonalscale, NULL, baij->lvec);
1633   PetscFunctionReturn(PETSC_SUCCESS);
1634 }
1635 
1636 static PetscErrorCode MatSetUnfactored_MPISBAIJ(Mat A)
1637 {
1638   Mat_MPISBAIJ *a = (Mat_MPISBAIJ *)A->data;
1639 
1640   PetscFunctionBegin;
1641   PetscCall(MatSetUnfactored(a->A));
1642   PetscFunctionReturn(PETSC_SUCCESS);
1643 }
1644 
1645 static PetscErrorCode MatDuplicate_MPISBAIJ(Mat, MatDuplicateOption, Mat *);
1646 
1647 static PetscErrorCode MatEqual_MPISBAIJ(Mat A, Mat B, PetscBool *flag)
1648 {
1649   Mat_MPISBAIJ *matB = (Mat_MPISBAIJ *)B->data, *matA = (Mat_MPISBAIJ *)A->data;
1650   Mat           a, b, c, d;
1651   PetscBool     flg;
1652 
1653   PetscFunctionBegin;
1654   a = matA->A;
1655   b = matA->B;
1656   c = matB->A;
1657   d = matB->B;
1658 
1659   PetscCall(MatEqual(a, c, &flg));
1660   if (flg) PetscCall(MatEqual(b, d, &flg));
1661   PetscCallMPI(MPIU_Allreduce(&flg, flag, 1, MPIU_BOOL, MPI_LAND, PetscObjectComm((PetscObject)A)));
1662   PetscFunctionReturn(PETSC_SUCCESS);
1663 }
1664 
1665 static PetscErrorCode MatCopy_MPISBAIJ(Mat A, Mat B, MatStructure str)
1666 {
1667   PetscBool isbaij;
1668 
1669   PetscFunctionBegin;
1670   PetscCall(PetscObjectTypeCompareAny((PetscObject)B, &isbaij, MATSEQSBAIJ, MATMPISBAIJ, ""));
1671   PetscCheck(isbaij, PetscObjectComm((PetscObject)B), PETSC_ERR_SUP, "Not for matrix type %s", ((PetscObject)B)->type_name);
1672   /* If the two matrices don't have the same copy implementation, they aren't compatible for fast copy. */
1673   if ((str != SAME_NONZERO_PATTERN) || (A->ops->copy != B->ops->copy)) {
1674     PetscCall(MatGetRowUpperTriangular(A));
1675     PetscCall(MatCopy_Basic(A, B, str));
1676     PetscCall(MatRestoreRowUpperTriangular(A));
1677   } else {
1678     Mat_MPISBAIJ *a = (Mat_MPISBAIJ *)A->data;
1679     Mat_MPISBAIJ *b = (Mat_MPISBAIJ *)B->data;
1680 
1681     PetscCall(MatCopy(a->A, b->A, str));
1682     PetscCall(MatCopy(a->B, b->B, str));
1683   }
1684   PetscCall(PetscObjectStateIncrease((PetscObject)B));
1685   PetscFunctionReturn(PETSC_SUCCESS);
1686 }
1687 
1688 static PetscErrorCode MatAXPY_MPISBAIJ(Mat Y, PetscScalar a, Mat X, MatStructure str)
1689 {
1690   Mat_MPISBAIJ *xx = (Mat_MPISBAIJ *)X->data, *yy = (Mat_MPISBAIJ *)Y->data;
1691   PetscBLASInt  bnz, one                          = 1;
1692   Mat_SeqSBAIJ *xa, *ya;
1693   Mat_SeqBAIJ  *xb, *yb;
1694 
1695   PetscFunctionBegin;
1696   if (str == SAME_NONZERO_PATTERN) {
1697     PetscScalar alpha = a;
1698     xa                = (Mat_SeqSBAIJ *)xx->A->data;
1699     ya                = (Mat_SeqSBAIJ *)yy->A->data;
1700     PetscCall(PetscBLASIntCast(xa->nz, &bnz));
1701     PetscCallBLAS("BLASaxpy", BLASaxpy_(&bnz, &alpha, xa->a, &one, ya->a, &one));
1702     xb = (Mat_SeqBAIJ *)xx->B->data;
1703     yb = (Mat_SeqBAIJ *)yy->B->data;
1704     PetscCall(PetscBLASIntCast(xb->nz, &bnz));
1705     PetscCallBLAS("BLASaxpy", BLASaxpy_(&bnz, &alpha, xb->a, &one, yb->a, &one));
1706     PetscCall(PetscObjectStateIncrease((PetscObject)Y));
1707   } else if (str == SUBSET_NONZERO_PATTERN) { /* nonzeros of X is a subset of Y's */
1708     PetscCall(MatSetOption(X, MAT_GETROW_UPPERTRIANGULAR, PETSC_TRUE));
1709     PetscCall(MatAXPY_Basic(Y, a, X, str));
1710     PetscCall(MatSetOption(X, MAT_GETROW_UPPERTRIANGULAR, PETSC_FALSE));
1711   } else {
1712     Mat       B;
1713     PetscInt *nnz_d, *nnz_o, bs = Y->rmap->bs;
1714     PetscCheck(bs == X->rmap->bs, PETSC_COMM_SELF, PETSC_ERR_ARG_SIZ, "Matrices must have same block size");
1715     PetscCall(MatGetRowUpperTriangular(X));
1716     PetscCall(MatGetRowUpperTriangular(Y));
1717     PetscCall(PetscMalloc1(yy->A->rmap->N, &nnz_d));
1718     PetscCall(PetscMalloc1(yy->B->rmap->N, &nnz_o));
1719     PetscCall(MatCreate(PetscObjectComm((PetscObject)Y), &B));
1720     PetscCall(PetscObjectSetName((PetscObject)B, ((PetscObject)Y)->name));
1721     PetscCall(MatSetSizes(B, Y->rmap->n, Y->cmap->n, Y->rmap->N, Y->cmap->N));
1722     PetscCall(MatSetBlockSizesFromMats(B, Y, Y));
1723     PetscCall(MatSetType(B, MATMPISBAIJ));
1724     PetscCall(MatAXPYGetPreallocation_SeqSBAIJ(yy->A, xx->A, nnz_d));
1725     PetscCall(MatAXPYGetPreallocation_MPIBAIJ(yy->B, yy->garray, xx->B, xx->garray, nnz_o));
1726     PetscCall(MatMPISBAIJSetPreallocation(B, bs, 0, nnz_d, 0, nnz_o));
1727     PetscCall(MatAXPY_BasicWithPreallocation(B, Y, a, X, str));
1728     PetscCall(MatHeaderMerge(Y, &B));
1729     PetscCall(PetscFree(nnz_d));
1730     PetscCall(PetscFree(nnz_o));
1731     PetscCall(MatRestoreRowUpperTriangular(X));
1732     PetscCall(MatRestoreRowUpperTriangular(Y));
1733   }
1734   PetscFunctionReturn(PETSC_SUCCESS);
1735 }
1736 
1737 static PetscErrorCode MatCreateSubMatrices_MPISBAIJ(Mat A, PetscInt n, const IS irow[], const IS icol[], MatReuse scall, Mat *B[])
1738 {
1739   PetscInt  i;
1740   PetscBool flg;
1741 
1742   PetscFunctionBegin;
1743   PetscCall(MatCreateSubMatrices_MPIBAIJ(A, n, irow, icol, scall, B)); /* B[] are sbaij matrices */
1744   for (i = 0; i < n; i++) {
1745     PetscCall(ISEqual(irow[i], icol[i], &flg));
1746     if (!flg) PetscCall(MatSeqSBAIJZeroOps_Private(*B[i]));
1747   }
1748   PetscFunctionReturn(PETSC_SUCCESS);
1749 }
1750 
1751 static PetscErrorCode MatShift_MPISBAIJ(Mat Y, PetscScalar a)
1752 {
1753   Mat_MPISBAIJ *maij = (Mat_MPISBAIJ *)Y->data;
1754   Mat_SeqSBAIJ *aij  = (Mat_SeqSBAIJ *)maij->A->data;
1755 
1756   PetscFunctionBegin;
1757   if (!Y->preallocated) {
1758     PetscCall(MatMPISBAIJSetPreallocation(Y, Y->rmap->bs, 1, NULL, 0, NULL));
1759   } else if (!aij->nz) {
1760     PetscInt nonew = aij->nonew;
1761     PetscCall(MatSeqSBAIJSetPreallocation(maij->A, Y->rmap->bs, 1, NULL));
1762     aij->nonew = nonew;
1763   }
1764   PetscCall(MatShift_Basic(Y, a));
1765   PetscFunctionReturn(PETSC_SUCCESS);
1766 }
1767 
1768 static PetscErrorCode MatMissingDiagonal_MPISBAIJ(Mat A, PetscBool *missing, PetscInt *d)
1769 {
1770   Mat_MPISBAIJ *a = (Mat_MPISBAIJ *)A->data;
1771 
1772   PetscFunctionBegin;
1773   PetscCheck(A->rmap->n == A->cmap->n, PETSC_COMM_SELF, PETSC_ERR_SUP, "Only works for square matrices");
1774   PetscCall(MatMissingDiagonal(a->A, missing, d));
1775   if (d) {
1776     PetscInt rstart;
1777     PetscCall(MatGetOwnershipRange(A, &rstart, NULL));
1778     *d += rstart / A->rmap->bs;
1779   }
1780   PetscFunctionReturn(PETSC_SUCCESS);
1781 }
1782 
1783 static PetscErrorCode MatGetDiagonalBlock_MPISBAIJ(Mat A, Mat *a)
1784 {
1785   PetscFunctionBegin;
1786   *a = ((Mat_MPISBAIJ *)A->data)->A;
1787   PetscFunctionReturn(PETSC_SUCCESS);
1788 }
1789 
1790 static PetscErrorCode MatEliminateZeros_MPISBAIJ(Mat A, PetscBool keep)
1791 {
1792   Mat_MPISBAIJ *a = (Mat_MPISBAIJ *)A->data;
1793 
1794   PetscFunctionBegin;
1795   PetscCall(MatEliminateZeros_SeqSBAIJ(a->A, keep));       // possibly keep zero diagonal coefficients
1796   PetscCall(MatEliminateZeros_SeqBAIJ(a->B, PETSC_FALSE)); // never keep zero diagonal coefficients
1797   PetscFunctionReturn(PETSC_SUCCESS);
1798 }
1799 
1800 static PetscErrorCode MatLoad_MPISBAIJ(Mat, PetscViewer);
1801 static PetscErrorCode MatGetRowMaxAbs_MPISBAIJ(Mat, Vec, PetscInt[]);
1802 static PetscErrorCode MatSOR_MPISBAIJ(Mat, Vec, PetscReal, MatSORType, PetscReal, PetscInt, PetscInt, Vec);
1803 
1804 static struct _MatOps MatOps_Values = {MatSetValues_MPISBAIJ,
1805                                        MatGetRow_MPISBAIJ,
1806                                        MatRestoreRow_MPISBAIJ,
1807                                        MatMult_MPISBAIJ,
1808                                        /*  4*/ MatMultAdd_MPISBAIJ,
1809                                        MatMult_MPISBAIJ, /* transpose versions are same as non-transpose */
1810                                        MatMultAdd_MPISBAIJ,
1811                                        NULL,
1812                                        NULL,
1813                                        NULL,
1814                                        /* 10*/ NULL,
1815                                        NULL,
1816                                        NULL,
1817                                        MatSOR_MPISBAIJ,
1818                                        MatTranspose_MPISBAIJ,
1819                                        /* 15*/ MatGetInfo_MPISBAIJ,
1820                                        MatEqual_MPISBAIJ,
1821                                        MatGetDiagonal_MPISBAIJ,
1822                                        MatDiagonalScale_MPISBAIJ,
1823                                        MatNorm_MPISBAIJ,
1824                                        /* 20*/ MatAssemblyBegin_MPISBAIJ,
1825                                        MatAssemblyEnd_MPISBAIJ,
1826                                        MatSetOption_MPISBAIJ,
1827                                        MatZeroEntries_MPISBAIJ,
1828                                        /* 24*/ NULL,
1829                                        NULL,
1830                                        NULL,
1831                                        NULL,
1832                                        NULL,
1833                                        /* 29*/ MatSetUp_MPI_Hash,
1834                                        NULL,
1835                                        NULL,
1836                                        MatGetDiagonalBlock_MPISBAIJ,
1837                                        NULL,
1838                                        /* 34*/ MatDuplicate_MPISBAIJ,
1839                                        NULL,
1840                                        NULL,
1841                                        NULL,
1842                                        NULL,
1843                                        /* 39*/ MatAXPY_MPISBAIJ,
1844                                        MatCreateSubMatrices_MPISBAIJ,
1845                                        MatIncreaseOverlap_MPISBAIJ,
1846                                        MatGetValues_MPISBAIJ,
1847                                        MatCopy_MPISBAIJ,
1848                                        /* 44*/ NULL,
1849                                        MatScale_MPISBAIJ,
1850                                        MatShift_MPISBAIJ,
1851                                        NULL,
1852                                        NULL,
1853                                        /* 49*/ NULL,
1854                                        NULL,
1855                                        NULL,
1856                                        NULL,
1857                                        NULL,
1858                                        /* 54*/ NULL,
1859                                        NULL,
1860                                        MatSetUnfactored_MPISBAIJ,
1861                                        NULL,
1862                                        MatSetValuesBlocked_MPISBAIJ,
1863                                        /* 59*/ MatCreateSubMatrix_MPISBAIJ,
1864                                        NULL,
1865                                        NULL,
1866                                        NULL,
1867                                        NULL,
1868                                        /* 64*/ NULL,
1869                                        NULL,
1870                                        NULL,
1871                                        NULL,
1872                                        NULL,
1873                                        /* 69*/ MatGetRowMaxAbs_MPISBAIJ,
1874                                        NULL,
1875                                        MatConvert_MPISBAIJ_Basic,
1876                                        NULL,
1877                                        NULL,
1878                                        /* 74*/ NULL,
1879                                        NULL,
1880                                        NULL,
1881                                        NULL,
1882                                        NULL,
1883                                        /* 79*/ NULL,
1884                                        NULL,
1885                                        NULL,
1886                                        NULL,
1887                                        MatLoad_MPISBAIJ,
1888                                        /* 84*/ NULL,
1889                                        NULL,
1890                                        NULL,
1891                                        NULL,
1892                                        NULL,
1893                                        /* 89*/ NULL,
1894                                        NULL,
1895                                        NULL,
1896                                        NULL,
1897                                        NULL,
1898                                        /* 94*/ NULL,
1899                                        NULL,
1900                                        NULL,
1901                                        NULL,
1902                                        NULL,
1903                                        /* 99*/ NULL,
1904                                        NULL,
1905                                        NULL,
1906                                        MatConjugate_MPISBAIJ,
1907                                        NULL,
1908                                        /*104*/ NULL,
1909                                        MatRealPart_MPISBAIJ,
1910                                        MatImaginaryPart_MPISBAIJ,
1911                                        MatGetRowUpperTriangular_MPISBAIJ,
1912                                        MatRestoreRowUpperTriangular_MPISBAIJ,
1913                                        /*109*/ NULL,
1914                                        NULL,
1915                                        NULL,
1916                                        NULL,
1917                                        MatMissingDiagonal_MPISBAIJ,
1918                                        /*114*/ NULL,
1919                                        NULL,
1920                                        NULL,
1921                                        NULL,
1922                                        NULL,
1923                                        /*119*/ NULL,
1924                                        NULL,
1925                                        NULL,
1926                                        NULL,
1927                                        NULL,
1928                                        /*124*/ NULL,
1929                                        NULL,
1930                                        NULL,
1931                                        NULL,
1932                                        NULL,
1933                                        /*129*/ NULL,
1934                                        NULL,
1935                                        NULL,
1936                                        NULL,
1937                                        NULL,
1938                                        /*134*/ NULL,
1939                                        NULL,
1940                                        NULL,
1941                                        NULL,
1942                                        NULL,
1943                                        /*139*/ MatSetBlockSizes_Default,
1944                                        NULL,
1945                                        NULL,
1946                                        NULL,
1947                                        NULL,
1948                                        /*144*/ MatCreateMPIMatConcatenateSeqMat_MPISBAIJ,
1949                                        NULL,
1950                                        NULL,
1951                                        NULL,
1952                                        NULL,
1953                                        NULL,
1954                                        /*150*/ NULL,
1955                                        MatEliminateZeros_MPISBAIJ,
1956                                        NULL,
1957                                        NULL,
1958                                        NULL,
1959                                        /*155*/ NULL,
1960                                        MatCopyHashToXAIJ_MPI_Hash};
1961 
1962 static PetscErrorCode MatMPISBAIJSetPreallocation_MPISBAIJ(Mat B, PetscInt bs, PetscInt d_nz, const PetscInt *d_nnz, PetscInt o_nz, const PetscInt *o_nnz)
1963 {
1964   Mat_MPISBAIJ *b = (Mat_MPISBAIJ *)B->data;
1965   PetscInt      i, mbs, Mbs;
1966   PetscMPIInt   size;
1967 
1968   PetscFunctionBegin;
1969   if (B->hash_active) {
1970     B->ops[0]      = b->cops;
1971     B->hash_active = PETSC_FALSE;
1972   }
1973   if (!B->preallocated) PetscCall(MatStashCreate_Private(PetscObjectComm((PetscObject)B), bs, &B->bstash));
1974   PetscCall(MatSetBlockSize(B, PetscAbs(bs)));
1975   PetscCall(PetscLayoutSetUp(B->rmap));
1976   PetscCall(PetscLayoutSetUp(B->cmap));
1977   PetscCall(PetscLayoutGetBlockSize(B->rmap, &bs));
1978   PetscCheck(B->rmap->N <= B->cmap->N, PetscObjectComm((PetscObject)B), PETSC_ERR_SUP, "MPISBAIJ matrix cannot have more rows %" PetscInt_FMT " than columns %" PetscInt_FMT, B->rmap->N, B->cmap->N);
1979   PetscCheck(B->rmap->n <= B->cmap->n, PETSC_COMM_SELF, PETSC_ERR_SUP, "MPISBAIJ matrix cannot have more local rows %" PetscInt_FMT " than columns %" PetscInt_FMT, B->rmap->n, B->cmap->n);
1980 
1981   mbs = B->rmap->n / bs;
1982   Mbs = B->rmap->N / bs;
1983   PetscCheck(mbs * bs == B->rmap->n, PETSC_COMM_SELF, PETSC_ERR_ARG_SIZ, "No of local rows %" PetscInt_FMT " must be divisible by blocksize %" PetscInt_FMT, B->rmap->N, bs);
1984 
1985   B->rmap->bs = bs;
1986   b->bs2      = bs * bs;
1987   b->mbs      = mbs;
1988   b->Mbs      = Mbs;
1989   b->nbs      = B->cmap->n / bs;
1990   b->Nbs      = B->cmap->N / bs;
1991 
1992   for (i = 0; i <= b->size; i++) b->rangebs[i] = B->rmap->range[i] / bs;
1993   b->rstartbs = B->rmap->rstart / bs;
1994   b->rendbs   = B->rmap->rend / bs;
1995 
1996   b->cstartbs = B->cmap->rstart / bs;
1997   b->cendbs   = B->cmap->rend / bs;
1998 
1999 #if defined(PETSC_USE_CTABLE)
2000   PetscCall(PetscHMapIDestroy(&b->colmap));
2001 #else
2002   PetscCall(PetscFree(b->colmap));
2003 #endif
2004   PetscCall(PetscFree(b->garray));
2005   PetscCall(VecDestroy(&b->lvec));
2006   PetscCall(VecScatterDestroy(&b->Mvctx));
2007   PetscCall(VecDestroy(&b->slvec0));
2008   PetscCall(VecDestroy(&b->slvec0b));
2009   PetscCall(VecDestroy(&b->slvec1));
2010   PetscCall(VecDestroy(&b->slvec1a));
2011   PetscCall(VecDestroy(&b->slvec1b));
2012   PetscCall(VecScatterDestroy(&b->sMvctx));
2013 
2014   PetscCallMPI(MPI_Comm_size(PetscObjectComm((PetscObject)B), &size));
2015 
2016   MatSeqXAIJGetOptions_Private(b->B);
2017   PetscCall(MatDestroy(&b->B));
2018   PetscCall(MatCreate(PETSC_COMM_SELF, &b->B));
2019   PetscCall(MatSetSizes(b->B, B->rmap->n, size > 1 ? B->cmap->N : 0, B->rmap->n, size > 1 ? B->cmap->N : 0));
2020   PetscCall(MatSetType(b->B, MATSEQBAIJ));
2021   MatSeqXAIJRestoreOptions_Private(b->B);
2022 
2023   MatSeqXAIJGetOptions_Private(b->A);
2024   PetscCall(MatDestroy(&b->A));
2025   PetscCall(MatCreate(PETSC_COMM_SELF, &b->A));
2026   PetscCall(MatSetSizes(b->A, B->rmap->n, B->cmap->n, B->rmap->n, B->cmap->n));
2027   PetscCall(MatSetType(b->A, MATSEQSBAIJ));
2028   MatSeqXAIJRestoreOptions_Private(b->A);
2029 
2030   PetscCall(MatSeqSBAIJSetPreallocation(b->A, bs, d_nz, d_nnz));
2031   PetscCall(MatSeqBAIJSetPreallocation(b->B, bs, o_nz, o_nnz));
2032 
2033   B->preallocated  = PETSC_TRUE;
2034   B->was_assembled = PETSC_FALSE;
2035   B->assembled     = PETSC_FALSE;
2036   PetscFunctionReturn(PETSC_SUCCESS);
2037 }
2038 
2039 static PetscErrorCode MatMPISBAIJSetPreallocationCSR_MPISBAIJ(Mat B, PetscInt bs, const PetscInt ii[], const PetscInt jj[], const PetscScalar V[])
2040 {
2041   PetscInt        m, rstart, cend;
2042   PetscInt        i, j, d, nz, bd, nz_max = 0, *d_nnz = NULL, *o_nnz = NULL;
2043   const PetscInt *JJ          = NULL;
2044   PetscScalar    *values      = NULL;
2045   PetscBool       roworiented = ((Mat_MPISBAIJ *)B->data)->roworiented;
2046   PetscBool       nooffprocentries;
2047 
2048   PetscFunctionBegin;
2049   PetscCheck(bs >= 1, PetscObjectComm((PetscObject)B), PETSC_ERR_ARG_OUTOFRANGE, "Invalid block size specified, must be positive but it is %" PetscInt_FMT, bs);
2050   PetscCall(PetscLayoutSetBlockSize(B->rmap, bs));
2051   PetscCall(PetscLayoutSetBlockSize(B->cmap, bs));
2052   PetscCall(PetscLayoutSetUp(B->rmap));
2053   PetscCall(PetscLayoutSetUp(B->cmap));
2054   PetscCall(PetscLayoutGetBlockSize(B->rmap, &bs));
2055   m      = B->rmap->n / bs;
2056   rstart = B->rmap->rstart / bs;
2057   cend   = B->cmap->rend / bs;
2058 
2059   PetscCheck(!ii[0], PETSC_COMM_SELF, PETSC_ERR_ARG_OUTOFRANGE, "ii[0] must be 0 but it is %" PetscInt_FMT, ii[0]);
2060   PetscCall(PetscMalloc2(m, &d_nnz, m, &o_nnz));
2061   for (i = 0; i < m; i++) {
2062     nz = ii[i + 1] - ii[i];
2063     PetscCheck(nz >= 0, PETSC_COMM_SELF, PETSC_ERR_ARG_OUTOFRANGE, "Local row %" PetscInt_FMT " has a negative number of columns %" PetscInt_FMT, i, nz);
2064     /* count the ones on the diagonal and above, split into diagonal and off-diagonal portions. */
2065     JJ = jj + ii[i];
2066     bd = 0;
2067     for (j = 0; j < nz; j++) {
2068       if (*JJ >= i + rstart) break;
2069       JJ++;
2070       bd++;
2071     }
2072     d = 0;
2073     for (; j < nz; j++) {
2074       if (*JJ++ >= cend) break;
2075       d++;
2076     }
2077     d_nnz[i] = d;
2078     o_nnz[i] = nz - d - bd;
2079     nz       = nz - bd;
2080     nz_max   = PetscMax(nz_max, nz);
2081   }
2082   PetscCall(MatMPISBAIJSetPreallocation(B, bs, 0, d_nnz, 0, o_nnz));
2083   PetscCall(MatSetOption(B, MAT_IGNORE_LOWER_TRIANGULAR, PETSC_TRUE));
2084   PetscCall(PetscFree2(d_nnz, o_nnz));
2085 
2086   values = (PetscScalar *)V;
2087   if (!values) PetscCall(PetscCalloc1(bs * bs * nz_max, &values));
2088   for (i = 0; i < m; i++) {
2089     PetscInt        row   = i + rstart;
2090     PetscInt        ncols = ii[i + 1] - ii[i];
2091     const PetscInt *icols = jj + ii[i];
2092     if (bs == 1 || !roworiented) { /* block ordering matches the non-nested layout of MatSetValues so we can insert entire rows */
2093       const PetscScalar *svals = values + (V ? (bs * bs * ii[i]) : 0);
2094       PetscCall(MatSetValuesBlocked_MPISBAIJ(B, 1, &row, ncols, icols, svals, INSERT_VALUES));
2095     } else { /* block ordering does not match so we can only insert one block at a time. */
2096       PetscInt j;
2097       for (j = 0; j < ncols; j++) {
2098         const PetscScalar *svals = values + (V ? (bs * bs * (ii[i] + j)) : 0);
2099         PetscCall(MatSetValuesBlocked_MPISBAIJ(B, 1, &row, 1, &icols[j], svals, INSERT_VALUES));
2100       }
2101     }
2102   }
2103 
2104   if (!V) PetscCall(PetscFree(values));
2105   nooffprocentries    = B->nooffprocentries;
2106   B->nooffprocentries = PETSC_TRUE;
2107   PetscCall(MatAssemblyBegin(B, MAT_FINAL_ASSEMBLY));
2108   PetscCall(MatAssemblyEnd(B, MAT_FINAL_ASSEMBLY));
2109   B->nooffprocentries = nooffprocentries;
2110 
2111   PetscCall(MatSetOption(B, MAT_NEW_NONZERO_LOCATION_ERR, PETSC_TRUE));
2112   PetscFunctionReturn(PETSC_SUCCESS);
2113 }
2114 
2115 /*MC
2116    MATMPISBAIJ - MATMPISBAIJ = "mpisbaij" - A matrix type to be used for distributed symmetric sparse block matrices,
2117    based on block compressed sparse row format.  Only the upper triangular portion of the "diagonal" portion of
2118    the matrix is stored.
2119 
2120    For complex numbers by default this matrix is symmetric, NOT Hermitian symmetric. To make it Hermitian symmetric you
2121    can call `MatSetOption`(`Mat`, `MAT_HERMITIAN`);
2122 
2123    Options Database Key:
2124 . -mat_type mpisbaij - sets the matrix type to "mpisbaij" during a call to `MatSetFromOptions()`
2125 
2126    Level: beginner
2127 
2128    Note:
2129      The number of rows in the matrix must be less than or equal to the number of columns. Similarly the number of rows in the
2130      diagonal portion of the matrix of each process has to less than or equal the number of columns.
2131 
2132 .seealso: [](ch_matrices), `Mat`, `MATSBAIJ`, `MATBAIJ`, `MatCreateBAIJ()`, `MATSEQSBAIJ`, `MatType`
2133 M*/
2134 
2135 PETSC_EXTERN PetscErrorCode MatCreate_MPISBAIJ(Mat B)
2136 {
2137   Mat_MPISBAIJ *b;
2138   PetscBool     flg = PETSC_FALSE;
2139 
2140   PetscFunctionBegin;
2141   PetscCall(PetscNew(&b));
2142   B->data   = (void *)b;
2143   B->ops[0] = MatOps_Values;
2144 
2145   B->ops->destroy = MatDestroy_MPISBAIJ;
2146   B->ops->view    = MatView_MPISBAIJ;
2147   B->assembled    = PETSC_FALSE;
2148   B->insertmode   = NOT_SET_VALUES;
2149 
2150   PetscCallMPI(MPI_Comm_rank(PetscObjectComm((PetscObject)B), &b->rank));
2151   PetscCallMPI(MPI_Comm_size(PetscObjectComm((PetscObject)B), &b->size));
2152 
2153   /* build local table of row and column ownerships */
2154   PetscCall(PetscMalloc1(b->size + 2, &b->rangebs));
2155 
2156   /* build cache for off array entries formed */
2157   PetscCall(MatStashCreate_Private(PetscObjectComm((PetscObject)B), 1, &B->stash));
2158 
2159   b->donotstash  = PETSC_FALSE;
2160   b->colmap      = NULL;
2161   b->garray      = NULL;
2162   b->roworiented = PETSC_TRUE;
2163 
2164   /* stuff used in block assembly */
2165   b->barray = NULL;
2166 
2167   /* stuff used for matrix vector multiply */
2168   b->lvec    = NULL;
2169   b->Mvctx   = NULL;
2170   b->slvec0  = NULL;
2171   b->slvec0b = NULL;
2172   b->slvec1  = NULL;
2173   b->slvec1a = NULL;
2174   b->slvec1b = NULL;
2175   b->sMvctx  = NULL;
2176 
2177   /* stuff for MatGetRow() */
2178   b->rowindices   = NULL;
2179   b->rowvalues    = NULL;
2180   b->getrowactive = PETSC_FALSE;
2181 
2182   /* hash table stuff */
2183   b->ht           = NULL;
2184   b->hd           = NULL;
2185   b->ht_size      = 0;
2186   b->ht_flag      = PETSC_FALSE;
2187   b->ht_fact      = 0;
2188   b->ht_total_ct  = 0;
2189   b->ht_insert_ct = 0;
2190 
2191   /* stuff for MatCreateSubMatrices_MPIBAIJ_local() */
2192   b->ijonly = PETSC_FALSE;
2193 
2194   b->in_loc = NULL;
2195   b->v_loc  = NULL;
2196   b->n_loc  = 0;
2197 
2198   PetscCall(PetscObjectComposeFunction((PetscObject)B, "MatStoreValues_C", MatStoreValues_MPISBAIJ));
2199   PetscCall(PetscObjectComposeFunction((PetscObject)B, "MatRetrieveValues_C", MatRetrieveValues_MPISBAIJ));
2200   PetscCall(PetscObjectComposeFunction((PetscObject)B, "MatMPISBAIJSetPreallocation_C", MatMPISBAIJSetPreallocation_MPISBAIJ));
2201   PetscCall(PetscObjectComposeFunction((PetscObject)B, "MatMPISBAIJSetPreallocationCSR_C", MatMPISBAIJSetPreallocationCSR_MPISBAIJ));
2202 #if defined(PETSC_HAVE_ELEMENTAL)
2203   PetscCall(PetscObjectComposeFunction((PetscObject)B, "MatConvert_mpisbaij_elemental_C", MatConvert_MPISBAIJ_Elemental));
2204 #endif
2205 #if defined(PETSC_HAVE_SCALAPACK)
2206   PetscCall(PetscObjectComposeFunction((PetscObject)B, "MatConvert_mpisbaij_scalapack_C", MatConvert_SBAIJ_ScaLAPACK));
2207 #endif
2208   PetscCall(PetscObjectComposeFunction((PetscObject)B, "MatConvert_mpisbaij_mpiaij_C", MatConvert_MPISBAIJ_Basic));
2209   PetscCall(PetscObjectComposeFunction((PetscObject)B, "MatConvert_mpisbaij_mpibaij_C", MatConvert_MPISBAIJ_Basic));
2210 
2211   B->symmetric                   = PETSC_BOOL3_TRUE;
2212   B->structurally_symmetric      = PETSC_BOOL3_TRUE;
2213   B->symmetry_eternal            = PETSC_TRUE;
2214   B->structural_symmetry_eternal = PETSC_TRUE;
2215 #if defined(PETSC_USE_COMPLEX)
2216   B->hermitian = PETSC_BOOL3_FALSE;
2217 #else
2218   B->hermitian = PETSC_BOOL3_TRUE;
2219 #endif
2220 
2221   PetscCall(PetscObjectChangeTypeName((PetscObject)B, MATMPISBAIJ));
2222   PetscOptionsBegin(PetscObjectComm((PetscObject)B), NULL, "Options for loading MPISBAIJ matrix 1", "Mat");
2223   PetscCall(PetscOptionsBool("-mat_use_hash_table", "Use hash table to save memory in constructing matrix", "MatSetOption", flg, &flg, NULL));
2224   if (flg) {
2225     PetscReal fact = 1.39;
2226     PetscCall(MatSetOption(B, MAT_USE_HASH_TABLE, PETSC_TRUE));
2227     PetscCall(PetscOptionsReal("-mat_use_hash_table", "Use hash table factor", "MatMPIBAIJSetHashTableFactor", fact, &fact, NULL));
2228     if (fact <= 1.0) fact = 1.39;
2229     PetscCall(MatMPIBAIJSetHashTableFactor(B, fact));
2230     PetscCall(PetscInfo(B, "Hash table Factor used %5.2g\n", (double)fact));
2231   }
2232   PetscOptionsEnd();
2233   PetscFunctionReturn(PETSC_SUCCESS);
2234 }
2235 
2236 // PetscClangLinter pragma disable: -fdoc-section-header-unknown
2237 /*MC
2238    MATSBAIJ - MATSBAIJ = "sbaij" - A matrix type to be used for symmetric block sparse matrices.
2239 
2240    This matrix type is identical to `MATSEQSBAIJ` when constructed with a single process communicator,
2241    and `MATMPISBAIJ` otherwise.
2242 
2243    Options Database Key:
2244 . -mat_type sbaij - sets the matrix type to `MATSBAIJ` during a call to `MatSetFromOptions()`
2245 
2246   Level: beginner
2247 
2248 .seealso: [](ch_matrices), `Mat`, `MATSEQSBAIJ`, `MATMPISBAIJ`, `MatCreateSBAIJ()`, `MATSEQSBAIJ`, `MATMPISBAIJ`
2249 M*/
2250 
2251 /*@
2252   MatMPISBAIJSetPreallocation - For good matrix assembly performance
2253   the user should preallocate the matrix storage by setting the parameters
2254   d_nz (or d_nnz) and o_nz (or o_nnz).  By setting these parameters accurately,
2255   performance can be increased by more than a factor of 50.
2256 
2257   Collective
2258 
2259   Input Parameters:
2260 + B     - the matrix
2261 . bs    - size of block, the blocks are ALWAYS square. One can use MatSetBlockSizes() to set a different row and column blocksize but the row
2262           blocksize always defines the size of the blocks. The column blocksize sets the blocksize of the vectors obtained with MatCreateVecs()
2263 . d_nz  - number of block nonzeros per block row in diagonal portion of local
2264           submatrix  (same for all local rows)
2265 . d_nnz - array containing the number of block nonzeros in the various block rows
2266           in the upper triangular and diagonal part of the in diagonal portion of the local
2267           (possibly different for each block row) or `NULL`.  If you plan to factor the matrix you must leave room
2268           for the diagonal entry and set a value even if it is zero.
2269 . o_nz  - number of block nonzeros per block row in the off-diagonal portion of local
2270           submatrix (same for all local rows).
2271 - o_nnz - array containing the number of nonzeros in the various block rows of the
2272           off-diagonal portion of the local submatrix that is right of the diagonal
2273           (possibly different for each block row) or `NULL`.
2274 
2275   Options Database Keys:
2276 + -mat_no_unroll  - uses code that does not unroll the loops in the
2277                     block calculations (much slower)
2278 - -mat_block_size - size of the blocks to use
2279 
2280   Level: intermediate
2281 
2282   Notes:
2283 
2284   If `PETSC_DECIDE` or `PETSC_DETERMINE` is used for a particular argument on one processor
2285   than it must be used on all processors that share the object for that argument.
2286 
2287   If the *_nnz parameter is given then the *_nz parameter is ignored
2288 
2289   Storage Information:
2290   For a square global matrix we define each processor's diagonal portion
2291   to be its local rows and the corresponding columns (a square submatrix);
2292   each processor's off-diagonal portion encompasses the remainder of the
2293   local matrix (a rectangular submatrix).
2294 
2295   The user can specify preallocated storage for the diagonal part of
2296   the local submatrix with either `d_nz` or `d_nnz` (not both).  Set
2297   `d_nz` = `PETSC_DEFAULT` and `d_nnz` = `NULL` for PETSc to control dynamic
2298   memory allocation.  Likewise, specify preallocated storage for the
2299   off-diagonal part of the local submatrix with `o_nz` or `o_nnz` (not both).
2300 
2301   You can call `MatGetInfo()` to get information on how effective the preallocation was;
2302   for example the fields mallocs,nz_allocated,nz_used,nz_unneeded;
2303   You can also run with the option `-info` and look for messages with the string
2304   malloc in them to see if additional memory allocation was needed.
2305 
2306   Consider a processor that owns rows 3, 4 and 5 of a parallel matrix. In
2307   the figure below we depict these three local rows and all columns (0-11).
2308 
2309 .vb
2310            0 1 2 3 4 5 6 7 8 9 10 11
2311           --------------------------
2312    row 3  |. . . d d d o o o o  o  o
2313    row 4  |. . . d d d o o o o  o  o
2314    row 5  |. . . d d d o o o o  o  o
2315           --------------------------
2316 .ve
2317 
2318   Thus, any entries in the d locations are stored in the d (diagonal)
2319   submatrix, and any entries in the o locations are stored in the
2320   o (off-diagonal) submatrix.  Note that the d matrix is stored in
2321   `MATSEQSBAIJ` format and the o submatrix in `MATSEQBAIJ` format.
2322 
2323   Now `d_nz` should indicate the number of block nonzeros per row in the upper triangular
2324   plus the diagonal part of the d matrix,
2325   and `o_nz` should indicate the number of block nonzeros per row in the o matrix
2326 
2327   In general, for PDE problems in which most nonzeros are near the diagonal,
2328   one expects `d_nz` >> `o_nz`.
2329 
2330 .seealso: [](ch_matrices), `Mat`, `MATMPISBAIJ`, `MATSBAIJ`, `MatCreate()`, `MatCreateSeqSBAIJ()`, `MatSetValues()`, `MatCreateBAIJ()`, `PetscSplitOwnership()`
2331 @*/
2332 PetscErrorCode MatMPISBAIJSetPreallocation(Mat B, PetscInt bs, PetscInt d_nz, const PetscInt d_nnz[], PetscInt o_nz, const PetscInt o_nnz[])
2333 {
2334   PetscFunctionBegin;
2335   PetscValidHeaderSpecific(B, MAT_CLASSID, 1);
2336   PetscValidType(B, 1);
2337   PetscValidLogicalCollectiveInt(B, bs, 2);
2338   PetscTryMethod(B, "MatMPISBAIJSetPreallocation_C", (Mat, PetscInt, PetscInt, const PetscInt[], PetscInt, const PetscInt[]), (B, bs, d_nz, d_nnz, o_nz, o_nnz));
2339   PetscFunctionReturn(PETSC_SUCCESS);
2340 }
2341 
2342 // PetscClangLinter pragma disable: -fdoc-section-header-unknown
2343 /*@
2344   MatCreateSBAIJ - Creates a sparse parallel matrix in symmetric block AIJ format, `MATSBAIJ`,
2345   (block compressed row).  For good matrix assembly performance
2346   the user should preallocate the matrix storage by setting the parameters
2347   `d_nz` (or `d_nnz`) and `o_nz` (or `o_nnz`).
2348 
2349   Collective
2350 
2351   Input Parameters:
2352 + comm  - MPI communicator
2353 . bs    - size of block, the blocks are ALWAYS square. One can use `MatSetBlockSizes()` to set a different row and column blocksize but the row
2354           blocksize always defines the size of the blocks. The column blocksize sets the blocksize of the vectors obtained with `MatCreateVecs()`
2355 . m     - number of local rows (or `PETSC_DECIDE` to have calculated if `M` is given)
2356           This value should be the same as the local size used in creating the
2357           y vector for the matrix-vector product y = Ax.
2358 . n     - number of local columns (or `PETSC_DECIDE` to have calculated if `N` is given)
2359           This value should be the same as the local size used in creating the
2360           x vector for the matrix-vector product y = Ax.
2361 . M     - number of global rows (or `PETSC_DETERMINE` to have calculated if `m` is given)
2362 . N     - number of global columns (or `PETSC_DETERMINE` to have calculated if `n` is given)
2363 . d_nz  - number of block nonzeros per block row in diagonal portion of local
2364           submatrix (same for all local rows)
2365 . d_nnz - array containing the number of block nonzeros in the various block rows
2366           in the upper triangular portion of the in diagonal portion of the local
2367           (possibly different for each block block row) or `NULL`.
2368           If you plan to factor the matrix you must leave room for the diagonal entry and
2369           set its value even if it is zero.
2370 . o_nz  - number of block nonzeros per block row in the off-diagonal portion of local
2371           submatrix (same for all local rows).
2372 - o_nnz - array containing the number of nonzeros in the various block rows of the
2373           off-diagonal portion of the local submatrix (possibly different for
2374           each block row) or `NULL`.
2375 
2376   Output Parameter:
2377 . A - the matrix
2378 
2379   Options Database Keys:
2380 + -mat_no_unroll  - uses code that does not unroll the loops in the
2381                     block calculations (much slower)
2382 . -mat_block_size - size of the blocks to use
2383 - -mat_mpi        - use the parallel matrix data structures even on one processor
2384                     (defaults to using SeqBAIJ format on one processor)
2385 
2386   Level: intermediate
2387 
2388   Notes:
2389   It is recommended that one use `MatCreateFromOptions()` or the `MatCreate()`, `MatSetType()` and/or `MatSetFromOptions()`,
2390   MatXXXXSetPreallocation() paradigm instead of this routine directly.
2391   [MatXXXXSetPreallocation() is, for example, `MatSeqAIJSetPreallocation()`]
2392 
2393   The number of rows and columns must be divisible by blocksize.
2394   This matrix type does not support complex Hermitian operation.
2395 
2396   The user MUST specify either the local or global matrix dimensions
2397   (possibly both).
2398 
2399   If `PETSC_DECIDE` or `PETSC_DETERMINE` is used for a particular argument on one processor
2400   than it must be used on all processors that share the object for that argument.
2401 
2402   If `m` and `n` are not `PETSC_DECIDE`, then the values determines the `PetscLayout` of the matrix and the ranges returned by
2403   `MatGetOwnershipRange()`,  `MatGetOwnershipRanges()`, `MatGetOwnershipRangeColumn()`, and `MatGetOwnershipRangesColumn()`.
2404 
2405   If the *_nnz parameter is given then the *_nz parameter is ignored
2406 
2407   Storage Information:
2408   For a square global matrix we define each processor's diagonal portion
2409   to be its local rows and the corresponding columns (a square submatrix);
2410   each processor's off-diagonal portion encompasses the remainder of the
2411   local matrix (a rectangular submatrix).
2412 
2413   The user can specify preallocated storage for the diagonal part of
2414   the local submatrix with either `d_nz` or `d_nnz` (not both). Set
2415   `d_nz` = `PETSC_DEFAULT` and `d_nnz` = `NULL` for PETSc to control dynamic
2416   memory allocation. Likewise, specify preallocated storage for the
2417   off-diagonal part of the local submatrix with `o_nz` or `o_nnz` (not both).
2418 
2419   Consider a processor that owns rows 3, 4 and 5 of a parallel matrix. In
2420   the figure below we depict these three local rows and all columns (0-11).
2421 
2422 .vb
2423            0 1 2 3 4 5 6 7 8 9 10 11
2424           --------------------------
2425    row 3  |. . . d d d o o o o  o  o
2426    row 4  |. . . d d d o o o o  o  o
2427    row 5  |. . . d d d o o o o  o  o
2428           --------------------------
2429 .ve
2430 
2431   Thus, any entries in the d locations are stored in the d (diagonal)
2432   submatrix, and any entries in the o locations are stored in the
2433   o (off-diagonal) submatrix. Note that the d matrix is stored in
2434   `MATSEQSBAIJ` format and the o submatrix in `MATSEQBAIJ` format.
2435 
2436   Now `d_nz` should indicate the number of block nonzeros per row in the upper triangular
2437   plus the diagonal part of the d matrix,
2438   and `o_nz` should indicate the number of block nonzeros per row in the o matrix.
2439   In general, for PDE problems in which most nonzeros are near the diagonal,
2440   one expects `d_nz` >> `o_nz`.
2441 
2442 .seealso: [](ch_matrices), `Mat`, `MATSBAIJ`, `MatCreate()`, `MatCreateSeqSBAIJ()`, `MatSetValues()`, `MatCreateBAIJ()`,
2443           `MatGetOwnershipRange()`,  `MatGetOwnershipRanges()`, `MatGetOwnershipRangeColumn()`, `MatGetOwnershipRangesColumn()`, `PetscLayout`
2444 @*/
2445 PetscErrorCode MatCreateSBAIJ(MPI_Comm comm, PetscInt bs, PetscInt m, PetscInt n, PetscInt M, PetscInt N, PetscInt d_nz, const PetscInt d_nnz[], PetscInt o_nz, const PetscInt o_nnz[], Mat *A)
2446 {
2447   PetscMPIInt size;
2448 
2449   PetscFunctionBegin;
2450   PetscCall(MatCreate(comm, A));
2451   PetscCall(MatSetSizes(*A, m, n, M, N));
2452   PetscCallMPI(MPI_Comm_size(comm, &size));
2453   if (size > 1) {
2454     PetscCall(MatSetType(*A, MATMPISBAIJ));
2455     PetscCall(MatMPISBAIJSetPreallocation(*A, bs, d_nz, d_nnz, o_nz, o_nnz));
2456   } else {
2457     PetscCall(MatSetType(*A, MATSEQSBAIJ));
2458     PetscCall(MatSeqSBAIJSetPreallocation(*A, bs, d_nz, d_nnz));
2459   }
2460   PetscFunctionReturn(PETSC_SUCCESS);
2461 }
2462 
2463 static PetscErrorCode MatDuplicate_MPISBAIJ(Mat matin, MatDuplicateOption cpvalues, Mat *newmat)
2464 {
2465   Mat           mat;
2466   Mat_MPISBAIJ *a, *oldmat = (Mat_MPISBAIJ *)matin->data;
2467   PetscInt      len = 0, nt, bs = matin->rmap->bs, mbs = oldmat->mbs;
2468   PetscScalar  *array;
2469 
2470   PetscFunctionBegin;
2471   *newmat = NULL;
2472 
2473   PetscCall(MatCreate(PetscObjectComm((PetscObject)matin), &mat));
2474   PetscCall(MatSetSizes(mat, matin->rmap->n, matin->cmap->n, matin->rmap->N, matin->cmap->N));
2475   PetscCall(MatSetType(mat, ((PetscObject)matin)->type_name));
2476   PetscCall(PetscLayoutReference(matin->rmap, &mat->rmap));
2477   PetscCall(PetscLayoutReference(matin->cmap, &mat->cmap));
2478 
2479   if (matin->hash_active) {
2480     PetscCall(MatSetUp(mat));
2481   } else {
2482     mat->factortype   = matin->factortype;
2483     mat->preallocated = PETSC_TRUE;
2484     mat->assembled    = PETSC_TRUE;
2485     mat->insertmode   = NOT_SET_VALUES;
2486 
2487     a      = (Mat_MPISBAIJ *)mat->data;
2488     a->bs2 = oldmat->bs2;
2489     a->mbs = oldmat->mbs;
2490     a->nbs = oldmat->nbs;
2491     a->Mbs = oldmat->Mbs;
2492     a->Nbs = oldmat->Nbs;
2493 
2494     a->size         = oldmat->size;
2495     a->rank         = oldmat->rank;
2496     a->donotstash   = oldmat->donotstash;
2497     a->roworiented  = oldmat->roworiented;
2498     a->rowindices   = NULL;
2499     a->rowvalues    = NULL;
2500     a->getrowactive = PETSC_FALSE;
2501     a->barray       = NULL;
2502     a->rstartbs     = oldmat->rstartbs;
2503     a->rendbs       = oldmat->rendbs;
2504     a->cstartbs     = oldmat->cstartbs;
2505     a->cendbs       = oldmat->cendbs;
2506 
2507     /* hash table stuff */
2508     a->ht           = NULL;
2509     a->hd           = NULL;
2510     a->ht_size      = 0;
2511     a->ht_flag      = oldmat->ht_flag;
2512     a->ht_fact      = oldmat->ht_fact;
2513     a->ht_total_ct  = 0;
2514     a->ht_insert_ct = 0;
2515 
2516     PetscCall(PetscArraycpy(a->rangebs, oldmat->rangebs, a->size + 2));
2517     if (oldmat->colmap) {
2518 #if defined(PETSC_USE_CTABLE)
2519       PetscCall(PetscHMapIDuplicate(oldmat->colmap, &a->colmap));
2520 #else
2521       PetscCall(PetscMalloc1(a->Nbs, &a->colmap));
2522       PetscCall(PetscArraycpy(a->colmap, oldmat->colmap, a->Nbs));
2523 #endif
2524     } else a->colmap = NULL;
2525 
2526     if (oldmat->garray && (len = ((Mat_SeqBAIJ *)oldmat->B->data)->nbs)) {
2527       PetscCall(PetscMalloc1(len, &a->garray));
2528       PetscCall(PetscArraycpy(a->garray, oldmat->garray, len));
2529     } else a->garray = NULL;
2530 
2531     PetscCall(MatStashCreate_Private(PetscObjectComm((PetscObject)matin), matin->rmap->bs, &mat->bstash));
2532     PetscCall(VecDuplicate(oldmat->lvec, &a->lvec));
2533     PetscCall(VecScatterCopy(oldmat->Mvctx, &a->Mvctx));
2534 
2535     PetscCall(VecDuplicate(oldmat->slvec0, &a->slvec0));
2536     PetscCall(VecDuplicate(oldmat->slvec1, &a->slvec1));
2537 
2538     PetscCall(VecGetLocalSize(a->slvec1, &nt));
2539     PetscCall(VecGetArray(a->slvec1, &array));
2540     PetscCall(VecCreateSeqWithArray(PETSC_COMM_SELF, 1, bs * mbs, array, &a->slvec1a));
2541     PetscCall(VecCreateSeqWithArray(PETSC_COMM_SELF, 1, nt - bs * mbs, array + bs * mbs, &a->slvec1b));
2542     PetscCall(VecRestoreArray(a->slvec1, &array));
2543     PetscCall(VecGetArray(a->slvec0, &array));
2544     PetscCall(VecCreateSeqWithArray(PETSC_COMM_SELF, 1, nt - bs * mbs, array + bs * mbs, &a->slvec0b));
2545     PetscCall(VecRestoreArray(a->slvec0, &array));
2546 
2547     /* ierr =  VecScatterCopy(oldmat->sMvctx,&a->sMvctx); - not written yet, replaced by the lazy trick: */
2548     PetscCall(PetscObjectReference((PetscObject)oldmat->sMvctx));
2549     a->sMvctx = oldmat->sMvctx;
2550 
2551     PetscCall(MatDuplicate(oldmat->A, cpvalues, &a->A));
2552     PetscCall(MatDuplicate(oldmat->B, cpvalues, &a->B));
2553   }
2554   PetscCall(PetscFunctionListDuplicate(((PetscObject)matin)->qlist, &((PetscObject)mat)->qlist));
2555   *newmat = mat;
2556   PetscFunctionReturn(PETSC_SUCCESS);
2557 }
2558 
2559 /* Used for both MPIBAIJ and MPISBAIJ matrices */
2560 #define MatLoad_MPISBAIJ_Binary MatLoad_MPIBAIJ_Binary
2561 
2562 static PetscErrorCode MatLoad_MPISBAIJ(Mat mat, PetscViewer viewer)
2563 {
2564   PetscBool isbinary;
2565 
2566   PetscFunctionBegin;
2567   PetscCall(PetscObjectTypeCompare((PetscObject)viewer, PETSCVIEWERBINARY, &isbinary));
2568   PetscCheck(isbinary, PetscObjectComm((PetscObject)viewer), PETSC_ERR_SUP, "Viewer type %s not yet supported for reading %s matrices", ((PetscObject)viewer)->type_name, ((PetscObject)mat)->type_name);
2569   PetscCall(MatLoad_MPISBAIJ_Binary(mat, viewer));
2570   PetscFunctionReturn(PETSC_SUCCESS);
2571 }
2572 
2573 static PetscErrorCode MatGetRowMaxAbs_MPISBAIJ(Mat A, Vec v, PetscInt idx[])
2574 {
2575   Mat_MPISBAIJ *a = (Mat_MPISBAIJ *)A->data;
2576   Mat_SeqBAIJ  *b = (Mat_SeqBAIJ *)a->B->data;
2577   PetscReal     atmp;
2578   PetscReal    *work, *svalues, *rvalues;
2579   PetscInt      i, bs, mbs, *bi, *bj, brow, j, ncols, krow, kcol, col, row, Mbs, bcol;
2580   PetscMPIInt   rank, size;
2581   PetscInt     *rowners_bs, count, source;
2582   PetscScalar  *va;
2583   MatScalar    *ba;
2584   MPI_Status    stat;
2585 
2586   PetscFunctionBegin;
2587   PetscCheck(!idx, PETSC_COMM_SELF, PETSC_ERR_SUP, "Send email to petsc-maint@mcs.anl.gov");
2588   PetscCall(MatGetRowMaxAbs(a->A, v, NULL));
2589   PetscCall(VecGetArray(v, &va));
2590 
2591   PetscCallMPI(MPI_Comm_size(PetscObjectComm((PetscObject)A), &size));
2592   PetscCallMPI(MPI_Comm_rank(PetscObjectComm((PetscObject)A), &rank));
2593 
2594   bs  = A->rmap->bs;
2595   mbs = a->mbs;
2596   Mbs = a->Mbs;
2597   ba  = b->a;
2598   bi  = b->i;
2599   bj  = b->j;
2600 
2601   /* find ownerships */
2602   rowners_bs = A->rmap->range;
2603 
2604   /* each proc creates an array to be distributed */
2605   PetscCall(PetscCalloc1(bs * Mbs, &work));
2606 
2607   /* row_max for B */
2608   if (rank != size - 1) {
2609     for (i = 0; i < mbs; i++) {
2610       ncols = bi[1] - bi[0];
2611       bi++;
2612       brow = bs * i;
2613       for (j = 0; j < ncols; j++) {
2614         bcol = bs * (*bj);
2615         for (kcol = 0; kcol < bs; kcol++) {
2616           col = bcol + kcol;           /* local col index */
2617           col += rowners_bs[rank + 1]; /* global col index */
2618           for (krow = 0; krow < bs; krow++) {
2619             atmp = PetscAbsScalar(*ba);
2620             ba++;
2621             row = brow + krow; /* local row index */
2622             if (PetscRealPart(va[row]) < atmp) va[row] = atmp;
2623             if (work[col] < atmp) work[col] = atmp;
2624           }
2625         }
2626         bj++;
2627       }
2628     }
2629 
2630     /* send values to its owners */
2631     for (PetscMPIInt dest = rank + 1; dest < size; dest++) {
2632       svalues = work + rowners_bs[dest];
2633       count   = rowners_bs[dest + 1] - rowners_bs[dest];
2634       PetscCallMPI(MPIU_Send(svalues, count, MPIU_REAL, dest, rank, PetscObjectComm((PetscObject)A)));
2635     }
2636   }
2637 
2638   /* receive values */
2639   if (rank) {
2640     rvalues = work;
2641     count   = rowners_bs[rank + 1] - rowners_bs[rank];
2642     for (source = 0; source < rank; source++) {
2643       PetscCallMPI(MPIU_Recv(rvalues, count, MPIU_REAL, MPI_ANY_SOURCE, MPI_ANY_TAG, PetscObjectComm((PetscObject)A), &stat));
2644       /* process values */
2645       for (i = 0; i < count; i++) {
2646         if (PetscRealPart(va[i]) < rvalues[i]) va[i] = rvalues[i];
2647       }
2648     }
2649   }
2650 
2651   PetscCall(VecRestoreArray(v, &va));
2652   PetscCall(PetscFree(work));
2653   PetscFunctionReturn(PETSC_SUCCESS);
2654 }
2655 
2656 static PetscErrorCode MatSOR_MPISBAIJ(Mat matin, Vec bb, PetscReal omega, MatSORType flag, PetscReal fshift, PetscInt its, PetscInt lits, Vec xx)
2657 {
2658   Mat_MPISBAIJ      *mat = (Mat_MPISBAIJ *)matin->data;
2659   PetscInt           mbs = mat->mbs, bs = matin->rmap->bs;
2660   PetscScalar       *x, *ptr, *from;
2661   Vec                bb1;
2662   const PetscScalar *b;
2663 
2664   PetscFunctionBegin;
2665   PetscCheck(its > 0 && lits > 0, PETSC_COMM_SELF, PETSC_ERR_ARG_WRONG, "Relaxation requires global its %" PetscInt_FMT " and local its %" PetscInt_FMT " both positive", its, lits);
2666   PetscCheck(bs <= 1, PETSC_COMM_SELF, PETSC_ERR_SUP, "SSOR for block size > 1 is not yet implemented");
2667 
2668   if (flag == SOR_APPLY_UPPER) {
2669     PetscCall((*mat->A->ops->sor)(mat->A, bb, omega, flag, fshift, lits, 1, xx));
2670     PetscFunctionReturn(PETSC_SUCCESS);
2671   }
2672 
2673   if ((flag & SOR_LOCAL_SYMMETRIC_SWEEP) == SOR_LOCAL_SYMMETRIC_SWEEP) {
2674     if (flag & SOR_ZERO_INITIAL_GUESS) {
2675       PetscCall((*mat->A->ops->sor)(mat->A, bb, omega, flag, fshift, lits, lits, xx));
2676       its--;
2677     }
2678 
2679     PetscCall(VecDuplicate(bb, &bb1));
2680     while (its--) {
2681       /* lower triangular part: slvec0b = - B^T*xx */
2682       PetscCall((*mat->B->ops->multtranspose)(mat->B, xx, mat->slvec0b));
2683 
2684       /* copy xx into slvec0a */
2685       PetscCall(VecGetArray(mat->slvec0, &ptr));
2686       PetscCall(VecGetArray(xx, &x));
2687       PetscCall(PetscArraycpy(ptr, x, bs * mbs));
2688       PetscCall(VecRestoreArray(mat->slvec0, &ptr));
2689 
2690       PetscCall(VecScale(mat->slvec0, -1.0));
2691 
2692       /* copy bb into slvec1a */
2693       PetscCall(VecGetArray(mat->slvec1, &ptr));
2694       PetscCall(VecGetArrayRead(bb, &b));
2695       PetscCall(PetscArraycpy(ptr, b, bs * mbs));
2696       PetscCall(VecRestoreArray(mat->slvec1, &ptr));
2697 
2698       /* set slvec1b = 0 */
2699       PetscCall(PetscObjectStateIncrease((PetscObject)mat->slvec1b));
2700       PetscCall(VecZeroEntries(mat->slvec1b));
2701 
2702       PetscCall(VecScatterBegin(mat->sMvctx, mat->slvec0, mat->slvec1, ADD_VALUES, SCATTER_FORWARD));
2703       PetscCall(VecRestoreArray(xx, &x));
2704       PetscCall(VecRestoreArrayRead(bb, &b));
2705       PetscCall(VecScatterEnd(mat->sMvctx, mat->slvec0, mat->slvec1, ADD_VALUES, SCATTER_FORWARD));
2706 
2707       /* upper triangular part: bb1 = bb1 - B*x */
2708       PetscCall((*mat->B->ops->multadd)(mat->B, mat->slvec1b, mat->slvec1a, bb1));
2709 
2710       /* local diagonal sweep */
2711       PetscCall((*mat->A->ops->sor)(mat->A, bb1, omega, SOR_SYMMETRIC_SWEEP, fshift, lits, lits, xx));
2712     }
2713     PetscCall(VecDestroy(&bb1));
2714   } else if ((flag & SOR_LOCAL_FORWARD_SWEEP) && (its == 1) && (flag & SOR_ZERO_INITIAL_GUESS)) {
2715     PetscCall((*mat->A->ops->sor)(mat->A, bb, omega, flag, fshift, lits, 1, xx));
2716   } else if ((flag & SOR_LOCAL_BACKWARD_SWEEP) && (its == 1) && (flag & SOR_ZERO_INITIAL_GUESS)) {
2717     PetscCall((*mat->A->ops->sor)(mat->A, bb, omega, flag, fshift, lits, 1, xx));
2718   } else if (flag & SOR_EISENSTAT) {
2719     Vec                xx1;
2720     PetscBool          hasop;
2721     const PetscScalar *diag;
2722     PetscScalar       *sl, scale = (omega - 2.0) / omega;
2723     PetscInt           i, n;
2724 
2725     if (!mat->xx1) {
2726       PetscCall(VecDuplicate(bb, &mat->xx1));
2727       PetscCall(VecDuplicate(bb, &mat->bb1));
2728     }
2729     xx1 = mat->xx1;
2730     bb1 = mat->bb1;
2731 
2732     PetscCall((*mat->A->ops->sor)(mat->A, bb, omega, (MatSORType)(SOR_ZERO_INITIAL_GUESS | SOR_LOCAL_BACKWARD_SWEEP), fshift, lits, 1, xx));
2733 
2734     if (!mat->diag) {
2735       /* this is wrong for same matrix with new nonzero values */
2736       PetscCall(MatCreateVecs(matin, &mat->diag, NULL));
2737       PetscCall(MatGetDiagonal(matin, mat->diag));
2738     }
2739     PetscCall(MatHasOperation(matin, MATOP_MULT_DIAGONAL_BLOCK, &hasop));
2740 
2741     if (hasop) {
2742       PetscCall(MatMultDiagonalBlock(matin, xx, bb1));
2743       PetscCall(VecAYPX(mat->slvec1a, scale, bb));
2744     } else {
2745       /*
2746           These two lines are replaced by code that may be a bit faster for a good compiler
2747       PetscCall(VecPointwiseMult(mat->slvec1a,mat->diag,xx));
2748       PetscCall(VecAYPX(mat->slvec1a,scale,bb));
2749       */
2750       PetscCall(VecGetArray(mat->slvec1a, &sl));
2751       PetscCall(VecGetArrayRead(mat->diag, &diag));
2752       PetscCall(VecGetArrayRead(bb, &b));
2753       PetscCall(VecGetArray(xx, &x));
2754       PetscCall(VecGetLocalSize(xx, &n));
2755       if (omega == 1.0) {
2756         for (i = 0; i < n; i++) sl[i] = b[i] - diag[i] * x[i];
2757         PetscCall(PetscLogFlops(2.0 * n));
2758       } else {
2759         for (i = 0; i < n; i++) sl[i] = b[i] + scale * diag[i] * x[i];
2760         PetscCall(PetscLogFlops(3.0 * n));
2761       }
2762       PetscCall(VecRestoreArray(mat->slvec1a, &sl));
2763       PetscCall(VecRestoreArrayRead(mat->diag, &diag));
2764       PetscCall(VecRestoreArrayRead(bb, &b));
2765       PetscCall(VecRestoreArray(xx, &x));
2766     }
2767 
2768     /* multiply off-diagonal portion of matrix */
2769     PetscCall(PetscObjectStateIncrease((PetscObject)mat->slvec1b));
2770     PetscCall(VecZeroEntries(mat->slvec1b));
2771     PetscCall((*mat->B->ops->multtranspose)(mat->B, xx, mat->slvec0b));
2772     PetscCall(VecGetArray(mat->slvec0, &from));
2773     PetscCall(VecGetArray(xx, &x));
2774     PetscCall(PetscArraycpy(from, x, bs * mbs));
2775     PetscCall(VecRestoreArray(mat->slvec0, &from));
2776     PetscCall(VecRestoreArray(xx, &x));
2777     PetscCall(VecScatterBegin(mat->sMvctx, mat->slvec0, mat->slvec1, ADD_VALUES, SCATTER_FORWARD));
2778     PetscCall(VecScatterEnd(mat->sMvctx, mat->slvec0, mat->slvec1, ADD_VALUES, SCATTER_FORWARD));
2779     PetscCall((*mat->B->ops->multadd)(mat->B, mat->slvec1b, mat->slvec1a, mat->slvec1a));
2780 
2781     /* local sweep */
2782     PetscCall((*mat->A->ops->sor)(mat->A, mat->slvec1a, omega, (MatSORType)(SOR_ZERO_INITIAL_GUESS | SOR_LOCAL_FORWARD_SWEEP), fshift, lits, 1, xx1));
2783     PetscCall(VecAXPY(xx, 1.0, xx1));
2784   } else SETERRQ(PETSC_COMM_SELF, PETSC_ERR_SUP, "MatSORType is not supported for SBAIJ matrix format");
2785   PetscFunctionReturn(PETSC_SUCCESS);
2786 }
2787 
2788 /*@
2789   MatCreateMPISBAIJWithArrays - creates a `MATMPISBAIJ` matrix using arrays that contain in standard CSR format for the local rows.
2790 
2791   Collective
2792 
2793   Input Parameters:
2794 + comm - MPI communicator
2795 . bs   - the block size, only a block size of 1 is supported
2796 . m    - number of local rows (Cannot be `PETSC_DECIDE`)
2797 . n    - This value should be the same as the local size used in creating the
2798          x vector for the matrix-vector product $ y = Ax $. (or `PETSC_DECIDE` to have
2799          calculated if `N` is given) For square matrices `n` is almost always `m`.
2800 . M    - number of global rows (or `PETSC_DETERMINE` to have calculated if `m` is given)
2801 . N    - number of global columns (or `PETSC_DETERMINE` to have calculated if `n` is given)
2802 . i    - row indices; that is i[0] = 0, i[row] = i[row-1] + number of block elements in that row block row of the matrix
2803 . j    - column indices
2804 - a    - matrix values
2805 
2806   Output Parameter:
2807 . mat - the matrix
2808 
2809   Level: intermediate
2810 
2811   Notes:
2812   The `i`, `j`, and `a` arrays ARE copied by this routine into the internal format used by PETSc;
2813   thus you CANNOT change the matrix entries by changing the values of `a` after you have
2814   called this routine. Use `MatCreateMPIAIJWithSplitArrays()` to avoid needing to copy the arrays.
2815 
2816   The `i` and `j` indices are 0 based, and `i` indices are indices corresponding to the local `j` array.
2817 
2818 .seealso: [](ch_matrices), `Mat`, `MATMPISBAIJ`, `MatCreate()`, `MatCreateSeqAIJ()`, `MatSetValues()`, `MatMPIAIJSetPreallocation()`, `MatMPIAIJSetPreallocationCSR()`,
2819           `MATMPIAIJ`, `MatCreateAIJ()`, `MatCreateMPIAIJWithSplitArrays()`, `MatMPISBAIJSetPreallocationCSR()`
2820 @*/
2821 PetscErrorCode MatCreateMPISBAIJWithArrays(MPI_Comm comm, PetscInt bs, PetscInt m, PetscInt n, PetscInt M, PetscInt N, const PetscInt i[], const PetscInt j[], const PetscScalar a[], Mat *mat)
2822 {
2823   PetscFunctionBegin;
2824   PetscCheck(!i[0], PETSC_COMM_SELF, PETSC_ERR_ARG_OUTOFRANGE, "i (row indices) must start with 0");
2825   PetscCheck(m >= 0, PETSC_COMM_SELF, PETSC_ERR_ARG_OUTOFRANGE, "local number of rows (m) cannot be PETSC_DECIDE, or negative");
2826   PetscCall(MatCreate(comm, mat));
2827   PetscCall(MatSetSizes(*mat, m, n, M, N));
2828   PetscCall(MatSetType(*mat, MATMPISBAIJ));
2829   PetscCall(MatMPISBAIJSetPreallocationCSR(*mat, bs, i, j, a));
2830   PetscFunctionReturn(PETSC_SUCCESS);
2831 }
2832 
2833 /*@
2834   MatMPISBAIJSetPreallocationCSR - Creates a sparse parallel matrix in `MATMPISBAIJ` format using the given nonzero structure and (optional) numerical values
2835 
2836   Collective
2837 
2838   Input Parameters:
2839 + B  - the matrix
2840 . bs - the block size
2841 . i  - the indices into `j` for the start of each local row (indices start with zero)
2842 . j  - the column indices for each local row (indices start with zero) these must be sorted for each row
2843 - v  - optional values in the matrix, pass `NULL` if not provided
2844 
2845   Level: advanced
2846 
2847   Notes:
2848   The `i`, `j`, and `v` arrays ARE copied by this routine into the internal format used by PETSc;
2849   thus you CANNOT change the matrix entries by changing the values of `v` after you have
2850   called this routine.
2851 
2852   Though this routine has Preallocation() in the name it also sets the exact nonzero locations of the matrix entries
2853   and usually the numerical values as well
2854 
2855   Any entries passed in that are below the diagonal are ignored
2856 
2857 .seealso: [](ch_matrices), `Mat`, `MATMPISBAIJ`, `MatCreate()`, `MatCreateSeqAIJ()`, `MatSetValues()`, `MatMPIBAIJSetPreallocation()`, `MatCreateAIJ()`, `MATMPIAIJ`,
2858           `MatCreateMPISBAIJWithArrays()`
2859 @*/
2860 PetscErrorCode MatMPISBAIJSetPreallocationCSR(Mat B, PetscInt bs, const PetscInt i[], const PetscInt j[], const PetscScalar v[])
2861 {
2862   PetscFunctionBegin;
2863   PetscTryMethod(B, "MatMPISBAIJSetPreallocationCSR_C", (Mat, PetscInt, const PetscInt[], const PetscInt[], const PetscScalar[]), (B, bs, i, j, v));
2864   PetscFunctionReturn(PETSC_SUCCESS);
2865 }
2866 
2867 PetscErrorCode MatCreateMPIMatConcatenateSeqMat_MPISBAIJ(MPI_Comm comm, Mat inmat, PetscInt n, MatReuse scall, Mat *outmat)
2868 {
2869   PetscInt     m, N, i, rstart, nnz, Ii, bs, cbs;
2870   PetscInt    *indx;
2871   PetscScalar *values;
2872 
2873   PetscFunctionBegin;
2874   PetscCall(MatGetSize(inmat, &m, &N));
2875   if (scall == MAT_INITIAL_MATRIX) { /* symbolic phase */
2876     Mat_SeqSBAIJ *a = (Mat_SeqSBAIJ *)inmat->data;
2877     PetscInt     *dnz, *onz, mbs, Nbs, nbs;
2878     PetscInt     *bindx, rmax = a->rmax, j;
2879     PetscMPIInt   rank, size;
2880 
2881     PetscCall(MatGetBlockSizes(inmat, &bs, &cbs));
2882     mbs = m / bs;
2883     Nbs = N / cbs;
2884     if (n == PETSC_DECIDE) PetscCall(PetscSplitOwnershipBlock(comm, cbs, &n, &N));
2885     nbs = n / cbs;
2886 
2887     PetscCall(PetscMalloc1(rmax, &bindx));
2888     MatPreallocateBegin(comm, mbs, nbs, dnz, onz); /* inline function, output __end and __rstart are used below */
2889 
2890     PetscCallMPI(MPI_Comm_rank(comm, &rank));
2891     PetscCallMPI(MPI_Comm_rank(comm, &size));
2892     if (rank == size - 1) {
2893       /* Check sum(nbs) = Nbs */
2894       PetscCheck(__end == Nbs, PETSC_COMM_SELF, PETSC_ERR_ARG_INCOMP, "Sum of local block columns %" PetscInt_FMT " != global block columns %" PetscInt_FMT, __end, Nbs);
2895     }
2896 
2897     rstart = __rstart; /* block rstart of *outmat; see inline function MatPreallocateBegin */
2898     PetscCall(MatSetOption(inmat, MAT_GETROW_UPPERTRIANGULAR, PETSC_TRUE));
2899     for (i = 0; i < mbs; i++) {
2900       PetscCall(MatGetRow_SeqSBAIJ(inmat, i * bs, &nnz, &indx, NULL)); /* non-blocked nnz and indx */
2901       nnz = nnz / bs;
2902       for (j = 0; j < nnz; j++) bindx[j] = indx[j * bs] / bs;
2903       PetscCall(MatPreallocateSet(i + rstart, nnz, bindx, dnz, onz));
2904       PetscCall(MatRestoreRow_SeqSBAIJ(inmat, i * bs, &nnz, &indx, NULL));
2905     }
2906     PetscCall(MatSetOption(inmat, MAT_GETROW_UPPERTRIANGULAR, PETSC_FALSE));
2907     PetscCall(PetscFree(bindx));
2908 
2909     PetscCall(MatCreate(comm, outmat));
2910     PetscCall(MatSetSizes(*outmat, m, n, PETSC_DETERMINE, PETSC_DETERMINE));
2911     PetscCall(MatSetBlockSizes(*outmat, bs, cbs));
2912     PetscCall(MatSetType(*outmat, MATSBAIJ));
2913     PetscCall(MatSeqSBAIJSetPreallocation(*outmat, bs, 0, dnz));
2914     PetscCall(MatMPISBAIJSetPreallocation(*outmat, bs, 0, dnz, 0, onz));
2915     MatPreallocateEnd(dnz, onz);
2916   }
2917 
2918   /* numeric phase */
2919   PetscCall(MatGetBlockSizes(inmat, &bs, &cbs));
2920   PetscCall(MatGetOwnershipRange(*outmat, &rstart, NULL));
2921 
2922   PetscCall(MatSetOption(inmat, MAT_GETROW_UPPERTRIANGULAR, PETSC_TRUE));
2923   for (i = 0; i < m; i++) {
2924     PetscCall(MatGetRow_SeqSBAIJ(inmat, i, &nnz, &indx, &values));
2925     Ii = i + rstart;
2926     PetscCall(MatSetValues(*outmat, 1, &Ii, nnz, indx, values, INSERT_VALUES));
2927     PetscCall(MatRestoreRow_SeqSBAIJ(inmat, i, &nnz, &indx, &values));
2928   }
2929   PetscCall(MatSetOption(inmat, MAT_GETROW_UPPERTRIANGULAR, PETSC_FALSE));
2930   PetscCall(MatAssemblyBegin(*outmat, MAT_FINAL_ASSEMBLY));
2931   PetscCall(MatAssemblyEnd(*outmat, MAT_FINAL_ASSEMBLY));
2932   PetscFunctionReturn(PETSC_SUCCESS);
2933 }
2934