xref: /petsc/src/mat/impls/elemental/matelem.cxx (revision 4fe7bbcaf558b286201a0623a1c750d8657ebdd1)
1 #include <../src/mat/impls/elemental/matelemimpl.h> /*I "petscmat.h" I*/
2 
3 /*
4     The variable Petsc_Elemental_keyval is used to indicate an MPI attribute that
5   is attached to a communicator, in this case the attribute is a Mat_Elemental_Grid
6 */
7 static PetscMPIInt Petsc_Elemental_keyval = MPI_KEYVAL_INVALID;
8 
9 #undef __FUNCT__
10 #define __FUNCT__ "PetscElementalInitializePackage"
11 /*@C
12    PetscElementalInitializePackage - Initialize Elemental package
13 
14    Logically Collective
15 
16    Input Arguments:
17 .  path - the dynamic library path or PETSC_NULL
18 
19    Level: developer
20 
21 .seealso: MATELEMENTAL, PetscElementalFinalizePackage()
22 @*/
23 PetscErrorCode PetscElementalInitializePackage(const char *path)
24 {
25   PetscErrorCode ierr;
26 
27   PetscFunctionBegin;
28   if (elem::Initialized()) PetscFunctionReturn(0);
29   { /* We have already initialized MPI, so this song and dance is just to pass these variables (which won't be used by Elemental) through the interface that needs references */
30     int zero = 0;
31     char **nothing = 0;
32     elem::Initialize(zero,nothing);
33   }
34   ierr = PetscRegisterFinalize(PetscElementalFinalizePackage);CHKERRQ(ierr);
35   PetscFunctionReturn(0);
36 }
37 
38 #undef __FUNCT__
39 #define __FUNCT__ "PetscElementalFinalizePackage"
40 /*@C
41    PetscElementalFinalizePackage - Finalize Elemental package
42 
43    Logically Collective
44 
45    Level: developer
46 
47 .seealso: MATELEMENTAL, PetscElementalInitializePackage()
48 @*/
49 PetscErrorCode PetscElementalFinalizePackage(void)
50 {
51 
52   PetscFunctionBegin;
53   elem::Finalize();
54   PetscFunctionReturn(0);
55 }
56 
57 #undef __FUNCT__
58 #define __FUNCT__ "MatView_Elemental"
59 static PetscErrorCode MatView_Elemental(Mat A,PetscViewer viewer)
60 {
61   PetscErrorCode ierr;
62   Mat_Elemental  *a = (Mat_Elemental*)A->data;
63   PetscBool      iascii;
64 
65   PetscFunctionBegin;
66   ierr = PetscObjectTypeCompare((PetscObject)viewer,PETSCVIEWERASCII,&iascii);CHKERRQ(ierr);
67   if (iascii) {
68     PetscViewerFormat format;
69     ierr = PetscViewerGetFormat(viewer,&format);CHKERRQ(ierr);
70     if (format == PETSC_VIEWER_ASCII_INFO) {
71       if (format == PETSC_VIEWER_ASCII_FACTOR_INFO){
72         SETERRQ(((PetscObject)viewer)->comm,PETSC_ERR_SUP,"FactorInfo viewer not implemented yet");
73         /* ierr = MatInfo_Elemental(A,viewer);CHKERRQ(ierr); */
74       } else {
75         SETERRQ(((PetscObject)viewer)->comm,PETSC_ERR_SUP,"Info viewer not implemented yet");
76       }
77     } else if (format == PETSC_VIEWER_DEFAULT) {
78       Mat Aaij;
79       ierr = PetscViewerASCIIUseTabs(viewer,PETSC_FALSE);CHKERRQ(ierr);
80       ierr = PetscObjectPrintClassNamePrefixType((PetscObject)A,viewer,"Matrix Object");CHKERRQ(ierr);
81       a->emat->Print("Elemental matrix (cyclic ordering)");
82       ierr = PetscViewerASCIIUseTabs(viewer,PETSC_TRUE);CHKERRQ(ierr);
83       if (A->factortype == MAT_FACTOR_NONE){
84         ierr = PetscPrintf(((PetscObject)viewer)->comm,"Elemental matrix (explicit ordering)\n");CHKERRQ(ierr);
85         ierr = MatComputeExplicitOperator(A,&Aaij);CHKERRQ(ierr);
86         ierr = MatView(Aaij,PETSC_VIEWER_STDOUT_WORLD);CHKERRQ(ierr);
87         ierr = MatDestroy(&Aaij);CHKERRQ(ierr);
88       }
89     } else SETERRQ(((PetscObject)viewer)->comm,PETSC_ERR_SUP,"Format");
90   } else SETERRQ1(PETSC_COMM_SELF,PETSC_ERR_SUP,"Viewer type %s not supported by Elemental matrices",((PetscObject)viewer)->type_name);
91   PetscFunctionReturn(0);
92 }
93 
94 #undef __FUNCT__
95 #define __FUNCT__ "MatSetValues_Elemental"
96 static PetscErrorCode MatSetValues_Elemental(Mat A,PetscInt nr,const PetscInt *rows,PetscInt nc,const PetscInt *cols,const PetscScalar *vals,InsertMode imode)
97 {
98   PetscErrorCode ierr;
99   Mat_Elemental  *a = (Mat_Elemental*)A->data;
100   PetscMPIInt    rank;
101   PetscInt       i,j,rrank,ridx,crank,cidx;
102 
103   PetscFunctionBegin;
104   ierr = MPI_Comm_rank(((PetscObject)A)->comm,&rank);CHKERRQ(ierr);
105 
106   const elem::Grid &grid = a->emat->Grid();
107   for (i=0; i<nr; i++) {
108     PetscInt erow,ecol,elrow,elcol;
109     if (rows[i] < 0) continue;
110     P2RO(A,0,rows[i],&rrank,&ridx);
111     RO2E(A,0,rrank,ridx,&erow);
112     if (rrank < 0 || ridx < 0 || erow < 0) SETERRQ(((PetscObject)A)->comm,PETSC_ERR_PLIB,"Incorrect row translation");
113     for (j=0; j<nc; j++) {
114       if (cols[j] < 0) continue;
115       P2RO(A,1,cols[j],&crank,&cidx);
116       RO2E(A,1,crank,cidx,&ecol);
117       if (crank < 0 || cidx < 0 || ecol < 0) SETERRQ(((PetscObject)A)->comm,PETSC_ERR_PLIB,"Incorrect col translation");
118       if (erow % grid.MCSize() != grid.MCRank() || ecol % grid.MRSize() != grid.MRRank()){ /* off-proc entry */
119         if (imode != ADD_VALUES) SETERRQ(PETSC_COMM_SELF,PETSC_ERR_SUP,"Only ADD_VALUES to off-processor entry is supported");
120         /* PetscPrintf(PETSC_COMM_SELF,"[%D] add off-proc entry (%D,%D, %g) (%D %D)\n",rank,rows[i],cols[j],*(vals+i*nc),erow,ecol); */
121         a->esubmat->Set(0,0, vals[i*nc+j]);
122         a->interface->Axpy(1.0,*(a->esubmat),erow,ecol);
123         continue;
124       }
125       elrow = erow / grid.MCSize();
126       elcol = ecol / grid.MRSize();
127       switch (imode) {
128       case INSERT_VALUES: a->emat->SetLocal(elrow,elcol,vals[i*nc+j]); break;
129       case ADD_VALUES: a->emat->UpdateLocal(elrow,elcol,vals[i*nc+j]); break;
130       default: SETERRQ1(((PetscObject)A)->comm,PETSC_ERR_SUP,"No support for InsertMode %d",(int)imode);
131       }
132     }
133   }
134   PetscFunctionReturn(0);
135 }
136 
137 #undef __FUNCT__
138 #define __FUNCT__ "MatMult_Elemental"
139 static PetscErrorCode MatMult_Elemental(Mat A,Vec X,Vec Y)
140 {
141   Mat_Elemental     *a = (Mat_Elemental*)A->data;
142   PetscErrorCode    ierr;
143   const PetscScalar *x;
144   PetscScalar       *y;
145   PetscScalar       one = 1,zero = 0;
146 
147   PetscFunctionBegin;
148   ierr = VecGetArrayRead(X,&x);CHKERRQ(ierr);
149   ierr = VecGetArray(Y,&y);CHKERRQ(ierr);
150   { /* Scoping so that constructor is called before pointer is returned */
151     elem::DistMatrix<PetscScalar,elem::VC,elem::STAR> xe(A->cmap->N,1,0,x,A->cmap->n,*a->grid);
152     elem::DistMatrix<PetscScalar,elem::VC,elem::STAR> ye(A->rmap->N,1,0,y,A->rmap->n,*a->grid);
153     elem::Gemv(elem::NORMAL,one,*a->emat,xe,zero,ye);
154   }
155   ierr = VecRestoreArrayRead(X,&x);CHKERRQ(ierr);
156   ierr = VecRestoreArray(Y,&y);CHKERRQ(ierr);
157   PetscFunctionReturn(0);
158 }
159 
160 #undef __FUNCT__
161 #define __FUNCT__ "MatMultAdd_Elemental"
162 static PetscErrorCode MatMultAdd_Elemental(Mat A,Vec X,Vec Y,Vec Z)
163 {
164   Mat_Elemental     *a = (Mat_Elemental*)A->data;
165   PetscErrorCode    ierr;
166   const PetscScalar *x;
167   PetscScalar       *z;
168   PetscScalar       one = 1.0;
169 
170   PetscFunctionBegin;
171   if (Y != Z) {ierr = VecCopy(Y,Z);CHKERRQ(ierr);}
172   ierr = VecGetArrayRead(X,&x);CHKERRQ(ierr);
173   ierr = VecGetArray(Z,&z);CHKERRQ(ierr);
174   { /* Scoping so that constructor is called before pointer is returned */
175     elem::DistMatrix<PetscScalar,elem::VC,elem::STAR> xe(A->cmap->N,1,0,x,A->cmap->n,*a->grid);
176     elem::DistMatrix<PetscScalar,elem::VC,elem::STAR> ze(A->rmap->N,1,0,z,A->rmap->n,*a->grid);
177     elem::Gemv(elem::NORMAL,one,*a->emat,xe,one,ze);
178   }
179   ierr = VecRestoreArrayRead(X,&x);CHKERRQ(ierr);
180   ierr = VecRestoreArray(Z,&z);CHKERRQ(ierr);
181   PetscFunctionReturn(0);
182 }
183 
184 #undef __FUNCT__
185 #define __FUNCT__ "MatMatMultNumeric_Elemental"
186 static PetscErrorCode MatMatMultNumeric_Elemental(Mat A,Mat B,Mat C)
187 {
188   Mat_Elemental  *a = (Mat_Elemental*)A->data;
189   Mat_Elemental  *b = (Mat_Elemental*)B->data;
190   Mat_Elemental  *c = (Mat_Elemental*)C->data;
191   PetscScalar    one = 1.0,zero = 0.0;
192 
193   PetscFunctionBegin;
194   { /* Scoping so that constructor is called before pointer is returned */
195     elem::Gemm(elem::NORMAL,elem::NORMAL,one,*a->emat,*b->emat,zero,*c->emat);
196   }
197   C->assembled = PETSC_TRUE;
198   PetscFunctionReturn(0);
199 }
200 
201 #undef __FUNCT__
202 #define __FUNCT__ "MatMatMultSymbolic_Elemental"
203 static PetscErrorCode MatMatMultSymbolic_Elemental(Mat A,Mat B,PetscReal fill,Mat *C)
204 {
205   PetscErrorCode ierr;
206   Mat            Ce;
207   MPI_Comm       comm=((PetscObject)A)->comm;
208 
209   PetscFunctionBegin;
210   ierr = MatCreate(comm,&Ce);CHKERRQ(ierr);
211   ierr = MatSetSizes(Ce,A->rmap->n,B->cmap->n,PETSC_DECIDE,PETSC_DECIDE);CHKERRQ(ierr);
212   ierr = MatSetType(Ce,MATELEMENTAL);CHKERRQ(ierr);
213   ierr = MatSetUp(Ce);CHKERRQ(ierr);
214   *C = Ce;
215   PetscFunctionReturn(0);
216 }
217 
218 #undef __FUNCT__
219 #define __FUNCT__ "MatMatMult_Elemental"
220 static PetscErrorCode MatMatMult_Elemental(Mat A,Mat B,MatReuse scall,PetscReal fill,Mat *C)
221 {
222   PetscErrorCode ierr;
223 
224   PetscFunctionBegin;
225   if (scall == MAT_INITIAL_MATRIX){
226     ierr = PetscLogEventBegin(MAT_MatMultSymbolic,A,B,0,0);CHKERRQ(ierr);
227     ierr = MatMatMultSymbolic_Elemental(A,B,1.0,C);CHKERRQ(ierr);
228     ierr = PetscLogEventEnd(MAT_MatMultSymbolic,A,B,0,0);CHKERRQ(ierr);
229   }
230   ierr = PetscLogEventBegin(MAT_MatMultNumeric,A,B,0,0);CHKERRQ(ierr);
231   ierr = MatMatMultNumeric_Elemental(A,B,*C);CHKERRQ(ierr);
232   ierr = PetscLogEventEnd(MAT_MatMultNumeric,A,B,0,0);CHKERRQ(ierr);
233   PetscFunctionReturn(0);
234 }
235 
236 #undef __FUNCT__
237 #define __FUNCT__ "MatScale_Elemental"
238 static PetscErrorCode MatScale_Elemental(Mat X,PetscScalar a)
239 {
240   Mat_Elemental  *x = (Mat_Elemental*)X->data;
241 
242   PetscFunctionBegin;
243   elem::Scal(a,*x->emat);
244   PetscFunctionReturn(0);
245 }
246 
247 #undef __FUNCT__
248 #define __FUNCT__ "MatAXPY_Elemental"
249 static PetscErrorCode MatAXPY_Elemental(Mat Y,PetscScalar a,Mat X,MatStructure str)
250 {
251   Mat_Elemental  *x = (Mat_Elemental*)X->data;
252   Mat_Elemental  *y = (Mat_Elemental*)Y->data;
253 
254   PetscFunctionBegin;
255   elem::Axpy(a,*x->emat,*y->emat);
256   PetscFunctionReturn(0);
257 }
258 
259 #undef __FUNCT__
260 #define __FUNCT__ "MatCopy_Elemental"
261 static PetscErrorCode MatCopy_Elemental(Mat A,Mat B,MatStructure str)
262 {
263   Mat_Elemental *a=(Mat_Elemental*)A->data;
264   Mat_Elemental *b=(Mat_Elemental*)B->data;
265 
266   PetscFunctionBegin;
267   elem::Copy(*a->emat,*b->emat);
268   PetscFunctionReturn(0);
269 }
270 
271 #undef __FUNCT__
272 #define __FUNCT__ "MatTranspose_Elemental"
273 static PetscErrorCode MatTranspose_Elemental(Mat A,MatReuse reuse,Mat *B)
274 {
275   /* Only out-of-place supported */
276   Mat            Be;
277   PetscErrorCode ierr;
278   MPI_Comm       comm=((PetscObject)A)->comm;
279   Mat_Elemental  *a = (Mat_Elemental*)A->data, *b;
280 
281   PetscFunctionBegin;
282   if (reuse == MAT_INITIAL_MATRIX){
283     ierr = MatCreate(comm,&Be);CHKERRQ(ierr);
284     ierr = MatSetSizes(Be,A->cmap->n,A->rmap->n,PETSC_DECIDE,PETSC_DECIDE);CHKERRQ(ierr);
285     ierr = MatSetType(Be,MATELEMENTAL);CHKERRQ(ierr);
286     ierr = MatSetUp(Be);CHKERRQ(ierr);
287     *B = Be;
288   }
289   b = (Mat_Elemental*)Be->data;
290   elem::Transpose(*a->emat,*b->emat);
291   Be->assembled = PETSC_TRUE;
292   PetscFunctionReturn(0);
293 }
294 
295 #undef __FUNCT__
296 #define __FUNCT__ "MatSolve_Elemental"
297 static PetscErrorCode MatSolve_Elemental(Mat A,Vec B,Vec X)
298 {
299   Mat_Elemental     *a = (Mat_Elemental*)A->data;
300   PetscErrorCode    ierr;
301   PetscScalar       *x;
302 
303   PetscFunctionBegin;
304   ierr = VecCopy(B,X);CHKERRQ(ierr);
305   ierr = VecGetArray(X,&x);CHKERRQ(ierr);
306   elem::DistMatrix<PetscScalar,elem::VC,elem::STAR> xe(A->rmap->N,1,0,x,A->rmap->n,*a->grid);
307   elem::DistMatrix<PetscScalar,elem::MC,elem::MR> xer = xe;
308   switch (A->factortype) {
309   case MAT_FACTOR_LU:
310     if ((*a->pivot).AllocatedMemory()) {
311       elem::SolveAfterLU(elem::NORMAL,*a->emat,*a->pivot,xer);
312       elem::Copy(xer,xe);
313     } else {
314       elem::SolveAfterLU(elem::NORMAL,*a->emat,xer);
315       elem::Copy(xer,xe);
316     }
317     break;
318   case MAT_FACTOR_CHOLESKY:
319     elem::SolveAfterCholesky(elem::UPPER,elem::NORMAL,*a->emat,xer);
320     elem::Copy(xer,xe);
321     break;
322   default:
323     SETERRQ(PETSC_COMM_SELF,PETSC_ERR_SUP,"Unfactored Matrix or Unsupported MatFactorType");
324     break;
325   }
326   ierr = VecRestoreArray(X,&x);CHKERRQ(ierr);
327   PetscFunctionReturn(0);
328 }
329 
330 #undef __FUNCT__
331 #define __FUNCT__ "MatMatSolve_Elemental"
332 static PetscErrorCode MatMatSolve_Elemental(Mat A,Mat B,Mat X)
333 {
334   Mat_Elemental *a=(Mat_Elemental*)A->data;
335   Mat_Elemental *b=(Mat_Elemental*)B->data;
336   Mat_Elemental *x=(Mat_Elemental*)X->data;
337 
338   PetscFunctionBegin;
339   elem::Copy(*b->emat,*x->emat);
340   switch (A->factortype) {
341   case MAT_FACTOR_LU:
342     if ((*a->pivot).AllocatedMemory()) {
343       elem::SolveAfterLU(elem::NORMAL,*a->emat,*a->pivot,*x->emat);
344     } else {
345       elem::SolveAfterLU(elem::NORMAL,*a->emat,*x->emat);
346     }
347     break;
348   case MAT_FACTOR_CHOLESKY:
349     elem::SolveAfterCholesky(elem::UPPER,elem::NORMAL,*a->emat,*x->emat);
350     break;
351   default:
352     SETERRQ(PETSC_COMM_SELF,PETSC_ERR_SUP,"Unfactored Matrix or Unsupported MatFactorType");
353     break;
354   }
355   PetscFunctionReturn(0);
356 }
357 
358 #undef __FUNCT__
359 #define __FUNCT__ "MatLUFactor_Elemental"
360 static PetscErrorCode MatLUFactor_Elemental(Mat A,IS row,IS col,const MatFactorInfo *info)
361 {
362   Mat_Elemental  *a = (Mat_Elemental*)A->data;
363 
364   PetscFunctionBegin;
365   if (info->dtcol){
366     elem::LU(*a->emat,*a->pivot);
367   } else {
368     elem::LU(*a->emat);
369   }
370   A->factortype = MAT_FACTOR_LU;
371   A->assembled  = PETSC_TRUE;
372   PetscFunctionReturn(0);
373 }
374 
375 #undef __FUNCT__
376 #define __FUNCT__ "MatLUFactorNumeric_Elemental"
377 static PetscErrorCode  MatLUFactorNumeric_Elemental(Mat F,Mat A,const MatFactorInfo *info)
378 {
379   PetscErrorCode ierr;
380 
381   PetscFunctionBegin;
382   ierr = MatCopy(A,F,SAME_NONZERO_PATTERN);CHKERRQ(ierr);
383   ierr = MatLUFactor_Elemental(F,0,0,info);CHKERRQ(ierr);
384   PetscFunctionReturn(0);
385 }
386 
387 #undef __FUNCT__
388 #define __FUNCT__ "MatLUFactorSymbolic_Elemental"
389 static PetscErrorCode  MatLUFactorSymbolic_Elemental(Mat F,Mat A,IS r,IS c,const MatFactorInfo *info)
390 {
391   PetscFunctionBegin;
392   /* F is create and allocated by MatGetFactor_elemental_petsc(), skip this routine. */
393   PetscFunctionReturn(0);
394 }
395 
396 #undef __FUNCT__
397 #define __FUNCT__ "MatCholeskyFactor_Elemental"
398 static PetscErrorCode MatCholeskyFactor_Elemental(Mat A,IS perm,const MatFactorInfo *info)
399 {
400   Mat_Elemental  *a = (Mat_Elemental*)A->data;
401   elem::DistMatrix<PetscScalar,elem::MC,elem::MR>   atrans;
402   elem::DistMatrix<PetscScalar,elem::MC,elem::STAR> d;
403 
404   PetscFunctionBegin;
405   if (info->dtcol){
406     /* A = U^T * U for SPD Matrix A */
407     printf("Cholesky Factorization for SPD Matrices...\n");
408     elem::Cholesky(elem::UPPER,*a->emat);
409   } else {
410     /* A = U^T * D * U * for Symmetric Matrix A */
411     printf("LDL^T Factorization for Symmetric Matrices\n");
412     PetscInt i,size;
413     elem::LDLT(*a->emat,d);
414     elem::Transpose(*a->emat,atrans);
415     elem::Copy(atrans,*a->emat);
416     size = (*a->emat).Height();
417     for (i=0;i<size;i++) {
418       (*a->emat).Set(i,i,1.0 / (*a->emat).Get(i,i));
419     }
420   }
421   A->factortype = MAT_FACTOR_CHOLESKY;
422   A->assembled  = PETSC_TRUE;
423   PetscFunctionReturn(0);
424 }
425 
426 EXTERN_C_BEGIN
427 #undef __FUNCT__
428 #define __FUNCT__ "MatGetFactor_elemental_petsc"
429 static PetscErrorCode MatGetFactor_elemental_petsc(Mat A,MatFactorType ftype,Mat *F)
430 {
431   Mat            B;
432   PetscErrorCode ierr;
433 
434   PetscFunctionBegin;
435   /* Create the factorization matrix */
436   ierr = MatCreate(((PetscObject)A)->comm,&B);CHKERRQ(ierr);
437   ierr = MatSetSizes(B,A->rmap->n,A->cmap->n,PETSC_DECIDE,PETSC_DECIDE);CHKERRQ(ierr);
438   ierr = MatSetType(B,MATELEMENTAL);CHKERRQ(ierr);
439   ierr = MatSetUp(B);CHKERRQ(ierr);
440   B->factortype = ftype;
441   *F            = B;
442   PetscFunctionReturn(0);
443 }
444 EXTERN_C_END
445 
446 #undef __FUNCT__
447 #define __FUNCT__ "MatNorm_Elemental"
448 static PetscErrorCode MatNorm_Elemental(Mat A,NormType type,PetscReal *nrm)
449 {
450   Mat_Elemental *a=(Mat_Elemental*)A->data;
451 
452   PetscFunctionBegin;
453   switch (type){
454   case NORM_1:
455     *nrm = elem::Norm(*a->emat,elem::ONE_NORM);
456     break;
457   case NORM_FROBENIUS:
458     *nrm = elem::Norm(*a->emat,elem::FROBENIUS_NORM);
459     break;
460   case NORM_INFINITY:
461     *nrm = elem::Norm(*a->emat,elem::INFINITY_NORM);
462     break;
463   default:
464     printf("Error: unsupported norm type!\n");
465   }
466   PetscFunctionReturn(0);
467 }
468 
469 #undef __FUNCT__
470 #define __FUNCT__ "MatZeroEntries_Elemental"
471 static PetscErrorCode MatZeroEntries_Elemental(Mat A)
472 {
473   Mat_Elemental *a=(Mat_Elemental*)A->data;
474 
475   PetscFunctionBegin;
476   elem::Zero(*a->emat);
477   PetscFunctionReturn(0);
478 }
479 
480 EXTERN_C_BEGIN
481 #undef __FUNCT__
482 #define __FUNCT__ "MatGetOwnershipIS_Elemental"
483 static PetscErrorCode MatGetOwnershipIS_Elemental(Mat A,IS *rows,IS *cols)
484 {
485   Mat_Elemental  *a = (Mat_Elemental*)A->data;
486   PetscErrorCode ierr;
487   PetscInt       i,m,shift,stride,*idx;
488 
489   PetscFunctionBegin;
490   if (rows) {
491     m = a->emat->LocalHeight();
492     shift = a->emat->ColShift();
493     stride = a->emat->ColStride();
494     ierr = PetscMalloc(m*sizeof(PetscInt),&idx);CHKERRQ(ierr);
495     for (i=0; i<m; i++) {
496       PetscInt rank,offset;
497       E2RO(A,0,shift+i*stride,&rank,&offset);
498       RO2P(A,0,rank,offset,&idx[i]);
499     }
500     ierr = ISCreateGeneral(PETSC_COMM_SELF,m,idx,PETSC_OWN_POINTER,rows);CHKERRQ(ierr);
501   }
502   if (cols) {
503     m = a->emat->LocalWidth();
504     shift = a->emat->RowShift();
505     stride = a->emat->RowStride();
506     ierr = PetscMalloc(m*sizeof(PetscInt),&idx);CHKERRQ(ierr);
507     for (i=0; i<m; i++) {
508       PetscInt rank,offset;
509       E2RO(A,1,shift+i*stride,&rank,&offset);
510       RO2P(A,1,rank,offset,&idx[i]);
511     }
512     ierr = ISCreateGeneral(PETSC_COMM_SELF,m,idx,PETSC_OWN_POINTER,cols);CHKERRQ(ierr);
513   }
514   PetscFunctionReturn(0);
515 }
516 EXTERN_C_END
517 
518 #undef __FUNCT__
519 #define __FUNCT__ "MatDestroy_Elemental"
520 static PetscErrorCode MatDestroy_Elemental(Mat A)
521 {
522   Mat_Elemental      *a = (Mat_Elemental*)A->data;
523   PetscErrorCode     ierr;
524   Mat_Elemental_Grid *commgrid;
525   PetscBool          flg;
526   MPI_Comm           icomm;
527 
528   PetscFunctionBegin;
529   delete a->interface;
530   delete a->esubmat;
531   delete a->emat;
532   elem::mpi::Comm cxxcomm(((PetscObject)A)->comm);
533 
534   ierr = PetscCommDuplicate(cxxcomm,&icomm,PETSC_NULL);CHKERRQ(ierr);
535   ierr = MPI_Attr_get(icomm,Petsc_Elemental_keyval,(void**)&commgrid,(int*)&flg);CHKERRQ(ierr);
536   if (--commgrid->grid_refct == 0) {
537     delete commgrid->grid;
538     ierr = PetscFree(commgrid);CHKERRQ(ierr);
539   }
540   ierr = PetscCommDestroy(&icomm);CHKERRQ(ierr);
541   ierr = PetscObjectComposeFunctionDynamic((PetscObject)A,"MatGetOwnershipIS_C","",PETSC_NULL);CHKERRQ(ierr);
542   ierr = PetscObjectComposeFunctionDynamic((PetscObject)A,"MatGetFactor_petsc_C","",PETSC_NULL);CHKERRQ(ierr);
543   ierr = PetscFree(A->data);CHKERRQ(ierr);
544   PetscFunctionReturn(0);
545 }
546 
547 #undef __FUNCT__
548 #define __FUNCT__ "MatSetUp_Elemental"
549 PetscErrorCode MatSetUp_Elemental(Mat A)
550 {
551   Mat_Elemental  *a = (Mat_Elemental*)A->data;
552   PetscErrorCode ierr;
553   PetscMPIInt    rsize,csize;
554 
555   PetscFunctionBegin;
556   ierr = PetscLayoutSetUp(A->rmap);CHKERRQ(ierr);
557   ierr = PetscLayoutSetUp(A->cmap);CHKERRQ(ierr);
558 
559   a->emat->ResizeTo(A->rmap->N,A->cmap->N);CHKERRQ(ierr);
560   elem::Zero(*a->emat);
561 
562   ierr = MPI_Comm_size(A->rmap->comm,&rsize);CHKERRQ(ierr);
563   ierr = MPI_Comm_size(A->cmap->comm,&csize);CHKERRQ(ierr);
564   if (csize != rsize) SETERRQ(((PetscObject)A)->comm,PETSC_ERR_ARG_INCOMP,"Cannot use row and column communicators of different sizes");
565   a->commsize = rsize;
566   a->mr[0] = A->rmap->N % rsize; if (!a->mr[0]) a->mr[0] = rsize;
567   a->mr[1] = A->cmap->N % csize; if (!a->mr[1]) a->mr[1] = csize;
568   a->m[0] = A->rmap->N / rsize + (a->mr[0] != rsize);
569   a->m[1] = A->cmap->N / csize + (a->mr[1] != csize);
570   PetscFunctionReturn(0);
571 }
572 
573 #undef __FUNCT__
574 #define __FUNCT__ "MatAssemblyBegin_Elemental"
575 PetscErrorCode MatAssemblyBegin_Elemental(Mat A, MatAssemblyType type)
576 {
577   Mat_Elemental  *a = (Mat_Elemental*)A->data;
578 
579   PetscFunctionBegin;
580   a->interface->Detach();
581   a->interface->Attach(elem::LOCAL_TO_GLOBAL,*(a->emat));
582   PetscFunctionReturn(0);
583 }
584 
585 #undef __FUNCT__
586 #define __FUNCT__ "MatAssemblyEnd_Elemental"
587 PetscErrorCode MatAssemblyEnd_Elemental(Mat A, MatAssemblyType type)
588 {
589   PetscFunctionBegin;
590   /* Currently does nothing */
591   PetscFunctionReturn(0);
592 }
593 
594 /*MC
595    MATELEMENTAL = "elemental" - A matrix type for dense matrices using the Elemental package
596 
597    Options Database Keys:
598 . -mat_type elemental - sets the matrix type to "elemental" during a call to MatSetFromOptions()
599 
600   Level: beginner
601 
602 .seealso: MATDENSE,MatCreateElemental()
603 M*/
604 EXTERN_C_BEGIN
605 #undef __FUNCT__
606 #define __FUNCT__ "MatCreate_Elemental"
607 PETSC_EXTERN_C PetscErrorCode MatCreate_Elemental(Mat A)
608 {
609   Mat_Elemental      *a;
610   PetscErrorCode     ierr;
611   PetscBool          flg;
612   Mat_Elemental_Grid *commgrid;
613   MPI_Comm           icomm;
614 
615   PetscFunctionBegin;
616   ierr = PetscElementalInitializePackage(PETSC_NULL);CHKERRQ(ierr);
617 
618   ierr = PetscNewLog(A,Mat_Elemental,&a);CHKERRQ(ierr);
619   A->data = (void*)a;
620 
621   A->ops->view            = MatView_Elemental;
622   A->ops->destroy         = MatDestroy_Elemental;
623   A->ops->setup           = MatSetUp_Elemental;
624   A->ops->setvalues       = MatSetValues_Elemental;
625   A->ops->mult            = MatMult_Elemental;
626   A->ops->multadd         = MatMultAdd_Elemental;
627   A->ops->matmult         = MatMatMult_Elemental;
628   A->ops->matmultsymbolic = MatMatMultSymbolic_Elemental;
629   A->ops->matmultnumeric  = MatMatMultNumeric_Elemental;
630   A->ops->assemblybegin   = MatAssemblyBegin_Elemental;
631   A->ops->assemblyend     = MatAssemblyEnd_Elemental;
632   A->ops->scale           = MatScale_Elemental;
633   A->ops->axpy            = MatAXPY_Elemental;
634   A->ops->lufactor        = MatLUFactor_Elemental;
635   A->ops->lufactorsymbolic = MatLUFactorSymbolic_Elemental;
636   A->ops->lufactornumeric = MatLUFactorNumeric_Elemental;
637   A->ops->matsolve        = MatMatSolve_Elemental;
638   A->ops->copy            = MatCopy_Elemental;
639   A->ops->transpose       = MatTranspose_Elemental;
640   A->ops->norm            = MatNorm_Elemental;
641   A->ops->solve           = MatSolve_Elemental;
642   A->ops->zeroentries     = MatZeroEntries_Elemental;
643   A->ops->choleskyfactor  = MatCholeskyFactor_Elemental;
644 
645   A->insertmode = NOT_SET_VALUES;
646 
647   /* Set up the elemental matrix */
648   elem::mpi::Comm cxxcomm(((PetscObject)A)->comm);
649 
650   /* Grid needs to be shared between multiple Mats on the same communicator, implement by attribute caching on the MPI_Comm */
651   if (Petsc_Elemental_keyval == MPI_KEYVAL_INVALID) {
652     ierr = MPI_Keyval_create(MPI_NULL_COPY_FN,MPI_NULL_DELETE_FN,&Petsc_Elemental_keyval,(void*)0); // MPI_Keyval_free()?
653   }
654   ierr = PetscCommDuplicate(cxxcomm,&icomm,PETSC_NULL);CHKERRQ(ierr);
655   ierr = MPI_Attr_get(icomm,Petsc_Elemental_keyval,(void**)&commgrid,(int*)&flg);CHKERRQ(ierr);
656   if (!flg) {
657     ierr = PetscNewLog(A,Mat_Elemental_Grid,&commgrid);CHKERRQ(ierr);
658     commgrid->grid       = new elem::Grid(cxxcomm);
659     commgrid->grid_refct = 1;
660     ierr = MPI_Attr_put(icomm,Petsc_Elemental_keyval,(void*)commgrid);CHKERRQ(ierr);
661   } else {
662     commgrid->grid_refct++;
663   }
664   ierr = PetscCommDestroy(&icomm);CHKERRQ(ierr);
665   a->grid      = commgrid->grid;
666   a->emat      = new elem::DistMatrix<PetscScalar>(*a->grid);
667   a->esubmat   = new elem::Matrix<PetscScalar>(1,1);
668   a->interface = new elem::AxpyInterface<PetscScalar>;
669   a->pivot     = new elem::DistMatrix<PetscInt,elem::VC,elem::STAR>;
670 
671   /* build cache for off array entries formed */
672   a->interface->Attach(elem::LOCAL_TO_GLOBAL,*(a->emat));
673 
674   ierr = PetscObjectComposeFunctionDynamic((PetscObject)A,"MatGetOwnershipIS_C","MatGetOwnershipIS_Elemental",MatGetOwnershipIS_Elemental);CHKERRQ(ierr);
675   ierr = PetscObjectComposeFunctionDynamic((PetscObject)A,"MatGetFactor_petsc_C","MatGetFactor_elemental_petsc",MatGetFactor_elemental_petsc);CHKERRQ(ierr);
676 
677   ierr = PetscObjectChangeTypeName((PetscObject)A,MATELEMENTAL);CHKERRQ(ierr);
678   PetscFunctionReturn(0);
679 }
680 EXTERN_C_END
681