Bug Summary

File:ksp/pc/impls/redundant/redundant.c
Warning:line 2427, column 76
Access to field 'child' results in a dereference of a null pointer (loaded from variable 'scomm')

Annotated Source Code

[?] Use j/k keys for keyboard navigation

/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c

1
2/*
3 This file defines a "solve the problem redundantly on each subgroup of processor" preconditioner.
4*/
5#include <petsc/private/pcimpl.h>
6#include <petscksp.h> /*I "petscksp.h" I*/
7
8typedef struct {
9 KSP ksp;
10 PC pc; /* actual preconditioner used on each processor */
11 Vec xsub,ysub; /* vectors of a subcommunicator to hold parallel vectors of PetscObjectComm((PetscObject)pc) */
12 Vec xdup,ydup; /* parallel vector that congregates xsub or ysub facilitating vector scattering */
13 Mat pmats; /* matrix and optional preconditioner matrix belong to a subcommunicator */
14 VecScatter scatterin,scatterout; /* scatter used to move all values to each processor group (subcommunicator) */
15 PetscBool useparallelmat;
16 PetscSubcomm psubcomm;
17 PetscInt nsubcomm; /* num of data structure PetscSubcomm */
18 PetscBool shifttypeset;
19 MatFactorShiftType shifttype;
20} PC_Redundant;
21
22PetscErrorCode PCFactorSetShiftType_Redundant(PC pc,MatFactorShiftType shifttype)
23{
24 PC_Redundant *red = (PC_Redundant*)pc->data;
25 PetscErrorCode ierr;
26
27 PetscFunctionBegindo { do { ; if (petscstack && (petscstack->currentsize
< 64)) { petscstack->function[petscstack->currentsize
] = __func__; petscstack->file[petscstack->currentsize]
= "/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
; petscstack->line[petscstack->currentsize] = 27; petscstack
->petscroutine[petscstack->currentsize] = PETSC_TRUE; petscstack
->currentsize++; } if (petscstack) { petscstack->hotdepth
+= (PETSC_FALSE || petscstack->hotdepth); } ; } while (0)
; ; } while (0)
;
28 if (red->ksp) {
29 PC pc;
30 ierr = KSPGetPC(red->ksp,&pc);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),30,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
31 ierr = PCFactorSetShiftType(pc,shifttype);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),31,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
32 } else {
33 red->shifttypeset = PETSC_TRUE;
34 red->shifttype = shifttype;
35 }
36 PetscFunctionReturn(0)do { do { ; if (petscstack && petscstack->currentsize
> 0) { petscstack->currentsize--; petscstack->function
[petscstack->currentsize] = 0; petscstack->file[petscstack
->currentsize] = 0; petscstack->line[petscstack->currentsize
] = 0; petscstack->petscroutine[petscstack->currentsize
] = PETSC_FALSE; } if (petscstack) { petscstack->hotdepth =
(((petscstack->hotdepth-1)<(0)) ? (0) : (petscstack->
hotdepth-1)); } ; } while (0); return(0);} while (0)
;
37}
38
39static PetscErrorCode PCView_Redundant(PC pc,PetscViewer viewer)
40{
41 PC_Redundant *red = (PC_Redundant*)pc->data;
42 PetscErrorCode ierr;
43 PetscBool iascii,isstring;
44 PetscViewer subviewer;
45
46 PetscFunctionBegindo { do { ; if (petscstack && (petscstack->currentsize
< 64)) { petscstack->function[petscstack->currentsize
] = __func__; petscstack->file[petscstack->currentsize]
= "/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
; petscstack->line[petscstack->currentsize] = 46; petscstack
->petscroutine[petscstack->currentsize] = PETSC_TRUE; petscstack
->currentsize++; } if (petscstack) { petscstack->hotdepth
+= (PETSC_FALSE || petscstack->hotdepth); } ; } while (0)
; ; } while (0)
;
47 ierr = PetscObjectTypeCompare((PetscObject)viewer,PETSCVIEWERASCII"ascii",&iascii);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),47,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
48 ierr = PetscObjectTypeCompare((PetscObject)viewer,PETSCVIEWERSTRING"string",&isstring);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),48,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
49 if (iascii) {
50 if (!red->psubcomm) {
51 ierr = PetscViewerASCIIPrintf(viewer," Not yet setup\n");CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),51,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
52 } else {
53 ierr = PetscViewerASCIIPrintf(viewer," First (color=0) of %D PCs follows\n",red->nsubcomm);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),53,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
54 ierr = PetscViewerGetSubViewer(viewer,((PetscObject)red->pc)->comm,&subviewer);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),54,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
55 if (!red->psubcomm->color) { /* only view first redundant pc */
56 ierr = PetscViewerASCIIPushTab(subviewer);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),56,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
57 ierr = KSPView(red->ksp,subviewer);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),57,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
58 ierr = PetscViewerASCIIPopTab(subviewer);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),58,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
59 }
60 ierr = PetscViewerRestoreSubViewer(viewer,((PetscObject)red->pc)->comm,&subviewer);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),60,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
61 }
62 } else if (isstring) {
63 ierr = PetscViewerStringSPrintf(viewer," Redundant solver preconditioner");CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),63,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
64 }
65 PetscFunctionReturn(0)do { do { ; if (petscstack && petscstack->currentsize
> 0) { petscstack->currentsize--; petscstack->function
[petscstack->currentsize] = 0; petscstack->file[petscstack
->currentsize] = 0; petscstack->line[petscstack->currentsize
] = 0; petscstack->petscroutine[petscstack->currentsize
] = PETSC_FALSE; } if (petscstack) { petscstack->hotdepth =
(((petscstack->hotdepth-1)<(0)) ? (0) : (petscstack->
hotdepth-1)); } ; } while (0); return(0);} while (0)
;
66}
67
68#include <../src/mat/impls/aij/mpi/mpiaij.h>
69static PetscErrorCode PCSetUp_Redundant(PC pc)
70{
71 PC_Redundant *red = (PC_Redundant*)pc->data;
72 PetscErrorCode ierr;
73 PetscInt mstart,mend,mlocal,M;
74 PetscMPIInt size;
75 MPI_Comm comm,subcomm;
76 Vec x;
77
78 PetscFunctionBegindo { do { ; if (petscstack && (petscstack->currentsize
< 64)) { petscstack->function[petscstack->currentsize
] = __func__; petscstack->file[petscstack->currentsize]
= "/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
; petscstack->line[petscstack->currentsize] = 78; petscstack
->petscroutine[petscstack->currentsize] = PETSC_TRUE; petscstack
->currentsize++; } if (petscstack) { petscstack->hotdepth
+= (PETSC_FALSE || petscstack->hotdepth); } ; } while (0)
; ; } while (0)
;
79 ierr = PetscObjectGetComm((PetscObject)pc,&comm);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),79,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
80
81 /* if pmatrix set by user is sequential then we do not need to gather the parallel matrix */
82 ierr = MPI_Comm_size(comm,&size);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),82,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
83 if (size == 1) red->useparallelmat = PETSC_FALSE;
1
Assuming 'size' is not equal to 1
2
Taking false branch
84
85 if (!pc->setupcalled) {
3
Assuming the condition is true
4
Taking true branch
86 PetscInt mloc_sub;
87 if (!red->psubcomm) { /* create red->psubcomm, new ksp and pc over subcomm */
5
Assuming pointer value is null
6
Taking true branch
88 KSP ksp;
89 ierr = PCRedundantGetKSP(pc,&ksp);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),89,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
90 }
91 subcomm = PetscSubcommChild(red->psubcomm);
7
Passing null pointer value via 1st parameter 'scomm'
8
Calling 'PetscSubcommChild'
92
93 if (red->useparallelmat) {
94 /* grab the parallel matrix and put it into processors of a subcomminicator */
95 ierr = MatCreateRedundantMatrix(pc->pmat,red->psubcomm->n,subcomm,MAT_INITIAL_MATRIX,&red->pmats);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),95,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
96
97 ierr = MPI_Comm_size(subcomm,&size);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),97,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
98 if (size > 1) {
99 PetscBool foundpack;
100 ierr = MatGetFactorAvailable(red->pmats,NULL((void*)0),MAT_FACTOR_LU,&foundpack);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),100,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
101 if (!foundpack) { /* reset default ksp and pc */
102 ierr = KSPSetType(red->ksp,KSPGMRES"gmres");CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),102,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
103 ierr = PCSetType(red->pc,PCBJACOBI"bjacobi");CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),103,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
104 } else {
105 ierr = PCFactorSetMatSolverType(red->pc,NULL((void*)0));CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),105,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
106 }
107 }
108
109 ierr = KSPSetOperators(red->ksp,red->pmats,red->pmats);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),109,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
110
111 /* get working vectors xsub and ysub */
112 ierr = MatCreateVecs(red->pmats,&red->xsub,&red->ysub);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),112,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
113
114 /* create working vectors xdup and ydup.
115 xdup concatenates all xsub's contigously to form a mpi vector over dupcomm (see PetscSubcommCreate_interlaced())
116 ydup concatenates all ysub and has empty local arrays because ysub's arrays will be place into it.
117 Note: we use communicator dupcomm, not PetscObjectComm((PetscObject)pc)! */
118 ierr = MatGetLocalSize(red->pmats,&mloc_sub,NULL((void*)0));CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),118,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
119 ierr = VecCreateMPI(PetscSubcommContiguousParent(red->psubcomm),mloc_sub,PETSC_DECIDE-1,&red->xdup);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),119,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
120 ierr = VecCreateMPIWithArray(PetscSubcommContiguousParent(red->psubcomm),1,mloc_sub,PETSC_DECIDE-1,NULL((void*)0),&red->ydup);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),120,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
121
122 /* create vecscatters */
123 if (!red->scatterin) { /* efficiency of scatterin is independent from psubcomm_type! */
124 IS is1,is2;
125 PetscInt *idx1,*idx2,i,j,k;
126
127 ierr = MatCreateVecs(pc->pmat,&x,0);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),127,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
128 ierr = VecGetSize(x,&M);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),128,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
129 ierr = VecGetOwnershipRange(x,&mstart,&mend);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),129,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
130 mlocal = mend - mstart;
131 ierr = PetscMalloc2(red->psubcomm->n*mlocal,&idx1,red->psubcomm->n*mlocal,&idx2)PetscMallocA(2,PETSC_FALSE,131,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,(size_t)(red->psubcomm->n*mlocal)*sizeof(**(&idx1)
),(&idx1),(size_t)(red->psubcomm->n*mlocal)*sizeof(
**(&idx2)),(&idx2))
;CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),131,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
132 j = 0;
133 for (k=0; k<red->psubcomm->n; k++) {
134 for (i=mstart; i<mend; i++) {
135 idx1[j] = i;
136 idx2[j++] = i + M*k;
137 }
138 }
139 ierr = ISCreateGeneral(comm,red->psubcomm->n*mlocal,idx1,PETSC_COPY_VALUES,&is1);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),139,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
140 ierr = ISCreateGeneral(comm,red->psubcomm->n*mlocal,idx2,PETSC_COPY_VALUES,&is2);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),140,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
141 ierr = VecScatterCreate(x,is1,red->xdup,is2,&red->scatterin);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),141,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
142 ierr = ISDestroy(&is1);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),142,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
143 ierr = ISDestroy(&is2);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),143,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
144
145 /* Impl below is good for PETSC_SUBCOMM_INTERLACED (no inter-process communication) and PETSC_SUBCOMM_CONTIGUOUS (communication within subcomm) */
146 ierr = ISCreateStride(comm,mlocal,mstart+ red->psubcomm->color*M,1,&is1);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),146,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
147 ierr = ISCreateStride(comm,mlocal,mstart,1,&is2);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),147,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
148 ierr = VecScatterCreate(red->xdup,is1,x,is2,&red->scatterout);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),148,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
149 ierr = ISDestroy(&is1);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),149,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
150 ierr = ISDestroy(&is2);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),150,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
151 ierr = PetscFree2(idx1,idx2)PetscFreeA(2,151,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,&(idx1),&(idx2))
;CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),151,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
152 ierr = VecDestroy(&x);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),152,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
153 }
154 } else { /* !red->useparallelmat */
155 ierr = KSPSetOperators(red->ksp,pc->mat,pc->pmat);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),155,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
156 }
157 } else { /* pc->setupcalled */
158 if (red->useparallelmat) {
159 MatReuse reuse;
160 /* grab the parallel matrix and put it into processors of a subcomminicator */
161 /*--------------------------------------------------------------------------*/
162 if (pc->flag == DIFFERENT_NONZERO_PATTERN) {
163 /* destroy old matrices */
164 ierr = MatDestroy(&red->pmats);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),164,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
165 reuse = MAT_INITIAL_MATRIX;
166 } else {
167 reuse = MAT_REUSE_MATRIX;
168 }
169 ierr = MatCreateRedundantMatrix(pc->pmat,red->psubcomm->n,PetscSubcommChild(red->psubcomm),reuse,&red->pmats);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),169,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
170 ierr = KSPSetOperators(red->ksp,red->pmats,red->pmats);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),170,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
171 } else { /* !red->useparallelmat */
172 ierr = KSPSetOperators(red->ksp,pc->mat,pc->pmat);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),172,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
173 }
174 }
175
176 if (pc->setfromoptionscalled) {
177 ierr = KSPSetFromOptions(red->ksp);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),177,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
178 }
179 ierr = KSPSetUp(red->ksp);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),179,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
180 PetscFunctionReturn(0)do { do { ; if (petscstack && petscstack->currentsize
> 0) { petscstack->currentsize--; petscstack->function
[petscstack->currentsize] = 0; petscstack->file[petscstack
->currentsize] = 0; petscstack->line[petscstack->currentsize
] = 0; petscstack->petscroutine[petscstack->currentsize
] = PETSC_FALSE; } if (petscstack) { petscstack->hotdepth =
(((petscstack->hotdepth-1)<(0)) ? (0) : (petscstack->
hotdepth-1)); } ; } while (0); return(0);} while (0)
;
181}
182
183static PetscErrorCode PCApply_Redundant(PC pc,Vec x,Vec y)
184{
185 PC_Redundant *red = (PC_Redundant*)pc->data;
186 PetscErrorCode ierr;
187 PetscScalar *array;
188
189 PetscFunctionBegindo { do { ; if (petscstack && (petscstack->currentsize
< 64)) { petscstack->function[petscstack->currentsize
] = __func__; petscstack->file[petscstack->currentsize]
= "/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
; petscstack->line[petscstack->currentsize] = 189; petscstack
->petscroutine[petscstack->currentsize] = PETSC_TRUE; petscstack
->currentsize++; } if (petscstack) { petscstack->hotdepth
+= (PETSC_FALSE || petscstack->hotdepth); } ; } while (0)
; ; } while (0)
;
190 if (!red->useparallelmat) {
191 ierr = KSPSolve(red->ksp,x,y);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),191,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
192 ierr = KSPCheckSolve(red->ksp,pc,y);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),192,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
193 PetscFunctionReturn(0)do { do { ; if (petscstack && petscstack->currentsize
> 0) { petscstack->currentsize--; petscstack->function
[petscstack->currentsize] = 0; petscstack->file[petscstack
->currentsize] = 0; petscstack->line[petscstack->currentsize
] = 0; petscstack->petscroutine[petscstack->currentsize
] = PETSC_FALSE; } if (petscstack) { petscstack->hotdepth =
(((petscstack->hotdepth-1)<(0)) ? (0) : (petscstack->
hotdepth-1)); } ; } while (0); return(0);} while (0)
;
194 }
195
196 /* scatter x to xdup */
197 ierr = VecScatterBegin(red->scatterin,x,red->xdup,INSERT_VALUES,SCATTER_FORWARD);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),197,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
198 ierr = VecScatterEnd(red->scatterin,x,red->xdup,INSERT_VALUES,SCATTER_FORWARD);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),198,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
199
200 /* place xdup's local array into xsub */
201 ierr = VecGetArray(red->xdup,&array);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),201,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
202 ierr = VecPlaceArray(red->xsub,(const PetscScalar*)array);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),202,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
203
204 /* apply preconditioner on each processor */
205 ierr = KSPSolve(red->ksp,red->xsub,red->ysub);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),205,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
206 ierr = KSPCheckSolve(red->ksp,pc,red->ysub);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),206,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
207 ierr = VecResetArray(red->xsub);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),207,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
208 ierr = VecRestoreArray(red->xdup,&array);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),208,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
209
210 /* place ysub's local array into ydup */
211 ierr = VecGetArray(red->ysub,&array);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),211,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
212 ierr = VecPlaceArray(red->ydup,(const PetscScalar*)array);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),212,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
213
214 /* scatter ydup to y */
215 ierr = VecScatterBegin(red->scatterout,red->ydup,y,INSERT_VALUES,SCATTER_FORWARD);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),215,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
216 ierr = VecScatterEnd(red->scatterout,red->ydup,y,INSERT_VALUES,SCATTER_FORWARD);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),216,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
217 ierr = VecResetArray(red->ydup);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),217,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
218 ierr = VecRestoreArray(red->ysub,&array);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),218,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
219 PetscFunctionReturn(0)do { do { ; if (petscstack && petscstack->currentsize
> 0) { petscstack->currentsize--; petscstack->function
[petscstack->currentsize] = 0; petscstack->file[petscstack
->currentsize] = 0; petscstack->line[petscstack->currentsize
] = 0; petscstack->petscroutine[petscstack->currentsize
] = PETSC_FALSE; } if (petscstack) { petscstack->hotdepth =
(((petscstack->hotdepth-1)<(0)) ? (0) : (petscstack->
hotdepth-1)); } ; } while (0); return(0);} while (0)
;
220}
221
222static PetscErrorCode PCApplyTranspose_Redundant(PC pc,Vec x,Vec y)
223{
224 PC_Redundant *red = (PC_Redundant*)pc->data;
225 PetscErrorCode ierr;
226 PetscScalar *array;
227
228 PetscFunctionBegindo { do { ; if (petscstack && (petscstack->currentsize
< 64)) { petscstack->function[petscstack->currentsize
] = __func__; petscstack->file[petscstack->currentsize]
= "/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
; petscstack->line[petscstack->currentsize] = 228; petscstack
->petscroutine[petscstack->currentsize] = PETSC_TRUE; petscstack
->currentsize++; } if (petscstack) { petscstack->hotdepth
+= (PETSC_FALSE || petscstack->hotdepth); } ; } while (0)
; ; } while (0)
;
229 if (!red->useparallelmat) {
230 ierr = KSPSolveTranspose(red->ksp,x,y);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),230,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
231 ierr = KSPCheckSolve(red->ksp,pc,y);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),231,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
232 PetscFunctionReturn(0)do { do { ; if (petscstack && petscstack->currentsize
> 0) { petscstack->currentsize--; petscstack->function
[petscstack->currentsize] = 0; petscstack->file[petscstack
->currentsize] = 0; petscstack->line[petscstack->currentsize
] = 0; petscstack->petscroutine[petscstack->currentsize
] = PETSC_FALSE; } if (petscstack) { petscstack->hotdepth =
(((petscstack->hotdepth-1)<(0)) ? (0) : (petscstack->
hotdepth-1)); } ; } while (0); return(0);} while (0)
;
233 }
234
235 /* scatter x to xdup */
236 ierr = VecScatterBegin(red->scatterin,x,red->xdup,INSERT_VALUES,SCATTER_FORWARD);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),236,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
237 ierr = VecScatterEnd(red->scatterin,x,red->xdup,INSERT_VALUES,SCATTER_FORWARD);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),237,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
238
239 /* place xdup's local array into xsub */
240 ierr = VecGetArray(red->xdup,&array);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),240,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
241 ierr = VecPlaceArray(red->xsub,(const PetscScalar*)array);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),241,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
242
243 /* apply preconditioner on each processor */
244 ierr = KSPSolveTranspose(red->ksp,red->xsub,red->ysub);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),244,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
245 ierr = KSPCheckSolve(red->ksp,pc,red->ysub);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),245,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
246 ierr = VecResetArray(red->xsub);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),246,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
247 ierr = VecRestoreArray(red->xdup,&array);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),247,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
248
249 /* place ysub's local array into ydup */
250 ierr = VecGetArray(red->ysub,&array);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),250,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
251 ierr = VecPlaceArray(red->ydup,(const PetscScalar*)array);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),251,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
252
253 /* scatter ydup to y */
254 ierr = VecScatterBegin(red->scatterout,red->ydup,y,INSERT_VALUES,SCATTER_FORWARD);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),254,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
255 ierr = VecScatterEnd(red->scatterout,red->ydup,y,INSERT_VALUES,SCATTER_FORWARD);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),255,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
256 ierr = VecResetArray(red->ydup);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),256,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
257 ierr = VecRestoreArray(red->ysub,&array);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),257,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
258 PetscFunctionReturn(0)do { do { ; if (petscstack && petscstack->currentsize
> 0) { petscstack->currentsize--; petscstack->function
[petscstack->currentsize] = 0; petscstack->file[petscstack
->currentsize] = 0; petscstack->line[petscstack->currentsize
] = 0; petscstack->petscroutine[petscstack->currentsize
] = PETSC_FALSE; } if (petscstack) { petscstack->hotdepth =
(((petscstack->hotdepth-1)<(0)) ? (0) : (petscstack->
hotdepth-1)); } ; } while (0); return(0);} while (0)
;
259}
260
261static PetscErrorCode PCReset_Redundant(PC pc)
262{
263 PC_Redundant *red = (PC_Redundant*)pc->data;
264 PetscErrorCode ierr;
265
266 PetscFunctionBegindo { do { ; if (petscstack && (petscstack->currentsize
< 64)) { petscstack->function[petscstack->currentsize
] = __func__; petscstack->file[petscstack->currentsize]
= "/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
; petscstack->line[petscstack->currentsize] = 266; petscstack
->petscroutine[petscstack->currentsize] = PETSC_TRUE; petscstack
->currentsize++; } if (petscstack) { petscstack->hotdepth
+= (PETSC_FALSE || petscstack->hotdepth); } ; } while (0)
; ; } while (0)
;
267 if (red->useparallelmat) {
268 ierr = VecScatterDestroy(&red->scatterin);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),268,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
269 ierr = VecScatterDestroy(&red->scatterout);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),269,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
270 ierr = VecDestroy(&red->ysub);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),270,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
271 ierr = VecDestroy(&red->xsub);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),271,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
272 ierr = VecDestroy(&red->xdup);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),272,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
273 ierr = VecDestroy(&red->ydup);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),273,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
274 }
275 ierr = MatDestroy(&red->pmats);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),275,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
276 ierr = KSPReset(red->ksp);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),276,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
277 PetscFunctionReturn(0)do { do { ; if (petscstack && petscstack->currentsize
> 0) { petscstack->currentsize--; petscstack->function
[petscstack->currentsize] = 0; petscstack->file[petscstack
->currentsize] = 0; petscstack->line[petscstack->currentsize
] = 0; petscstack->petscroutine[petscstack->currentsize
] = PETSC_FALSE; } if (petscstack) { petscstack->hotdepth =
(((petscstack->hotdepth-1)<(0)) ? (0) : (petscstack->
hotdepth-1)); } ; } while (0); return(0);} while (0)
;
278}
279
280static PetscErrorCode PCDestroy_Redundant(PC pc)
281{
282 PC_Redundant *red = (PC_Redundant*)pc->data;
283 PetscErrorCode ierr;
284
285 PetscFunctionBegindo { do { ; if (petscstack && (petscstack->currentsize
< 64)) { petscstack->function[petscstack->currentsize
] = __func__; petscstack->file[petscstack->currentsize]
= "/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
; petscstack->line[petscstack->currentsize] = 285; petscstack
->petscroutine[petscstack->currentsize] = PETSC_TRUE; petscstack
->currentsize++; } if (petscstack) { petscstack->hotdepth
+= (PETSC_FALSE || petscstack->hotdepth); } ; } while (0)
; ; } while (0)
;
286 ierr = PCReset_Redundant(pc);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),286,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
287 ierr = KSPDestroy(&red->ksp);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),287,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
288 ierr = PetscSubcommDestroy(&red->psubcomm);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),288,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
289 ierr = PetscFree(pc->data)((*PetscTrFree)((void*)(pc->data),289,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
) || ((pc->data) = 0,0))
;CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),289,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
290 PetscFunctionReturn(0)do { do { ; if (petscstack && petscstack->currentsize
> 0) { petscstack->currentsize--; petscstack->function
[petscstack->currentsize] = 0; petscstack->file[petscstack
->currentsize] = 0; petscstack->line[petscstack->currentsize
] = 0; petscstack->petscroutine[petscstack->currentsize
] = PETSC_FALSE; } if (petscstack) { petscstack->hotdepth =
(((petscstack->hotdepth-1)<(0)) ? (0) : (petscstack->
hotdepth-1)); } ; } while (0); return(0);} while (0)
;
291}
292
293static PetscErrorCode PCSetFromOptions_Redundant(PetscOptionItems *PetscOptionsObject,PC pc)
294{
295 PetscErrorCode ierr;
296 PC_Redundant *red = (PC_Redundant*)pc->data;
297
298 PetscFunctionBegindo { do { ; if (petscstack && (petscstack->currentsize
< 64)) { petscstack->function[petscstack->currentsize
] = __func__; petscstack->file[petscstack->currentsize]
= "/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
; petscstack->line[petscstack->currentsize] = 298; petscstack
->petscroutine[petscstack->currentsize] = PETSC_TRUE; petscstack
->currentsize++; } if (petscstack) { petscstack->hotdepth
+= (PETSC_FALSE || petscstack->hotdepth); } ; } while (0)
; ; } while (0)
;
299 ierr = PetscOptionsHead(PetscOptionsObject,"Redundant options");CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),299,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
300 ierr = PetscOptionsInt("-pc_redundant_number","Number of redundant pc","PCRedundantSetNumber",red->nsubcomm,&red->nsubcomm,0)PetscOptionsInt_Private(PetscOptionsObject,"-pc_redundant_number"
,"Number of redundant pc","PCRedundantSetNumber",red->nsubcomm
,&red->nsubcomm,0,(-2147483647 - 1),2147483647)
;CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),300,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
301 ierr = PetscOptionsTail()0; do {if (PetscOptionsObject->count != 1) do { do { ; if (
petscstack && petscstack->currentsize > 0) { petscstack
->currentsize--; petscstack->function[petscstack->currentsize
] = 0; petscstack->file[petscstack->currentsize] = 0; petscstack
->line[petscstack->currentsize] = 0; petscstack->petscroutine
[petscstack->currentsize] = PETSC_FALSE; } if (petscstack)
{ petscstack->hotdepth = (((petscstack->hotdepth-1)<
(0)) ? (0) : (petscstack->hotdepth-1)); } ; } while (0); return
(0);} while (0);} while(0)
;CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),301,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
302 PetscFunctionReturn(0)do { do { ; if (petscstack && petscstack->currentsize
> 0) { petscstack->currentsize--; petscstack->function
[petscstack->currentsize] = 0; petscstack->file[petscstack
->currentsize] = 0; petscstack->line[petscstack->currentsize
] = 0; petscstack->petscroutine[petscstack->currentsize
] = PETSC_FALSE; } if (petscstack) { petscstack->hotdepth =
(((petscstack->hotdepth-1)<(0)) ? (0) : (petscstack->
hotdepth-1)); } ; } while (0); return(0);} while (0)
;
303}
304
305static PetscErrorCode PCRedundantSetNumber_Redundant(PC pc,PetscInt nreds)
306{
307 PC_Redundant *red = (PC_Redundant*)pc->data;
308
309 PetscFunctionBegindo { do { ; if (petscstack && (petscstack->currentsize
< 64)) { petscstack->function[petscstack->currentsize
] = __func__; petscstack->file[petscstack->currentsize]
= "/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
; petscstack->line[petscstack->currentsize] = 309; petscstack
->petscroutine[petscstack->currentsize] = PETSC_TRUE; petscstack
->currentsize++; } if (petscstack) { petscstack->hotdepth
+= (PETSC_FALSE || petscstack->hotdepth); } ; } while (0)
; ; } while (0)
;
310 red->nsubcomm = nreds;
311 PetscFunctionReturn(0)do { do { ; if (petscstack && petscstack->currentsize
> 0) { petscstack->currentsize--; petscstack->function
[petscstack->currentsize] = 0; petscstack->file[petscstack
->currentsize] = 0; petscstack->line[petscstack->currentsize
] = 0; petscstack->petscroutine[petscstack->currentsize
] = PETSC_FALSE; } if (petscstack) { petscstack->hotdepth =
(((petscstack->hotdepth-1)<(0)) ? (0) : (petscstack->
hotdepth-1)); } ; } while (0); return(0);} while (0)
;
312}
313
314/*@
315 PCRedundantSetNumber - Sets the number of redundant preconditioner contexts.
316
317 Logically Collective on PC
318
319 Input Parameters:
320+ pc - the preconditioner context
321- nredundant - number of redundant preconditioner contexts; for example if you are using 64 MPI processes and
322 use an nredundant of 4 there will be 4 parallel solves each on 16 = 64/4 processes.
323
324 Level: advanced
325
326@*/
327PetscErrorCode PCRedundantSetNumber(PC pc,PetscInt nredundant)
328{
329 PetscErrorCode ierr;
330
331 PetscFunctionBegindo { do { ; if (petscstack && (petscstack->currentsize
< 64)) { petscstack->function[petscstack->currentsize
] = __func__; petscstack->file[petscstack->currentsize]
= "/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
; petscstack->line[petscstack->currentsize] = 331; petscstack
->petscroutine[petscstack->currentsize] = PETSC_TRUE; petscstack
->currentsize++; } if (petscstack) { petscstack->hotdepth
+= (PETSC_FALSE || petscstack->hotdepth); } ; } while (0)
; ; } while (0)
;
332 PetscValidHeaderSpecific(pc,PC_CLASSID,1)do { if (!pc) return PetscError(((MPI_Comm)0x44000001),332,__func__
,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,85,PETSC_ERROR_INITIAL,"Null Object: Parameter # %d",1); if (
!PetscCheckPointer(pc,PETSC_OBJECT)) return PetscError(((MPI_Comm
)0x44000001),332,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,64,PETSC_ERROR_INITIAL,"Invalid Pointer to Object: Parameter # %d"
,1); if (((PetscObject)(pc))->classid != PC_CLASSID) { if (
((PetscObject)(pc))->classid == -1) return PetscError(((MPI_Comm
)0x44000001),332,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,64,PETSC_ERROR_INITIAL,"Object already free: Parameter # %d"
,1); else return PetscError(((MPI_Comm)0x44000001),332,__func__
,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,62,PETSC_ERROR_INITIAL,"Wrong type of object: Parameter # %d"
,1); } } while (0)
;
333 if (nredundant <= 0) SETERRQ1(PetscObjectComm((PetscObject)pc),PETSC_ERR_ARG_WRONG, "num of redundant pc %D must be positive",nredundant)return PetscError(PetscObjectComm((PetscObject)pc),333,__func__
,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,62,PETSC_ERROR_INITIAL,"num of redundant pc %D must be positive"
,nredundant)
;
334 ierr = PetscTryMethod(pc,"PCRedundantSetNumber_C",(PC,PetscInt),(pc,nredundant))0; do { PetscErrorCode (*f)(PC,PetscInt), __ierr; __ierr = PetscObjectQueryFunction_Private
(((PetscObject)pc),("PCRedundantSetNumber_C"),(PetscVoidFunction
*)(&f));do {if (__builtin_expect(!!(__ierr),0)) return PetscError
(((MPI_Comm)0x44000001),334,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,__ierr,PETSC_ERROR_REPEAT," ");} while (0); if (f) {__ierr =
(*f)(pc,nredundant);do {if (__builtin_expect(!!(__ierr),0)) return
PetscError(((MPI_Comm)0x44000001),334,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,__ierr,PETSC_ERROR_REPEAT," ");} while (0);} } while(0)
;CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),334,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
335 PetscFunctionReturn(0)do { do { ; if (petscstack && petscstack->currentsize
> 0) { petscstack->currentsize--; petscstack->function
[petscstack->currentsize] = 0; petscstack->file[petscstack
->currentsize] = 0; petscstack->line[petscstack->currentsize
] = 0; petscstack->petscroutine[petscstack->currentsize
] = PETSC_FALSE; } if (petscstack) { petscstack->hotdepth =
(((petscstack->hotdepth-1)<(0)) ? (0) : (petscstack->
hotdepth-1)); } ; } while (0); return(0);} while (0)
;
336}
337
338static PetscErrorCode PCRedundantSetScatter_Redundant(PC pc,VecScatter in,VecScatter out)
339{
340 PC_Redundant *red = (PC_Redundant*)pc->data;
341 PetscErrorCode ierr;
342
343 PetscFunctionBegindo { do { ; if (petscstack && (petscstack->currentsize
< 64)) { petscstack->function[petscstack->currentsize
] = __func__; petscstack->file[petscstack->currentsize]
= "/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
; petscstack->line[petscstack->currentsize] = 343; petscstack
->petscroutine[petscstack->currentsize] = PETSC_TRUE; petscstack
->currentsize++; } if (petscstack) { petscstack->hotdepth
+= (PETSC_FALSE || petscstack->hotdepth); } ; } while (0)
; ; } while (0)
;
344 ierr = PetscObjectReference((PetscObject)in);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),344,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
345 ierr = VecScatterDestroy(&red->scatterin);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),345,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
346
347 red->scatterin = in;
348
349 ierr = PetscObjectReference((PetscObject)out);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),349,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
350 ierr = VecScatterDestroy(&red->scatterout);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),350,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
351 red->scatterout = out;
352 PetscFunctionReturn(0)do { do { ; if (petscstack && petscstack->currentsize
> 0) { petscstack->currentsize--; petscstack->function
[petscstack->currentsize] = 0; petscstack->file[petscstack
->currentsize] = 0; petscstack->line[petscstack->currentsize
] = 0; petscstack->petscroutine[petscstack->currentsize
] = PETSC_FALSE; } if (petscstack) { petscstack->hotdepth =
(((petscstack->hotdepth-1)<(0)) ? (0) : (petscstack->
hotdepth-1)); } ; } while (0); return(0);} while (0)
;
353}
354
355/*@
356 PCRedundantSetScatter - Sets the scatter used to copy values into the
357 redundant local solve and the scatter to move them back into the global
358 vector.
359
360 Logically Collective on PC
361
362 Input Parameters:
363+ pc - the preconditioner context
364. in - the scatter to move the values in
365- out - the scatter to move them out
366
367 Level: advanced
368
369@*/
370PetscErrorCode PCRedundantSetScatter(PC pc,VecScatter in,VecScatter out)
371{
372 PetscErrorCode ierr;
373
374 PetscFunctionBegindo { do { ; if (petscstack && (petscstack->currentsize
< 64)) { petscstack->function[petscstack->currentsize
] = __func__; petscstack->file[petscstack->currentsize]
= "/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
; petscstack->line[petscstack->currentsize] = 374; petscstack
->petscroutine[petscstack->currentsize] = PETSC_TRUE; petscstack
->currentsize++; } if (petscstack) { petscstack->hotdepth
+= (PETSC_FALSE || petscstack->hotdepth); } ; } while (0)
; ; } while (0)
;
375 PetscValidHeaderSpecific(pc,PC_CLASSID,1)do { if (!pc) return PetscError(((MPI_Comm)0x44000001),375,__func__
,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,85,PETSC_ERROR_INITIAL,"Null Object: Parameter # %d",1); if (
!PetscCheckPointer(pc,PETSC_OBJECT)) return PetscError(((MPI_Comm
)0x44000001),375,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,64,PETSC_ERROR_INITIAL,"Invalid Pointer to Object: Parameter # %d"
,1); if (((PetscObject)(pc))->classid != PC_CLASSID) { if (
((PetscObject)(pc))->classid == -1) return PetscError(((MPI_Comm
)0x44000001),375,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,64,PETSC_ERROR_INITIAL,"Object already free: Parameter # %d"
,1); else return PetscError(((MPI_Comm)0x44000001),375,__func__
,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,62,PETSC_ERROR_INITIAL,"Wrong type of object: Parameter # %d"
,1); } } while (0)
;
376 PetscValidHeaderSpecific(in,VEC_SCATTER_CLASSID,2)do { if (!in) return PetscError(((MPI_Comm)0x44000001),376,__func__
,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,85,PETSC_ERROR_INITIAL,"Null Object: Parameter # %d",2); if (
!PetscCheckPointer(in,PETSC_OBJECT)) return PetscError(((MPI_Comm
)0x44000001),376,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,64,PETSC_ERROR_INITIAL,"Invalid Pointer to Object: Parameter # %d"
,2); if (((PetscObject)(in))->classid != VEC_SCATTER_CLASSID
) { if (((PetscObject)(in))->classid == -1) return PetscError
(((MPI_Comm)0x44000001),376,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,64,PETSC_ERROR_INITIAL,"Object already free: Parameter # %d"
,2); else return PetscError(((MPI_Comm)0x44000001),376,__func__
,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,62,PETSC_ERROR_INITIAL,"Wrong type of object: Parameter # %d"
,2); } } while (0)
;
377 PetscValidHeaderSpecific(out,VEC_SCATTER_CLASSID,3)do { if (!out) return PetscError(((MPI_Comm)0x44000001),377,__func__
,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,85,PETSC_ERROR_INITIAL,"Null Object: Parameter # %d",3); if (
!PetscCheckPointer(out,PETSC_OBJECT)) return PetscError(((MPI_Comm
)0x44000001),377,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,64,PETSC_ERROR_INITIAL,"Invalid Pointer to Object: Parameter # %d"
,3); if (((PetscObject)(out))->classid != VEC_SCATTER_CLASSID
) { if (((PetscObject)(out))->classid == -1) return PetscError
(((MPI_Comm)0x44000001),377,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,64,PETSC_ERROR_INITIAL,"Object already free: Parameter # %d"
,3); else return PetscError(((MPI_Comm)0x44000001),377,__func__
,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,62,PETSC_ERROR_INITIAL,"Wrong type of object: Parameter # %d"
,3); } } while (0)
;
378 ierr = PetscTryMethod(pc,"PCRedundantSetScatter_C",(PC,VecScatter,VecScatter),(pc,in,out))0; do { PetscErrorCode (*f)(PC,VecScatter,VecScatter), __ierr
; __ierr = PetscObjectQueryFunction_Private(((PetscObject)pc)
,("PCRedundantSetScatter_C"),(PetscVoidFunction*)(&f));do
{if (__builtin_expect(!!(__ierr),0)) return PetscError(((MPI_Comm
)0x44000001),378,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,__ierr,PETSC_ERROR_REPEAT," ");} while (0); if (f) {__ierr =
(*f)(pc,in,out);do {if (__builtin_expect(!!(__ierr),0)) return
PetscError(((MPI_Comm)0x44000001),378,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,__ierr,PETSC_ERROR_REPEAT," ");} while (0);} } while(0)
;CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),378,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
379 PetscFunctionReturn(0)do { do { ; if (petscstack && petscstack->currentsize
> 0) { petscstack->currentsize--; petscstack->function
[petscstack->currentsize] = 0; petscstack->file[petscstack
->currentsize] = 0; petscstack->line[petscstack->currentsize
] = 0; petscstack->petscroutine[petscstack->currentsize
] = PETSC_FALSE; } if (petscstack) { petscstack->hotdepth =
(((petscstack->hotdepth-1)<(0)) ? (0) : (petscstack->
hotdepth-1)); } ; } while (0); return(0);} while (0)
;
380}
381
382static PetscErrorCode PCRedundantGetKSP_Redundant(PC pc,KSP *innerksp)
383{
384 PetscErrorCode ierr;
385 PC_Redundant *red = (PC_Redundant*)pc->data;
386 MPI_Comm comm,subcomm;
387 const char *prefix;
388
389 PetscFunctionBegindo { do { ; if (petscstack && (petscstack->currentsize
< 64)) { petscstack->function[petscstack->currentsize
] = __func__; petscstack->file[petscstack->currentsize]
= "/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
; petscstack->line[petscstack->currentsize] = 389; petscstack
->petscroutine[petscstack->currentsize] = PETSC_TRUE; petscstack
->currentsize++; } if (petscstack) { petscstack->hotdepth
+= (PETSC_FALSE || petscstack->hotdepth); } ; } while (0)
; ; } while (0)
;
390 if (!red->psubcomm) {
391 ierr = PCGetOptionsPrefix(pc,&prefix);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),391,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
392
393 ierr = PetscObjectGetComm((PetscObject)pc,&comm);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),393,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
394 ierr = PetscSubcommCreate(comm,&red->psubcomm);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),394,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
395 ierr = PetscSubcommSetNumber(red->psubcomm,red->nsubcomm);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),395,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
396 ierr = PetscSubcommSetType(red->psubcomm,PETSC_SUBCOMM_CONTIGUOUS);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),396,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
397
398 ierr = PetscSubcommSetOptionsPrefix(red->psubcomm,prefix);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),398,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
399 ierr = PetscSubcommSetFromOptions(red->psubcomm);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),399,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
400 ierr = PetscLogObjectMemory((PetscObject)pc,sizeof(PetscSubcomm));CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),400,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
401
402 /* create a new PC that processors in each subcomm have copy of */
403 subcomm = PetscSubcommChild(red->psubcomm);
404
405 ierr = KSPCreate(subcomm,&red->ksp);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),405,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
406 ierr = KSPSetErrorIfNotConverged(red->ksp,pc->erroriffailure);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),406,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
407 ierr = PetscObjectIncrementTabLevel((PetscObject)red->ksp,(PetscObject)pc,1);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),407,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
408 ierr = PetscLogObjectParent((PetscObject)pc,(PetscObject)red->ksp);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),408,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
409 ierr = KSPSetType(red->ksp,KSPPREONLY"preonly");CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),409,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
410 ierr = KSPGetPC(red->ksp,&red->pc);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),410,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
411 ierr = PCSetType(red->pc,PCLU"lu");CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),411,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
412 if (red->shifttypeset) {
413 ierr = PCFactorSetShiftType(red->pc,red->shifttype);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),413,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
414 red->shifttypeset = PETSC_FALSE;
415 }
416 ierr = KSPSetOptionsPrefix(red->ksp,prefix);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),416,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
417 ierr = KSPAppendOptionsPrefix(red->ksp,"redundant_");CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),417,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
418 }
419 *innerksp = red->ksp;
420 PetscFunctionReturn(0)do { do { ; if (petscstack && petscstack->currentsize
> 0) { petscstack->currentsize--; petscstack->function
[petscstack->currentsize] = 0; petscstack->file[petscstack
->currentsize] = 0; petscstack->line[petscstack->currentsize
] = 0; petscstack->petscroutine[petscstack->currentsize
] = PETSC_FALSE; } if (petscstack) { petscstack->hotdepth =
(((petscstack->hotdepth-1)<(0)) ? (0) : (petscstack->
hotdepth-1)); } ; } while (0); return(0);} while (0)
;
421}
422
423/*@
424 PCRedundantGetKSP - Gets the less parallel KSP created by the redundant PC.
425
426 Not Collective
427
428 Input Parameter:
429. pc - the preconditioner context
430
431 Output Parameter:
432. innerksp - the KSP on the smaller set of processes
433
434 Level: advanced
435
436@*/
437PetscErrorCode PCRedundantGetKSP(PC pc,KSP *innerksp)
438{
439 PetscErrorCode ierr;
440
441 PetscFunctionBegindo { do { ; if (petscstack && (petscstack->currentsize
< 64)) { petscstack->function[petscstack->currentsize
] = __func__; petscstack->file[petscstack->currentsize]
= "/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
; petscstack->line[petscstack->currentsize] = 441; petscstack
->petscroutine[petscstack->currentsize] = PETSC_TRUE; petscstack
->currentsize++; } if (petscstack) { petscstack->hotdepth
+= (PETSC_FALSE || petscstack->hotdepth); } ; } while (0)
; ; } while (0)
;
442 PetscValidHeaderSpecific(pc,PC_CLASSID,1)do { if (!pc) return PetscError(((MPI_Comm)0x44000001),442,__func__
,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,85,PETSC_ERROR_INITIAL,"Null Object: Parameter # %d",1); if (
!PetscCheckPointer(pc,PETSC_OBJECT)) return PetscError(((MPI_Comm
)0x44000001),442,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,64,PETSC_ERROR_INITIAL,"Invalid Pointer to Object: Parameter # %d"
,1); if (((PetscObject)(pc))->classid != PC_CLASSID) { if (
((PetscObject)(pc))->classid == -1) return PetscError(((MPI_Comm
)0x44000001),442,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,64,PETSC_ERROR_INITIAL,"Object already free: Parameter # %d"
,1); else return PetscError(((MPI_Comm)0x44000001),442,__func__
,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,62,PETSC_ERROR_INITIAL,"Wrong type of object: Parameter # %d"
,1); } } while (0)
;
443 PetscValidPointer(innerksp,2)do { if (!innerksp) return PetscError(((MPI_Comm)0x44000001),
443,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,85,PETSC_ERROR_INITIAL,"Null Pointer: Parameter # %d",2); if
(!PetscCheckPointer(innerksp,PETSC_CHAR)) return PetscError(
((MPI_Comm)0x44000001),443,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,68,PETSC_ERROR_INITIAL,"Invalid Pointer: Parameter # %d",2);
} while (0)
;
444 ierr = PetscUseMethod(pc,"PCRedundantGetKSP_C",(PC,KSP*),(pc,innerksp))0; do { PetscErrorCode (*f)(PC,KSP*), __ierr; __ierr = PetscObjectQueryFunction_Private
(((PetscObject)pc),("PCRedundantGetKSP_C"),(PetscVoidFunction
*)(&f));do {if (__builtin_expect(!!(__ierr),0)) return PetscError
(((MPI_Comm)0x44000001),444,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,__ierr,PETSC_ERROR_REPEAT," ");} while (0); if (f) {__ierr =
(*f)(pc,innerksp);do {if (__builtin_expect(!!(__ierr),0)) return
PetscError(((MPI_Comm)0x44000001),444,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,__ierr,PETSC_ERROR_REPEAT," ");} while (0);} else return PetscError
(PetscObjectComm((PetscObject)pc),444,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,56,PETSC_ERROR_INITIAL,"Cannot locate function %s in object"
,"PCRedundantGetKSP_C"); } while(0)
;CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),444,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
445 PetscFunctionReturn(0)do { do { ; if (petscstack && petscstack->currentsize
> 0) { petscstack->currentsize--; petscstack->function
[petscstack->currentsize] = 0; petscstack->file[petscstack
->currentsize] = 0; petscstack->line[petscstack->currentsize
] = 0; petscstack->petscroutine[petscstack->currentsize
] = PETSC_FALSE; } if (petscstack) { petscstack->hotdepth =
(((petscstack->hotdepth-1)<(0)) ? (0) : (petscstack->
hotdepth-1)); } ; } while (0); return(0);} while (0)
;
446}
447
448static PetscErrorCode PCRedundantGetOperators_Redundant(PC pc,Mat *mat,Mat *pmat)
449{
450 PC_Redundant *red = (PC_Redundant*)pc->data;
451
452 PetscFunctionBegindo { do { ; if (petscstack && (petscstack->currentsize
< 64)) { petscstack->function[petscstack->currentsize
] = __func__; petscstack->file[petscstack->currentsize]
= "/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
; petscstack->line[petscstack->currentsize] = 452; petscstack
->petscroutine[petscstack->currentsize] = PETSC_TRUE; petscstack
->currentsize++; } if (petscstack) { petscstack->hotdepth
+= (PETSC_FALSE || petscstack->hotdepth); } ; } while (0)
; ; } while (0)
;
453 if (mat) *mat = red->pmats;
454 if (pmat) *pmat = red->pmats;
455 PetscFunctionReturn(0)do { do { ; if (petscstack && petscstack->currentsize
> 0) { petscstack->currentsize--; petscstack->function
[petscstack->currentsize] = 0; petscstack->file[petscstack
->currentsize] = 0; petscstack->line[petscstack->currentsize
] = 0; petscstack->petscroutine[petscstack->currentsize
] = PETSC_FALSE; } if (petscstack) { petscstack->hotdepth =
(((petscstack->hotdepth-1)<(0)) ? (0) : (petscstack->
hotdepth-1)); } ; } while (0); return(0);} while (0)
;
456}
457
458/*@
459 PCRedundantGetOperators - gets the sequential matrix and preconditioner matrix
460
461 Not Collective
462
463 Input Parameter:
464. pc - the preconditioner context
465
466 Output Parameters:
467+ mat - the matrix
468- pmat - the (possibly different) preconditioner matrix
469
470 Level: advanced
471
472@*/
473PetscErrorCode PCRedundantGetOperators(PC pc,Mat *mat,Mat *pmat)
474{
475 PetscErrorCode ierr;
476
477 PetscFunctionBegindo { do { ; if (petscstack && (petscstack->currentsize
< 64)) { petscstack->function[petscstack->currentsize
] = __func__; petscstack->file[petscstack->currentsize]
= "/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
; petscstack->line[petscstack->currentsize] = 477; petscstack
->petscroutine[petscstack->currentsize] = PETSC_TRUE; petscstack
->currentsize++; } if (petscstack) { petscstack->hotdepth
+= (PETSC_FALSE || petscstack->hotdepth); } ; } while (0)
; ; } while (0)
;
478 PetscValidHeaderSpecific(pc,PC_CLASSID,1)do { if (!pc) return PetscError(((MPI_Comm)0x44000001),478,__func__
,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,85,PETSC_ERROR_INITIAL,"Null Object: Parameter # %d",1); if (
!PetscCheckPointer(pc,PETSC_OBJECT)) return PetscError(((MPI_Comm
)0x44000001),478,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,64,PETSC_ERROR_INITIAL,"Invalid Pointer to Object: Parameter # %d"
,1); if (((PetscObject)(pc))->classid != PC_CLASSID) { if (
((PetscObject)(pc))->classid == -1) return PetscError(((MPI_Comm
)0x44000001),478,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,64,PETSC_ERROR_INITIAL,"Object already free: Parameter # %d"
,1); else return PetscError(((MPI_Comm)0x44000001),478,__func__
,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,62,PETSC_ERROR_INITIAL,"Wrong type of object: Parameter # %d"
,1); } } while (0)
;
479 if (mat) PetscValidPointer(mat,2)do { if (!mat) return PetscError(((MPI_Comm)0x44000001),479,__func__
,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,85,PETSC_ERROR_INITIAL,"Null Pointer: Parameter # %d",2); if
(!PetscCheckPointer(mat,PETSC_CHAR)) return PetscError(((MPI_Comm
)0x44000001),479,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,68,PETSC_ERROR_INITIAL,"Invalid Pointer: Parameter # %d",2);
} while (0)
;
480 if (pmat) PetscValidPointer(pmat,3)do { if (!pmat) return PetscError(((MPI_Comm)0x44000001),480,
__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,85,PETSC_ERROR_INITIAL,"Null Pointer: Parameter # %d",3); if
(!PetscCheckPointer(pmat,PETSC_CHAR)) return PetscError(((MPI_Comm
)0x44000001),480,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,68,PETSC_ERROR_INITIAL,"Invalid Pointer: Parameter # %d",3);
} while (0)
;
481 ierr = PetscUseMethod(pc,"PCRedundantGetOperators_C",(PC,Mat*,Mat*),(pc,mat,pmat))0; do { PetscErrorCode (*f)(PC,Mat*,Mat*), __ierr; __ierr = PetscObjectQueryFunction_Private
(((PetscObject)pc),("PCRedundantGetOperators_C"),(PetscVoidFunction
*)(&f));do {if (__builtin_expect(!!(__ierr),0)) return PetscError
(((MPI_Comm)0x44000001),481,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,__ierr,PETSC_ERROR_REPEAT," ");} while (0); if (f) {__ierr =
(*f)(pc,mat,pmat);do {if (__builtin_expect(!!(__ierr),0)) return
PetscError(((MPI_Comm)0x44000001),481,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,__ierr,PETSC_ERROR_REPEAT," ");} while (0);} else return PetscError
(PetscObjectComm((PetscObject)pc),481,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,56,PETSC_ERROR_INITIAL,"Cannot locate function %s in object"
,"PCRedundantGetOperators_C"); } while(0)
;CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),481,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
482 PetscFunctionReturn(0)do { do { ; if (petscstack && petscstack->currentsize
> 0) { petscstack->currentsize--; petscstack->function
[petscstack->currentsize] = 0; petscstack->file[petscstack
->currentsize] = 0; petscstack->line[petscstack->currentsize
] = 0; petscstack->petscroutine[petscstack->currentsize
] = PETSC_FALSE; } if (petscstack) { petscstack->hotdepth =
(((petscstack->hotdepth-1)<(0)) ? (0) : (petscstack->
hotdepth-1)); } ; } while (0); return(0);} while (0)
;
483}
484
485/* -------------------------------------------------------------------------------------*/
486/*MC
487 PCREDUNDANT - Runs a KSP solver with preconditioner for the entire problem on subgroups of processors
488
489 Options for the redundant preconditioners can be set with -redundant_pc_xxx for the redundant KSP with -redundant_ksp_xxx
490
491 Options Database:
492. -pc_redundant_number <n> - number of redundant solves, for example if you are using 64 MPI processes and
493 use an n of 4 there will be 4 parallel solves each on 16 = 64/4 processes.
494
495 Level: intermediate
496
497 Notes:
498 The default KSP is preonly and the default PC is LU.
499
500 PCFactorSetShiftType() applied to this PC will convey they shift type into the inner PC if it is factorization based.
501
502 Developer Notes:
503 Note that PCSetInitialGuessNonzero() is not used by this class but likely should be.
504
505.seealso: PCCreate(), PCSetType(), PCType (for list of available types), PCRedundantSetScatter(),
506 PCRedundantGetKSP(), PCRedundantGetOperators(), PCRedundantSetNumber()
507M*/
508
509PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PCCreate_Redundant(PC pc)
510{
511 PetscErrorCode ierr;
512 PC_Redundant *red;
513 PetscMPIInt size;
514
515 PetscFunctionBegindo { do { ; if (petscstack && (petscstack->currentsize
< 64)) { petscstack->function[petscstack->currentsize
] = __func__; petscstack->file[petscstack->currentsize]
= "/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
; petscstack->line[petscstack->currentsize] = 515; petscstack
->petscroutine[petscstack->currentsize] = PETSC_TRUE; petscstack
->currentsize++; } if (petscstack) { petscstack->hotdepth
+= (PETSC_FALSE || petscstack->hotdepth); } ; } while (0)
; ; } while (0)
;
516 ierr = PetscNewLog(pc,&red)(PetscMallocA(1,PETSC_TRUE,516,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,(size_t)(1)*sizeof(**(((&red)))),(((&red)))) || PetscLogObjectMemory
((PetscObject)pc,sizeof(**(&red))))
;CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),516,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
517 ierr = MPI_Comm_size(PetscObjectComm((PetscObject)pc),&size);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),517,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
518
519 red->nsubcomm = size;
520 red->useparallelmat = PETSC_TRUE;
521 pc->data = (void*)red;
522
523 pc->ops->apply = PCApply_Redundant;
524 pc->ops->applytranspose = PCApplyTranspose_Redundant;
525 pc->ops->setup = PCSetUp_Redundant;
526 pc->ops->destroy = PCDestroy_Redundant;
527 pc->ops->reset = PCReset_Redundant;
528 pc->ops->setfromoptions = PCSetFromOptions_Redundant;
529 pc->ops->view = PCView_Redundant;
530
531 ierr = PetscObjectComposeFunction((PetscObject)pc,"PCRedundantSetScatter_C",PCRedundantSetScatter_Redundant)PetscObjectComposeFunction_Private((PetscObject)pc,"PCRedundantSetScatter_C"
,(PetscVoidFunction)(PCRedundantSetScatter_Redundant))
;CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),531,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
532 ierr = PetscObjectComposeFunction((PetscObject)pc,"PCRedundantSetNumber_C",PCRedundantSetNumber_Redundant)PetscObjectComposeFunction_Private((PetscObject)pc,"PCRedundantSetNumber_C"
,(PetscVoidFunction)(PCRedundantSetNumber_Redundant))
;CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),532,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
533 ierr = PetscObjectComposeFunction((PetscObject)pc,"PCRedundantGetKSP_C",PCRedundantGetKSP_Redundant)PetscObjectComposeFunction_Private((PetscObject)pc,"PCRedundantGetKSP_C"
,(PetscVoidFunction)(PCRedundantGetKSP_Redundant))
;CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),533,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
534 ierr = PetscObjectComposeFunction((PetscObject)pc,"PCRedundantGetOperators_C",PCRedundantGetOperators_Redundant)PetscObjectComposeFunction_Private((PetscObject)pc,"PCRedundantGetOperators_C"
,(PetscVoidFunction)(PCRedundantGetOperators_Redundant))
;CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),534,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
535 ierr = PetscObjectComposeFunction((PetscObject)pc,"PCFactorSetShiftType_C",PCFactorSetShiftType_Redundant)PetscObjectComposeFunction_Private((PetscObject)pc,"PCFactorSetShiftType_C"
,(PetscVoidFunction)(PCFactorSetShiftType_Redundant))
;CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),535,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
536 PetscFunctionReturn(0)do { do { ; if (petscstack && petscstack->currentsize
> 0) { petscstack->currentsize--; petscstack->function
[petscstack->currentsize] = 0; petscstack->file[petscstack
->currentsize] = 0; petscstack->line[petscstack->currentsize
] = 0; petscstack->petscroutine[petscstack->currentsize
] = PETSC_FALSE; } if (petscstack) { petscstack->hotdepth =
(((petscstack->hotdepth-1)<(0)) ? (0) : (petscstack->
hotdepth-1)); } ; } while (0); return(0);} while (0)
;
537}
538

/sandbox/petsc/petsc.next/include/petscsys.h

1/*
2 This is the main PETSc include file (for C and C++). It is included by all
3 other PETSc include files, so it almost never has to be specifically included.
4*/
5#if !defined(PETSCSYS_H)
6#define PETSCSYS_H
7/* ========================================================================== */
8/*
9 petscconf.h is contained in ${PETSC_ARCH}/include/petscconf.h it is
10 found automatically by the compiler due to the -I${PETSC_DIR}/${PETSC_ARCH}/include.
11 For --prefix installs the ${PETSC_ARCH}/ does not exist and petscconf.h is in the same
12 directory as the other PETSc include files.
13*/
14#include <petscconf.h>
15#include <petscfix.h>
16
17#if defined(PETSC_DESIRE_FEATURE_TEST_MACROS)
18/*
19 Feature test macros must be included before headers defined by IEEE Std 1003.1-2001
20 We only turn these in PETSc source files that require them by setting PETSC_DESIRE_FEATURE_TEST_MACROS
21*/
22# if defined(PETSC__POSIX_C_SOURCE_200112L) && !defined(_POSIX_C_SOURCE200809L)
23# define _POSIX_C_SOURCE200809L 200112L
24# endif
25# if defined(PETSC__BSD_SOURCE1) && !defined(_BSD_SOURCE1)
26# define _BSD_SOURCE1
27# endif
28# if defined(PETSC__DEFAULT_SOURCE1) && !defined(_DEFAULT_SOURCE1)
29# define _DEFAULT_SOURCE1
30# endif
31# if defined(PETSC__GNU_SOURCE1) && !defined(_GNU_SOURCE)
32# define _GNU_SOURCE
33# endif
34#endif
35
36#include <petscsystypes.h>
37
38/* ========================================================================== */
39/*
40 This facilitates using the C version of PETSc from C++ and the C++ version from C.
41*/
42#if defined(__cplusplus)
43# define PETSC_FUNCTION_NAME__func__ PETSC_FUNCTION_NAME_CXX__func__
44#else
45# define PETSC_FUNCTION_NAME__func__ PETSC_FUNCTION_NAME_C__func__
46#endif
47
48/* ========================================================================== */
49/*
50 Since PETSc manages its own extern "C" handling users should never include PETSc include
51 files within extern "C". This will generate a compiler error if a user does put the include
52 file within an extern "C".
53*/
54#if defined(__cplusplus)
55void assert_never_put_petsc_headers_inside_an_extern_c(int); void assert_never_put_petsc_headers_inside_an_extern_c(double);
56#endif
57
58#if defined(__cplusplus)
59# define PETSC_RESTRICT__restrict PETSC_CXX_RESTRICT__restrict
60#else
61# define PETSC_RESTRICT__restrict PETSC_C_RESTRICT__restrict
62#endif
63
64#if defined(__cplusplus)
65# define PETSC_INLINEinline PETSC_CXX_INLINEinline
66#else
67# define PETSC_INLINEinline PETSC_C_INLINEinline
68#endif
69
70#define PETSC_STATIC_INLINEstatic inline static PETSC_INLINEinline
71
72#if defined(_WIN32) && defined(PETSC_USE_SHARED_LIBRARIES1) /* For Win32 shared libraries */
73# define PETSC_DLLEXPORT__attribute__((visibility ("default"))) __declspec(dllexport)
74# define PETSC_DLLIMPORT__attribute__((visibility ("default"))) __declspec(dllimport)
75# define PETSC_VISIBILITY_INTERNAL__attribute__((visibility ("hidden")))
76#elif defined(PETSC_USE_VISIBILITY_CXX1) && defined(__cplusplus)
77# define PETSC_DLLEXPORT__attribute__((visibility ("default"))) __attribute__((visibility ("default")))
78# define PETSC_DLLIMPORT__attribute__((visibility ("default"))) __attribute__((visibility ("default")))
79# define PETSC_VISIBILITY_INTERNAL__attribute__((visibility ("hidden"))) __attribute__((visibility ("hidden")))
80#elif defined(PETSC_USE_VISIBILITY_C1) && !defined(__cplusplus)
81# define PETSC_DLLEXPORT__attribute__((visibility ("default"))) __attribute__((visibility ("default")))
82# define PETSC_DLLIMPORT__attribute__((visibility ("default"))) __attribute__((visibility ("default")))
83# define PETSC_VISIBILITY_INTERNAL__attribute__((visibility ("hidden"))) __attribute__((visibility ("hidden")))
84#else
85# define PETSC_DLLEXPORT__attribute__((visibility ("default")))
86# define PETSC_DLLIMPORT__attribute__((visibility ("default")))
87# define PETSC_VISIBILITY_INTERNAL__attribute__((visibility ("hidden")))
88#endif
89
90#if defined(petsc_EXPORTS) /* CMake defines this when building the shared library */
91# define PETSC_VISIBILITY_PUBLIC__attribute__((visibility ("default"))) PETSC_DLLEXPORT__attribute__((visibility ("default")))
92#else /* Win32 users need this to import symbols from petsc.dll */
93# define PETSC_VISIBILITY_PUBLIC__attribute__((visibility ("default"))) PETSC_DLLIMPORT__attribute__((visibility ("default")))
94#endif
95
96/*
97 Functions tagged with PETSC_EXTERN in the header files are
98 always defined as extern "C" when compiled with C++ so they may be
99 used from C and are always visible in the shared libraries
100*/
101#if defined(__cplusplus)
102# define PETSC_EXTERNextern __attribute__((visibility ("default"))) extern "C" PETSC_VISIBILITY_PUBLIC__attribute__((visibility ("default")))
103# define PETSC_EXTERN_TYPEDEF extern "C"
104# define PETSC_INTERNextern __attribute__((visibility ("hidden"))) extern "C" PETSC_VISIBILITY_INTERNAL__attribute__((visibility ("hidden")))
105#else
106# define PETSC_EXTERNextern __attribute__((visibility ("default"))) extern PETSC_VISIBILITY_PUBLIC__attribute__((visibility ("default")))
107# define PETSC_EXTERN_TYPEDEF
108# define PETSC_INTERNextern __attribute__((visibility ("hidden"))) extern PETSC_VISIBILITY_INTERNAL__attribute__((visibility ("hidden")))
109#endif
110
111#include <petscversion.h>
112#define PETSC_AUTHOR_INFO" The PETSc Team\n [email protected]\n https://www.mcs.anl.gov/petsc/\n" " The PETSc Team\n [email protected]\n https://www.mcs.anl.gov/petsc/\n"
113
114/* ========================================================================== */
115
116/*
117 Defines the interface to MPI allowing the use of all MPI functions.
118
119 PETSc does not use the C++ binding of MPI at ALL. The following flag
120 makes sure the C++ bindings are not included. The C++ bindings REQUIRE
121 putting mpi.h before ANY C++ include files, we cannot control this
122 with all PETSc users. Users who want to use the MPI C++ bindings can include
123 mpicxx.h directly in their code
124*/
125#if !defined(MPICH_SKIP_MPICXX1)
126# define MPICH_SKIP_MPICXX1 1
127#endif
128#if !defined(OMPI_SKIP_MPICXX1)
129# define OMPI_SKIP_MPICXX1 1
130#endif
131#if defined(PETSC_HAVE_MPIUNI)
132# include <petsc/mpiuni/mpi.h>
133#else
134# include <mpi.h>
135#endif
136
137/*
138 Perform various sanity checks that the correct mpi.h is being included at compile time.
139 This usually happens because
140 * either an unexpected mpi.h is in the default compiler path (i.e. in /usr/include) or
141 * an extra include path -I/something (which contains the unexpected mpi.h) is being passed to the compiler
142*/
143#if defined(PETSC_HAVE_MPIUNI)
144# if !defined(MPIUNI_H)
145# error "PETSc was configured with --with-mpi=0 but now appears to be compiling using a different mpi.h"
146# endif
147#elif defined(PETSC_HAVE_I_MPI_NUMVERSION)
148# if !defined(I_MPI_NUMVERSION)
149# error "PETSc was configured with I_MPI but now appears to be compiling using a non-I_MPI mpi.h"
150# elif I_MPI_NUMVERSION != PETSC_HAVE_I_MPI_NUMVERSION
151# error "PETSc was configured with one I_MPI mpi.h version but now appears to be compiling using a different I_MPI mpi.h version"
152# endif
153#elif defined(PETSC_HAVE_MVAPICH2_NUMVERSION)
154# if !defined(MVAPICH2_NUMVERSION)
155# error "PETSc was configured with MVAPICH2 but now appears to be compiling using a non-MVAPICH2 mpi.h"
156# elif MVAPICH2_NUMVERSION != PETSC_HAVE_MVAPICH2_NUMVERSION
157# error "PETSc was configured with one MVAPICH2 mpi.h version but now appears to be compiling using a different MVAPICH2 mpi.h version"
158# endif
159#elif defined(PETSC_HAVE_MPICH_NUMVERSION30301300)
160# if !defined(MPICH_NUMVERSION30301300) || defined(MVAPICH2_NUMVERSION) || defined(I_MPI_NUMVERSION)
161# error "PETSc was configured with MPICH but now appears to be compiling using a non-MPICH mpi.h"
162# elif (MPICH_NUMVERSION30301300/100000 != PETSC_HAVE_MPICH_NUMVERSION30301300/100000) || (MPICH_NUMVERSION30301300%100000/1000 < PETSC_HAVE_MPICH_NUMVERSION30301300%100000/1000)
163# error "PETSc was configured with one MPICH mpi.h version but now appears to be compiling using a different MPICH mpi.h version"
164# endif
165#elif defined(PETSC_HAVE_OMPI_MAJOR_VERSION)
166# if !defined(OMPI_MAJOR_VERSION)
167# error "PETSc was configured with OpenMPI but now appears to be compiling using a non-OpenMPI mpi.h"
168# elif (OMPI_MAJOR_VERSION != PETSC_HAVE_OMPI_MAJOR_VERSION) || (OMPI_MINOR_VERSION != PETSC_HAVE_OMPI_MINOR_VERSION) || (OMPI_RELEASE_VERSION < PETSC_HAVE_OMPI_RELEASE_VERSION)
169# error "PETSc was configured with one OpenMPI mpi.h version but now appears to be compiling using a different OpenMPI mpi.h version"
170# endif
171#elif defined(OMPI_MAJOR_VERSION) || defined(MPICH_NUMVERSION30301300)
172# error "PETSc was configured with undetermined MPI - but now appears to be compiling using either of OpenMPI or a MPICH variant"
173#endif
174
175/*
176 Need to put stdio.h AFTER mpi.h for MPICH2 with C++ compiler
177 see the top of mpicxx.h in the MPICH2 distribution.
178*/
179#include <stdio.h>
180
181/* MSMPI on 32bit windows requires this yukky hack - that breaks MPI standard compliance */
182#if !defined(MPIAPI)
183#define MPIAPI
184#endif
185
186/*
187 Support for Clang (>=3.2) matching type tag arguments with void* buffer types.
188 This allows the compiler to detect cases where the MPI datatype argument passed to a MPI routine
189 does not match the actual type of the argument being passed in
190*/
191#if defined(__has_attribute)0 && defined(works_with_const_which_is_not_true)
192# if __has_attribute(argument_with_type_tag)1 && __has_attribute(pointer_with_type_tag)1 && __has_attribute(type_tag_for_datatype)1
193# define PetscAttrMPIPointerWithType(bufno,typeno) __attribute__((pointer_with_type_tag(MPI,bufno,typeno)))
194# define PetscAttrMPITypeTag(type) __attribute__((type_tag_for_datatype(MPI,type)))
195# define PetscAttrMPITypeTagLayoutCompatible(type) __attribute__((type_tag_for_datatype(MPI,type,layout_compatible)))
196# endif
197#endif
198#if !defined(PetscAttrMPIPointerWithType)
199# define PetscAttrMPIPointerWithType(bufno,typeno)
200# define PetscAttrMPITypeTag(type)
201# define PetscAttrMPITypeTagLayoutCompatible(type)
202#endif
203
204PETSC_EXTERNextern __attribute__((visibility ("default"))) MPI_Datatype MPIU_ENUM PetscAttrMPITypeTag(PetscEnum);
205PETSC_EXTERNextern __attribute__((visibility ("default"))) MPI_Datatype MPIU_BOOL PetscAttrMPITypeTag(PetscBool);
206
207/*MC
208 MPIU_INT - MPI datatype corresponding to PetscInt
209
210 Notes:
211 In MPI calls that require an MPI datatype that matches a PetscInt or array of PetscInt values, pass this value.
212
213 Level: beginner
214
215.seealso: PetscReal, PetscScalar, PetscComplex, PetscInt, MPIU_REAL, MPIU_SCALAR, MPIU_COMPLEX
216M*/
217
218#if defined(PETSC_HAVE_STDINT_H1) && defined(PETSC_HAVE_INTTYPES_H1) && defined(PETSC_HAVE_MPI_INT64_T1) /* MPI_INT64_T is not guaranteed to be a macro */
219# define MPIU_INT64((MPI_Datatype)0x4c00083a) MPI_INT64_T((MPI_Datatype)0x4c00083a)
220# define PetscInt64_FMT"l" "d" PRId64"l" "d"
221#elif (PETSC_SIZEOF_LONG_LONG8 == 8)
222# define MPIU_INT64((MPI_Datatype)0x4c00083a) MPI_LONG_LONG_INT((MPI_Datatype)0x4c000809)
223# define PetscInt64_FMT"l" "d" "lld"
224#elif defined(PETSC_HAVE___INT64)
225# define MPIU_INT64((MPI_Datatype)0x4c00083a) MPI_INT64_T((MPI_Datatype)0x4c00083a)
226# define PetscInt64_FMT"l" "d" "ld"
227#else
228# error "cannot determine PetscInt64 type"
229#endif
230
231PETSC_EXTERNextern __attribute__((visibility ("default"))) MPI_Datatype MPIU_FORTRANADDR;
232
233#if defined(PETSC_USE_64BIT_INDICES)
234# define MPIU_INT((MPI_Datatype)0x4c000405) MPIU_INT64((MPI_Datatype)0x4c00083a)
235# define PetscInt_FMT"d" PetscInt64_FMT"l" "d"
236#else
237# define MPIU_INT((MPI_Datatype)0x4c000405) MPI_INT((MPI_Datatype)0x4c000405)
238# define PetscInt_FMT"d" "d"
239#endif
240
241/*
242 For the rare cases when one needs to send a size_t object with MPI
243*/
244PETSC_EXTERNextern __attribute__((visibility ("default"))) MPI_Datatype MPIU_SIZE_T;
245
246/*
247 You can use PETSC_STDOUT as a replacement of stdout. You can also change
248 the value of PETSC_STDOUT to redirect all standard output elsewhere
249*/
250PETSC_EXTERNextern __attribute__((visibility ("default"))) FILE* PETSC_STDOUT;
251
252/*
253 You can use PETSC_STDERR as a replacement of stderr. You can also change
254 the value of PETSC_STDERR to redirect all standard error elsewhere
255*/
256PETSC_EXTERNextern __attribute__((visibility ("default"))) FILE* PETSC_STDERR;
257
258/*MC
259 PetscUnlikely - hints the compiler that the given condition is usually FALSE
260
261 Synopsis:
262 #include <petscsys.h>
263 PetscBool PetscUnlikely(PetscBool cond)
264
265 Not Collective
266
267 Input Parameters:
268. cond - condition or expression
269
270 Notes:
271 This returns the same truth value, it is only a hint to compilers that the resulting
272 branch is unlikely.
273
274 Level: advanced
275
276.seealso: PetscLikely(), CHKERRQ
277M*/
278
279/*MC
280 PetscLikely - hints the compiler that the given condition is usually TRUE
281
282 Synopsis:
283 #include <petscsys.h>
284 PetscBool PetscLikely(PetscBool cond)
285
286 Not Collective
287
288 Input Parameters:
289. cond - condition or expression
290
291 Notes:
292 This returns the same truth value, it is only a hint to compilers that the resulting
293 branch is likely.
294
295 Level: advanced
296
297.seealso: PetscUnlikely()
298M*/
299#if defined(PETSC_HAVE_BUILTIN_EXPECT1)
300# define PetscUnlikely(cond)__builtin_expect(!!(cond),0) __builtin_expect(!!(cond),0)
301# define PetscLikely(cond)__builtin_expect(!!(cond),1) __builtin_expect(!!(cond),1)
302#else
303# define PetscUnlikely(cond)__builtin_expect(!!(cond),0) (cond)
304# define PetscLikely(cond)__builtin_expect(!!(cond),1) (cond)
305#endif
306
307/*
308 Declare extern C stuff after including external header files
309*/
310
311PETSC_EXTERNextern __attribute__((visibility ("default"))) const char *const PetscBools[];
312
313/*
314 Defines elementary mathematics functions and constants.
315*/
316#include <petscmath.h>
317
318PETSC_EXTERNextern __attribute__((visibility ("default"))) const char *const PetscCopyModes[];
319
320/*MC
321 PETSC_IGNORE - same as NULL, means PETSc will ignore this argument
322
323 Level: beginner
324
325 Note:
326 Accepted by many PETSc functions to not set a parameter and instead use
327 some default
328
329 Fortran Notes:
330 This macro does not exist in Fortran; you must use PETSC_NULL_INTEGER,
331 PETSC_NULL_DOUBLE_PRECISION etc
332
333.seealso: PETSC_DECIDE, PETSC_DEFAULT, PETSC_DETERMINE
334
335M*/
336#define PETSC_IGNORE((void*)0) NULL((void*)0)
337
338/* This is deprecated */
339#define PETSC_NULL((void*)0) NULL((void*)0)
340
341/*MC
342 PETSC_DECIDE - standard way of passing in integer or floating point parameter
343 where you wish PETSc to use the default.
344
345 Level: beginner
346
347.seealso: PETSC_DEFAULT, PETSC_IGNORE, PETSC_DETERMINE
348
349M*/
350#define PETSC_DECIDE-1 -1
351
352/*MC
353 PETSC_DETERMINE - standard way of passing in integer or floating point parameter
354 where you wish PETSc to compute the required value.
355
356 Level: beginner
357
358 Developer Note:
359 I would like to use const PetscInt PETSC_DETERMINE = PETSC_DECIDE; but for
360 some reason this is not allowed by the standard even though PETSC_DECIDE is a constant value.
361
362.seealso: PETSC_DECIDE, PETSC_DEFAULT, PETSC_IGNORE, VecSetSizes()
363
364M*/
365#define PETSC_DETERMINE-1 PETSC_DECIDE-1
366
367/*MC
368 PETSC_DEFAULT - standard way of passing in integer or floating point parameter
369 where you wish PETSc to use the default.
370
371 Level: beginner
372
373 Fortran Notes:
374 You need to use PETSC_DEFAULT_INTEGER or PETSC_DEFAULT_REAL.
375
376.seealso: PETSC_DECIDE, PETSC_IGNORE, PETSC_DETERMINE
377
378M*/
379#define PETSC_DEFAULT-2 -2
380
381/*MC
382 PETSC_COMM_WORLD - the equivalent of the MPI_COMM_WORLD communicator which represents
383 all the processes that PETSc knows about.
384
385 Level: beginner
386
387 Notes:
388 By default PETSC_COMM_WORLD and MPI_COMM_WORLD are identical unless you wish to
389 run PETSc on ONLY a subset of MPI_COMM_WORLD. In that case create your new (smaller)
390 communicator, call it, say comm, and set PETSC_COMM_WORLD = comm BEFORE calling
391 PetscInitialize(), but after MPI_Init() has been called.
392
393 The value of PETSC_COMM_WORLD should never be USED/accessed before PetscInitialize()
394 is called because it may not have a valid value yet.
395
396.seealso: PETSC_COMM_SELF
397
398M*/
399PETSC_EXTERNextern __attribute__((visibility ("default"))) MPI_Comm PETSC_COMM_WORLD;
400
401/*MC
402 PETSC_COMM_SELF - This is always MPI_COMM_SELF
403
404 Level: beginner
405
406 Notes:
407 Do not USE/access or set this variable before PetscInitialize() has been called.
408
409.seealso: PETSC_COMM_WORLD
410
411M*/
412#define PETSC_COMM_SELF((MPI_Comm)0x44000001) MPI_COMM_SELF((MPI_Comm)0x44000001)
413
414PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscBool PetscBeganMPI;
415PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscBool PetscInitializeCalled;
416PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscBool PetscFinalizeCalled;
417PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscBool PetscViennaCLSynchronize;
418PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscBool PetscCUDASynchronize;
419
420PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSetHelpVersionFunctions(PetscErrorCode (*)(MPI_Comm),PetscErrorCode (*)(MPI_Comm));
421PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscCommDuplicate(MPI_Comm,MPI_Comm*,int*);
422PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscCommDestroy(MPI_Comm*);
423
424#if defined(PETSC_HAVE_CUDA)
425PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscCUDAInitialize(MPI_Comm);
426#endif
427
428/*MC
429 PetscMalloc - Allocates memory, One should use PetscNew(), PetscMalloc1() or PetscCalloc1() usually instead of this
430
431 Synopsis:
432 #include <petscsys.h>
433 PetscErrorCode PetscMalloc(size_t m,void **result)
434
435 Not Collective
436
437 Input Parameter:
438. m - number of bytes to allocate
439
440 Output Parameter:
441. result - memory allocated
442
443 Level: beginner
444
445 Notes:
446 Memory is always allocated at least double aligned
447
448 It is safe to allocate size 0 and pass the resulting pointer (which may or may not be NULL) to PetscFree().
449
450.seealso: PetscFree(), PetscNew()
451
452M*/
453#define PetscMalloc(a,b)((*PetscTrMalloc)((a),PETSC_FALSE,453,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,(void**)(b)))
((*PetscTrMalloc)((a),PETSC_FALSE,__LINE__453,PETSC_FUNCTION_NAME__func__,__FILE__"/sandbox/petsc/petsc.next/include/petscsys.h",(void**)(b)))
454
455/*MC
456 PetscRealloc - Rellocates memory
457
458 Synopsis:
459 #include <petscsys.h>
460 PetscErrorCode PetscRealloc(size_t m,void **result)
461
462 Not Collective
463
464 Input Parameters:
465+ m - number of bytes to allocate
466- result - initial memory allocated
467
468 Output Parameter:
469. result - new memory allocated
470
471 Level: developer
472
473 Notes:
474 Memory is always allocated at least double aligned
475
476.seealso: PetscMalloc(), PetscFree(), PetscNew()
477
478M*/
479#define PetscRealloc(a,b)((*PetscTrRealloc)((a),479,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,(void**)(b)))
((*PetscTrRealloc)((a),__LINE__479,PETSC_FUNCTION_NAME__func__,__FILE__"/sandbox/petsc/petsc.next/include/petscsys.h",(void**)(b)))
480
481/*MC
482 PetscAddrAlign - Rounds up an address to PETSC_MEMALIGN alignment
483
484 Synopsis:
485 #include <petscsys.h>
486 void *PetscAddrAlign(void *addr)
487
488 Not Collective
489
490 Input Parameters:
491. addr - address to align (any pointer type)
492
493 Level: developer
494
495.seealso: PetscMallocAlign()
496
497M*/
498#define PetscAddrAlign(a)(void*)((((uintptr_t)(a))+(16 -1)) & ~(16 -1)) (void*)((((PETSC_UINTPTR_Tuintptr_t)(a))+(PETSC_MEMALIGN16-1)) & ~(PETSC_MEMALIGN16-1))
499
500/*MC
501 PetscMalloc1 - Allocates an array of memory aligned to PETSC_MEMALIGN
502
503 Synopsis:
504 #include <petscsys.h>
505 PetscErrorCode PetscMalloc1(size_t m1,type **r1)
506
507 Not Collective
508
509 Input Parameter:
510. m1 - number of elements to allocate (may be zero)
511
512 Output Parameter:
513. r1 - memory allocated in first chunk
514
515 Note:
516 This uses the sizeof() of the memory type requested to determine the total memory to be allocated, therefore you should not
517 multiply the number of elements requested by the sizeof() the type. For example use
518$ PetscInt *id;
519$ PetscMalloc1(10,&id);
520 not
521$ PetscInt *id;
522$ PetscMalloc1(10*sizeof(PetscInt),&id);
523
524 Does not zero the memory allocatd, used PetscCalloc1() to obtain memory that has been zeroed.
525
526 Level: beginner
527
528.seealso: PetscFree(), PetscNew(), PetscMalloc(), PetscCalloc1(), PetscMalloc2()
529
530M*/
531#define PetscMalloc1(m1,r1)PetscMallocA(1,PETSC_FALSE,531,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,(size_t)(m1)*sizeof(**(r1)),(r1))
PetscMallocA(1,PETSC_FALSE,__LINE__531,PETSC_FUNCTION_NAME__func__,__FILE__"/sandbox/petsc/petsc.next/include/petscsys.h",(size_t)(m1)*sizeof(**(r1)),(r1))
532
533/*MC
534 PetscCalloc1 - Allocates a cleared (zeroed) array of memory aligned to PETSC_MEMALIGN
535
536 Synopsis:
537 #include <petscsys.h>
538 PetscErrorCode PetscCalloc1(size_t m1,type **r1)
539
540 Not Collective
541
542 Input Parameter:
543. m1 - number of elements to allocate in 1st chunk (may be zero)
544
545 Output Parameter:
546. r1 - memory allocated in first chunk
547
548 Notes:
549 See PetsMalloc1() for more details on usage.
550
551 Level: beginner
552
553.seealso: PetscFree(), PetscNew(), PetscMalloc(), PetscMalloc1(), PetscCalloc2()
554
555M*/
556#define PetscCalloc1(m1,r1)PetscMallocA(1,PETSC_TRUE,556,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,(size_t)(m1)*sizeof(**(r1)),(r1))
PetscMallocA(1,PETSC_TRUE,__LINE__556,PETSC_FUNCTION_NAME__func__,__FILE__"/sandbox/petsc/petsc.next/include/petscsys.h",(size_t)(m1)*sizeof(**(r1)),(r1))
557
558/*MC
559 PetscMalloc2 - Allocates 2 arrays of memory both aligned to PETSC_MEMALIGN
560
561 Synopsis:
562 #include <petscsys.h>
563 PetscErrorCode PetscMalloc2(size_t m1,type **r1,size_t m2,type **r2)
564
565 Not Collective
566
567 Input Parameter:
568+ m1 - number of elements to allocate in 1st chunk (may be zero)
569- m2 - number of elements to allocate in 2nd chunk (may be zero)
570
571 Output Parameter:
572+ r1 - memory allocated in first chunk
573- r2 - memory allocated in second chunk
574
575 Level: developer
576
577.seealso: PetscFree(), PetscNew(), PetscMalloc(), PetscMalloc1(), PetscCalloc2()
578
579M*/
580#define PetscMalloc2(m1,r1,m2,r2)PetscMallocA(2,PETSC_FALSE,580,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,(size_t)(m1)*sizeof(**(r1)),(r1),(size_t)(m2)*sizeof(**(r2))
,(r2))
PetscMallocA(2,PETSC_FALSE,__LINE__580,PETSC_FUNCTION_NAME__func__,__FILE__"/sandbox/petsc/petsc.next/include/petscsys.h",(size_t)(m1)*sizeof(**(r1)),(r1),(size_t)(m2)*sizeof(**(r2)),(r2))
581
582/*MC
583 PetscCalloc2 - Allocates 2 cleared (zeroed) arrays of memory both aligned to PETSC_MEMALIGN
584
585 Synopsis:
586 #include <petscsys.h>
587 PetscErrorCode PetscCalloc2(size_t m1,type **r1,size_t m2,type **r2)
588
589 Not Collective
590
591 Input Parameter:
592+ m1 - number of elements to allocate in 1st chunk (may be zero)
593- m2 - number of elements to allocate in 2nd chunk (may be zero)
594
595 Output Parameter:
596+ r1 - memory allocated in first chunk
597- r2 - memory allocated in second chunk
598
599 Level: developer
600
601.seealso: PetscFree(), PetscNew(), PetscMalloc(), PetscCalloc1(), PetscMalloc2()
602
603M*/
604#define PetscCalloc2(m1,r1,m2,r2)PetscMallocA(2,PETSC_TRUE,604,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,(size_t)(m1)*sizeof(**(r1)),(r1),(size_t)(m2)*sizeof(**(r2))
,(r2))
PetscMallocA(2,PETSC_TRUE,__LINE__604,PETSC_FUNCTION_NAME__func__,__FILE__"/sandbox/petsc/petsc.next/include/petscsys.h",(size_t)(m1)*sizeof(**(r1)),(r1),(size_t)(m2)*sizeof(**(r2)),(r2))
605
606/*MC
607 PetscMalloc3 - Allocates 3 arrays of memory, all aligned to PETSC_MEMALIGN
608
609 Synopsis:
610 #include <petscsys.h>
611 PetscErrorCode PetscMalloc3(size_t m1,type **r1,size_t m2,type **r2,size_t m3,type **r3)
612
613 Not Collective
614
615 Input Parameter:
616+ m1 - number of elements to allocate in 1st chunk (may be zero)
617. m2 - number of elements to allocate in 2nd chunk (may be zero)
618- m3 - number of elements to allocate in 3rd chunk (may be zero)
619
620 Output Parameter:
621+ r1 - memory allocated in first chunk
622. r2 - memory allocated in second chunk
623- r3 - memory allocated in third chunk
624
625 Level: developer
626
627.seealso: PetscFree(), PetscNew(), PetscMalloc(), PetscMalloc2(), PetscCalloc3(), PetscFree3()
628
629M*/
630#define PetscMalloc3(m1,r1,m2,r2,m3,r3)PetscMallocA(3,PETSC_FALSE,630,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,(size_t)(m1)*sizeof(**(r1)),(r1),(size_t)(m2)*sizeof(**(r2))
,(r2),(size_t)(m3)*sizeof(**(r3)),(r3))
PetscMallocA(3,PETSC_FALSE,__LINE__630,PETSC_FUNCTION_NAME__func__,__FILE__"/sandbox/petsc/petsc.next/include/petscsys.h",(size_t)(m1)*sizeof(**(r1)),(r1),(size_t)(m2)*sizeof(**(r2)),(r2),(size_t)(m3)*sizeof(**(r3)),(r3))
631
632/*MC
633 PetscCalloc3 - Allocates 3 cleared (zeroed) arrays of memory, all aligned to PETSC_MEMALIGN
634
635 Synopsis:
636 #include <petscsys.h>
637 PetscErrorCode PetscCalloc3(size_t m1,type **r1,size_t m2,type **r2,size_t m3,type **r3)
638
639 Not Collective
640
641 Input Parameter:
642+ m1 - number of elements to allocate in 1st chunk (may be zero)
643. m2 - number of elements to allocate in 2nd chunk (may be zero)
644- m3 - number of elements to allocate in 3rd chunk (may be zero)
645
646 Output Parameter:
647+ r1 - memory allocated in first chunk
648. r2 - memory allocated in second chunk
649- r3 - memory allocated in third chunk
650
651 Level: developer
652
653.seealso: PetscFree(), PetscNew(), PetscMalloc(), PetscCalloc2(), PetscMalloc3(), PetscFree3()
654
655M*/
656#define PetscCalloc3(m1,r1,m2,r2,m3,r3)PetscMallocA(3,PETSC_TRUE,656,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,(size_t)(m1)*sizeof(**(r1)),(r1),(size_t)(m2)*sizeof(**(r2))
,(r2),(size_t)(m3)*sizeof(**(r3)),(r3))
PetscMallocA(3,PETSC_TRUE,__LINE__656,PETSC_FUNCTION_NAME__func__,__FILE__"/sandbox/petsc/petsc.next/include/petscsys.h",(size_t)(m1)*sizeof(**(r1)),(r1),(size_t)(m2)*sizeof(**(r2)),(r2),(size_t)(m3)*sizeof(**(r3)),(r3))
657
658/*MC
659 PetscMalloc4 - Allocates 4 arrays of memory, all aligned to PETSC_MEMALIGN
660
661 Synopsis:
662 #include <petscsys.h>
663 PetscErrorCode PetscMalloc4(size_t m1,type **r1,size_t m2,type **r2,size_t m3,type **r3,size_t m4,type **r4)
664
665 Not Collective
666
667 Input Parameter:
668+ m1 - number of elements to allocate in 1st chunk (may be zero)
669. m2 - number of elements to allocate in 2nd chunk (may be zero)
670. m3 - number of elements to allocate in 3rd chunk (may be zero)
671- m4 - number of elements to allocate in 4th chunk (may be zero)
672
673 Output Parameter:
674+ r1 - memory allocated in first chunk
675. r2 - memory allocated in second chunk
676. r3 - memory allocated in third chunk
677- r4 - memory allocated in fourth chunk
678
679 Level: developer
680
681.seealso: PetscFree(), PetscNew(), PetscMalloc(), PetscMalloc2(), PetscCalloc4(), PetscFree4()
682
683M*/
684#define PetscMalloc4(m1,r1,m2,r2,m3,r3,m4,r4)PetscMallocA(4,PETSC_FALSE,684,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,(size_t)(m1)*sizeof(**(r1)),(r1),(size_t)(m2)*sizeof(**(r2))
,(r2),(size_t)(m3)*sizeof(**(r3)),(r3),(size_t)(m4)*sizeof(**
(r4)),(r4))
PetscMallocA(4,PETSC_FALSE,__LINE__684,PETSC_FUNCTION_NAME__func__,__FILE__"/sandbox/petsc/petsc.next/include/petscsys.h",(size_t)(m1)*sizeof(**(r1)),(r1),(size_t)(m2)*sizeof(**(r2)),(r2),(size_t)(m3)*sizeof(**(r3)),(r3),(size_t)(m4)*sizeof(**(r4)),(r4))
685
686/*MC
687 PetscCalloc4 - Allocates 4 cleared (zeroed) arrays of memory, all aligned to PETSC_MEMALIGN
688
689 Synopsis:
690 #include <petscsys.h>
691 PetscErrorCode PetscCalloc4(size_t m1,type **r1,size_t m2,type **r2,size_t m3,type **r3,size_t m4,type **r4)
692
693 Not Collective
694
695 Input Parameters:
696+ m1 - number of elements to allocate in 1st chunk (may be zero)
697. m2 - number of elements to allocate in 2nd chunk (may be zero)
698. m3 - number of elements to allocate in 3rd chunk (may be zero)
699- m4 - number of elements to allocate in 4th chunk (may be zero)
700
701 Output Parameters:
702+ r1 - memory allocated in first chunk
703. r2 - memory allocated in second chunk
704. r3 - memory allocated in third chunk
705- r4 - memory allocated in fourth chunk
706
707 Level: developer
708
709.seealso: PetscFree(), PetscNew(), PetscMalloc(), PetscMalloc2(), PetscCalloc4(), PetscFree4()
710
711M*/
712#define PetscCalloc4(m1,r1,m2,r2,m3,r3,m4,r4)PetscMallocA(4,PETSC_TRUE,712,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,(size_t)(m1)*sizeof(**(r1)),(r1),(size_t)(m2)*sizeof(**(r2))
,(r2),(size_t)(m3)*sizeof(**(r3)),(r3),(size_t)(m4)*sizeof(**
(r4)),(r4))
PetscMallocA(4,PETSC_TRUE,__LINE__712,PETSC_FUNCTION_NAME__func__,__FILE__"/sandbox/petsc/petsc.next/include/petscsys.h",(size_t)(m1)*sizeof(**(r1)),(r1),(size_t)(m2)*sizeof(**(r2)),(r2),(size_t)(m3)*sizeof(**(r3)),(r3),(size_t)(m4)*sizeof(**(r4)),(r4))
713
714/*MC
715 PetscMalloc5 - Allocates 5 arrays of memory, all aligned to PETSC_MEMALIGN
716
717 Synopsis:
718 #include <petscsys.h>
719 PetscErrorCode PetscMalloc5(size_t m1,type **r1,size_t m2,type **r2,size_t m3,type **r3,size_t m4,type **r4,size_t m5,type **r5)
720
721 Not Collective
722
723 Input Parameters:
724+ m1 - number of elements to allocate in 1st chunk (may be zero)
725. m2 - number of elements to allocate in 2nd chunk (may be zero)
726. m3 - number of elements to allocate in 3rd chunk (may be zero)
727. m4 - number of elements to allocate in 4th chunk (may be zero)
728- m5 - number of elements to allocate in 5th chunk (may be zero)
729
730 Output Parameters:
731+ r1 - memory allocated in first chunk
732. r2 - memory allocated in second chunk
733. r3 - memory allocated in third chunk
734. r4 - memory allocated in fourth chunk
735- r5 - memory allocated in fifth chunk
736
737 Level: developer
738
739.seealso: PetscFree(), PetscNew(), PetscMalloc(), PetscMalloc2(), PetscCalloc5(), PetscFree5()
740
741M*/
742#define PetscMalloc5(m1,r1,m2,r2,m3,r3,m4,r4,m5,r5)PetscMallocA(5,PETSC_FALSE,742,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,(size_t)(m1)*sizeof(**(r1)),(r1),(size_t)(m2)*sizeof(**(r2))
,(r2),(size_t)(m3)*sizeof(**(r3)),(r3),(size_t)(m4)*sizeof(**
(r4)),(r4),(size_t)(m5)*sizeof(**(r5)),(r5))
PetscMallocA(5,PETSC_FALSE,__LINE__742,PETSC_FUNCTION_NAME__func__,__FILE__"/sandbox/petsc/petsc.next/include/petscsys.h",(size_t)(m1)*sizeof(**(r1)),(r1),(size_t)(m2)*sizeof(**(r2)),(r2),(size_t)(m3)*sizeof(**(r3)),(r3),(size_t)(m4)*sizeof(**(r4)),(r4),(size_t)(m5)*sizeof(**(r5)),(r5))
743
744/*MC
745 PetscCalloc5 - Allocates 5 cleared (zeroed) arrays of memory, all aligned to PETSC_MEMALIGN
746
747 Synopsis:
748 #include <petscsys.h>
749 PetscErrorCode PetscCalloc5(size_t m1,type **r1,size_t m2,type **r2,size_t m3,type **r3,size_t m4,type **r4,size_t m5,type **r5)
750
751 Not Collective
752
753 Input Parameters:
754+ m1 - number of elements to allocate in 1st chunk (may be zero)
755. m2 - number of elements to allocate in 2nd chunk (may be zero)
756. m3 - number of elements to allocate in 3rd chunk (may be zero)
757. m4 - number of elements to allocate in 4th chunk (may be zero)
758- m5 - number of elements to allocate in 5th chunk (may be zero)
759
760 Output Parameters:
761+ r1 - memory allocated in first chunk
762. r2 - memory allocated in second chunk
763. r3 - memory allocated in third chunk
764. r4 - memory allocated in fourth chunk
765- r5 - memory allocated in fifth chunk
766
767 Level: developer
768
769.seealso: PetscFree(), PetscNew(), PetscMalloc(), PetscMalloc5(), PetscFree5()
770
771M*/
772#define PetscCalloc5(m1,r1,m2,r2,m3,r3,m4,r4,m5,r5)PetscMallocA(5,PETSC_TRUE,772,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,(size_t)(m1)*sizeof(**(r1)),(r1),(size_t)(m2)*sizeof(**(r2))
,(r2),(size_t)(m3)*sizeof(**(r3)),(r3),(size_t)(m4)*sizeof(**
(r4)),(r4),(size_t)(m5)*sizeof(**(r5)),(r5))
PetscMallocA(5,PETSC_TRUE,__LINE__772,PETSC_FUNCTION_NAME__func__,__FILE__"/sandbox/petsc/petsc.next/include/petscsys.h",(size_t)(m1)*sizeof(**(r1)),(r1),(size_t)(m2)*sizeof(**(r2)),(r2),(size_t)(m3)*sizeof(**(r3)),(r3),(size_t)(m4)*sizeof(**(r4)),(r4),(size_t)(m5)*sizeof(**(r5)),(r5))
773
774/*MC
775 PetscMalloc6 - Allocates 6 arrays of memory, all aligned to PETSC_MEMALIGN
776
777 Synopsis:
778 #include <petscsys.h>
779 PetscErrorCode PetscMalloc6(size_t m1,type **r1,size_t m2,type **r2,size_t m3,type **r3,size_t m4,type **r4,size_t m5,type **r5,size_t m6,type **r6)
780
781 Not Collective
782
783 Input Parameters:
784+ m1 - number of elements to allocate in 1st chunk (may be zero)
785. m2 - number of elements to allocate in 2nd chunk (may be zero)
786. m3 - number of elements to allocate in 3rd chunk (may be zero)
787. m4 - number of elements to allocate in 4th chunk (may be zero)
788. m5 - number of elements to allocate in 5th chunk (may be zero)
789- m6 - number of elements to allocate in 6th chunk (may be zero)
790
791 Output Parameteasr:
792+ r1 - memory allocated in first chunk
793. r2 - memory allocated in second chunk
794. r3 - memory allocated in third chunk
795. r4 - memory allocated in fourth chunk
796. r5 - memory allocated in fifth chunk
797- r6 - memory allocated in sixth chunk
798
799 Level: developer
800
801.seealso: PetscFree(), PetscNew(), PetscMalloc(), PetscMalloc2(), PetscCalloc6(), PetscFree3(), PetscFree4(), PetscFree5(), PetscFree6()
802
803M*/
804#define PetscMalloc6(m1,r1,m2,r2,m3,r3,m4,r4,m5,r5,m6,r6)PetscMallocA(6,PETSC_FALSE,804,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,(size_t)(m1)*sizeof(**(r1)),(r1),(size_t)(m2)*sizeof(**(r2))
,(r2),(size_t)(m3)*sizeof(**(r3)),(r3),(size_t)(m4)*sizeof(**
(r4)),(r4),(size_t)(m5)*sizeof(**(r5)),(r5),(size_t)(m6)*sizeof
(**(r6)),(r6))
PetscMallocA(6,PETSC_FALSE,__LINE__804,PETSC_FUNCTION_NAME__func__,__FILE__"/sandbox/petsc/petsc.next/include/petscsys.h",(size_t)(m1)*sizeof(**(r1)),(r1),(size_t)(m2)*sizeof(**(r2)),(r2),(size_t)(m3)*sizeof(**(r3)),(r3),(size_t)(m4)*sizeof(**(r4)),(r4),(size_t)(m5)*sizeof(**(r5)),(r5),(size_t)(m6)*sizeof(**(r6)),(r6))
805
806/*MC
807 PetscCalloc6 - Allocates 6 cleared (zeroed) arrays of memory, all aligned to PETSC_MEMALIGN
808
809 Synopsis:
810 #include <petscsys.h>
811 PetscErrorCode PetscCalloc6(size_t m1,type **r1,size_t m2,type **r2,size_t m3,type **r3,size_t m4,type **r4,size_t m5,type **r5,size_t m6,type **r6)
812
813 Not Collective
814
815 Input Parameters:
816+ m1 - number of elements to allocate in 1st chunk (may be zero)
817. m2 - number of elements to allocate in 2nd chunk (may be zero)
818. m3 - number of elements to allocate in 3rd chunk (may be zero)
819. m4 - number of elements to allocate in 4th chunk (may be zero)
820. m5 - number of elements to allocate in 5th chunk (may be zero)
821- m6 - number of elements to allocate in 6th chunk (may be zero)
822
823 Output Parameters:
824+ r1 - memory allocated in first chunk
825. r2 - memory allocated in second chunk
826. r3 - memory allocated in third chunk
827. r4 - memory allocated in fourth chunk
828. r5 - memory allocated in fifth chunk
829- r6 - memory allocated in sixth chunk
830
831 Level: developer
832
833.seealso: PetscFree(), PetscNew(), PetscMalloc(), PetscMalloc2(), PetscMalloc6(), PetscFree6()
834
835M*/
836#define PetscCalloc6(m1,r1,m2,r2,m3,r3,m4,r4,m5,r5,m6,r6)PetscMallocA(6,PETSC_TRUE,836,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,(size_t)(m1)*sizeof(**(r1)),(r1),(size_t)(m2)*sizeof(**(r2))
,(r2),(size_t)(m3)*sizeof(**(r3)),(r3),(size_t)(m4)*sizeof(**
(r4)),(r4),(size_t)(m5)*sizeof(**(r5)),(r5),(size_t)(m6)*sizeof
(**(r6)),(r6))
PetscMallocA(6,PETSC_TRUE,__LINE__836,PETSC_FUNCTION_NAME__func__,__FILE__"/sandbox/petsc/petsc.next/include/petscsys.h",(size_t)(m1)*sizeof(**(r1)),(r1),(size_t)(m2)*sizeof(**(r2)),(r2),(size_t)(m3)*sizeof(**(r3)),(r3),(size_t)(m4)*sizeof(**(r4)),(r4),(size_t)(m5)*sizeof(**(r5)),(r5),(size_t)(m6)*sizeof(**(r6)),(r6))
837
838/*MC
839 PetscMalloc7 - Allocates 7 arrays of memory, all aligned to PETSC_MEMALIGN
840
841 Synopsis:
842 #include <petscsys.h>
843 PetscErrorCode PetscMalloc7(size_t m1,type **r1,size_t m2,type **r2,size_t m3,type **r3,size_t m4,type **r4,size_t m5,type **r5,size_t m6,type **r6,size_t m7,type **r7)
844
845 Not Collective
846
847 Input Parameters:
848+ m1 - number of elements to allocate in 1st chunk (may be zero)
849. m2 - number of elements to allocate in 2nd chunk (may be zero)
850. m3 - number of elements to allocate in 3rd chunk (may be zero)
851. m4 - number of elements to allocate in 4th chunk (may be zero)
852. m5 - number of elements to allocate in 5th chunk (may be zero)
853. m6 - number of elements to allocate in 6th chunk (may be zero)
854- m7 - number of elements to allocate in 7th chunk (may be zero)
855
856 Output Parameters:
857+ r1 - memory allocated in first chunk
858. r2 - memory allocated in second chunk
859. r3 - memory allocated in third chunk
860. r4 - memory allocated in fourth chunk
861. r5 - memory allocated in fifth chunk
862. r6 - memory allocated in sixth chunk
863- r7 - memory allocated in seventh chunk
864
865 Level: developer
866
867.seealso: PetscFree(), PetscNew(), PetscMalloc(), PetscMalloc2(), PetscCalloc7(), PetscFree7()
868
869M*/
870#define PetscMalloc7(m1,r1,m2,r2,m3,r3,m4,r4,m5,r5,m6,r6,m7,r7)PetscMallocA(7,PETSC_FALSE,870,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,(size_t)(m1)*sizeof(**(r1)),(r1),(size_t)(m2)*sizeof(**(r2))
,(r2),(size_t)(m3)*sizeof(**(r3)),(r3),(size_t)(m4)*sizeof(**
(r4)),(r4),(size_t)(m5)*sizeof(**(r5)),(r5),(size_t)(m6)*sizeof
(**(r6)),(r6),(size_t)(m7)*sizeof(**(r7)),(r7))
PetscMallocA(7,PETSC_FALSE,__LINE__870,PETSC_FUNCTION_NAME__func__,__FILE__"/sandbox/petsc/petsc.next/include/petscsys.h",(size_t)(m1)*sizeof(**(r1)),(r1),(size_t)(m2)*sizeof(**(r2)),(r2),(size_t)(m3)*sizeof(**(r3)),(r3),(size_t)(m4)*sizeof(**(r4)),(r4),(size_t)(m5)*sizeof(**(r5)),(r5),(size_t)(m6)*sizeof(**(r6)),(r6),(size_t)(m7)*sizeof(**(r7)),(r7))
871
872/*MC
873 PetscCalloc7 - Allocates 7 cleared (zeroed) arrays of memory, all aligned to PETSC_MEMALIGN
874
875 Synopsis:
876 #include <petscsys.h>
877 PetscErrorCode PetscCalloc7(size_t m1,type **r1,size_t m2,type **r2,size_t m3,type **r3,size_t m4,type **r4,size_t m5,type **r5,size_t m6,type **r6,size_t m7,type **r7)
878
879 Not Collective
880
881 Input Parameters:
882+ m1 - number of elements to allocate in 1st chunk (may be zero)
883. m2 - number of elements to allocate in 2nd chunk (may be zero)
884. m3 - number of elements to allocate in 3rd chunk (may be zero)
885. m4 - number of elements to allocate in 4th chunk (may be zero)
886. m5 - number of elements to allocate in 5th chunk (may be zero)
887. m6 - number of elements to allocate in 6th chunk (may be zero)
888- m7 - number of elements to allocate in 7th chunk (may be zero)
889
890 Output Parameters:
891+ r1 - memory allocated in first chunk
892. r2 - memory allocated in second chunk
893. r3 - memory allocated in third chunk
894. r4 - memory allocated in fourth chunk
895. r5 - memory allocated in fifth chunk
896. r6 - memory allocated in sixth chunk
897- r7 - memory allocated in seventh chunk
898
899 Level: developer
900
901.seealso: PetscFree(), PetscNew(), PetscMalloc(), PetscMalloc2(), PetscMalloc7(), PetscFree7()
902
903M*/
904#define PetscCalloc7(m1,r1,m2,r2,m3,r3,m4,r4,m5,r5,m6,r6,m7,r7)PetscMallocA(7,PETSC_TRUE,904,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,(size_t)(m1)*sizeof(**(r1)),(r1),(size_t)(m2)*sizeof(**(r2))
,(r2),(size_t)(m3)*sizeof(**(r3)),(r3),(size_t)(m4)*sizeof(**
(r4)),(r4),(size_t)(m5)*sizeof(**(r5)),(r5),(size_t)(m6)*sizeof
(**(r6)),(r6),(size_t)(m7)*sizeof(**(r7)),(r7))
PetscMallocA(7,PETSC_TRUE,__LINE__904,PETSC_FUNCTION_NAME__func__,__FILE__"/sandbox/petsc/petsc.next/include/petscsys.h",(size_t)(m1)*sizeof(**(r1)),(r1),(size_t)(m2)*sizeof(**(r2)),(r2),(size_t)(m3)*sizeof(**(r3)),(r3),(size_t)(m4)*sizeof(**(r4)),(r4),(size_t)(m5)*sizeof(**(r5)),(r5),(size_t)(m6)*sizeof(**(r6)),(r6),(size_t)(m7)*sizeof(**(r7)),(r7))
905
906/*MC
907 PetscNew - Allocates memory of a particular type, zeros the memory! Aligned to PETSC_MEMALIGN
908
909 Synopsis:
910 #include <petscsys.h>
911 PetscErrorCode PetscNew(type **result)
912
913 Not Collective
914
915 Output Parameter:
916. result - memory allocated, sized to match pointer type
917
918 Level: beginner
919
920.seealso: PetscFree(), PetscMalloc(), PetscNewLog(), PetscCalloc1(), PetscMalloc1()
921
922M*/
923#define PetscNew(b)PetscMallocA(1,PETSC_TRUE,923,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,(size_t)(1)*sizeof(**((b))),((b)))
PetscCalloc1(1,(b))PetscMallocA(1,PETSC_TRUE,923,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,(size_t)(1)*sizeof(**((b))),((b)))
924
925/*MC
926 PetscNewLog - Allocates memory of a type matching pointer, zeros the memory! Aligned to PETSC_MEMALIGN. Associates the memory allocated
927 with the given object using PetscLogObjectMemory().
928
929 Synopsis:
930 #include <petscsys.h>
931 PetscErrorCode PetscNewLog(PetscObject obj,type **result)
932
933 Not Collective
934
935 Input Parameter:
936. obj - object memory is logged to
937
938 Output Parameter:
939. result - memory allocated, sized to match pointer type
940
941 Level: developer
942
943.seealso: PetscFree(), PetscMalloc(), PetscNew(), PetscLogObjectMemory(), PetscCalloc1(), PetscMalloc1()
944
945M*/
946#define PetscNewLog(o,b)(PetscMallocA(1,PETSC_TRUE,946,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,(size_t)(1)*sizeof(**(((b)))),(((b)))) || PetscLogObjectMemory
((PetscObject)o,sizeof(**(b))))
(PetscNew((b))PetscMallocA(1,PETSC_TRUE,946,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,(size_t)(1)*sizeof(**(((b)))),(((b))))
|| PetscLogObjectMemory((PetscObject)o,sizeof(**(b))))
947
948/*MC
949 PetscFree - Frees memory
950
951 Synopsis:
952 #include <petscsys.h>
953 PetscErrorCode PetscFree(void *memory)
954
955 Not Collective
956
957 Input Parameter:
958. memory - memory to free (the pointer is ALWAYS set to NULL upon sucess)
959
960 Level: beginner
961
962 Notes:
963 Do not free memory obtained with PetscMalloc2(), PetscCalloc2() etc, they must be freed with PetscFree2() etc.
964
965 It is safe to call PetscFree() on a NULL pointer.
966
967.seealso: PetscNew(), PetscMalloc(), PetscNewLog(), PetscMalloc1(), PetscCalloc1()
968
969M*/
970#define PetscFree(a)((*PetscTrFree)((void*)(a),970,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
) || ((a) = 0,0))
((*PetscTrFree)((void*)(a),__LINE__970,PETSC_FUNCTION_NAME__func__,__FILE__"/sandbox/petsc/petsc.next/include/petscsys.h") || ((a) = 0,0))
971
972/*MC
973 PetscFree2 - Frees 2 chunks of memory obtained with PetscMalloc2()
974
975 Synopsis:
976 #include <petscsys.h>
977 PetscErrorCode PetscFree2(void *memory1,void *memory2)
978
979 Not Collective
980
981 Input Parameters:
982+ memory1 - memory to free
983- memory2 - 2nd memory to free
984
985 Level: developer
986
987 Notes:
988 Memory must have been obtained with PetscMalloc2()
989
990.seealso: PetscNew(), PetscMalloc(), PetscMalloc2(), PetscFree()
991
992M*/
993#define PetscFree2(m1,m2)PetscFreeA(2,993,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,&(m1),&(m2))
PetscFreeA(2,__LINE__993,PETSC_FUNCTION_NAME__func__,__FILE__"/sandbox/petsc/petsc.next/include/petscsys.h",&(m1),&(m2))
994
995/*MC
996 PetscFree3 - Frees 3 chunks of memory obtained with PetscMalloc3()
997
998 Synopsis:
999 #include <petscsys.h>
1000 PetscErrorCode PetscFree3(void *memory1,void *memory2,void *memory3)
1001
1002 Not Collective
1003
1004 Input Parameters:
1005+ memory1 - memory to free
1006. memory2 - 2nd memory to free
1007- memory3 - 3rd memory to free
1008
1009 Level: developer
1010
1011 Notes:
1012 Memory must have been obtained with PetscMalloc3()
1013
1014.seealso: PetscNew(), PetscMalloc(), PetscMalloc2(), PetscFree(), PetscMalloc3()
1015
1016M*/
1017#define PetscFree3(m1,m2,m3)PetscFreeA(3,1017,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,&(m1),&(m2),&(m3))
PetscFreeA(3,__LINE__1017,PETSC_FUNCTION_NAME__func__,__FILE__"/sandbox/petsc/petsc.next/include/petscsys.h",&(m1),&(m2),&(m3))
1018
1019/*MC
1020 PetscFree4 - Frees 4 chunks of memory obtained with PetscMalloc4()
1021
1022 Synopsis:
1023 #include <petscsys.h>
1024 PetscErrorCode PetscFree4(void *m1,void *m2,void *m3,void *m4)
1025
1026 Not Collective
1027
1028 Input Parameters:
1029+ m1 - memory to free
1030. m2 - 2nd memory to free
1031. m3 - 3rd memory to free
1032- m4 - 4th memory to free
1033
1034 Level: developer
1035
1036 Notes:
1037 Memory must have been obtained with PetscMalloc4()
1038
1039.seealso: PetscNew(), PetscMalloc(), PetscMalloc2(), PetscFree(), PetscMalloc3(), PetscMalloc4()
1040
1041M*/
1042#define PetscFree4(m1,m2,m3,m4)PetscFreeA(4,1042,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,&(m1),&(m2),&(m3),&(m4))
PetscFreeA(4,__LINE__1042,PETSC_FUNCTION_NAME__func__,__FILE__"/sandbox/petsc/petsc.next/include/petscsys.h",&(m1),&(m2),&(m3),&(m4))
1043
1044/*MC
1045 PetscFree5 - Frees 5 chunks of memory obtained with PetscMalloc5()
1046
1047 Synopsis:
1048 #include <petscsys.h>
1049 PetscErrorCode PetscFree5(void *m1,void *m2,void *m3,void *m4,void *m5)
1050
1051 Not Collective
1052
1053 Input Parameters:
1054+ m1 - memory to free
1055. m2 - 2nd memory to free
1056. m3 - 3rd memory to free
1057. m4 - 4th memory to free
1058- m5 - 5th memory to free
1059
1060 Level: developer
1061
1062 Notes:
1063 Memory must have been obtained with PetscMalloc5()
1064
1065.seealso: PetscNew(), PetscMalloc(), PetscMalloc2(), PetscFree(), PetscMalloc3(), PetscMalloc4(), PetscMalloc5()
1066
1067M*/
1068#define PetscFree5(m1,m2,m3,m4,m5)PetscFreeA(5,1068,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,&(m1),&(m2),&(m3),&(m4),&(m5))
PetscFreeA(5,__LINE__1068,PETSC_FUNCTION_NAME__func__,__FILE__"/sandbox/petsc/petsc.next/include/petscsys.h",&(m1),&(m2),&(m3),&(m4),&(m5))
1069
1070/*MC
1071 PetscFree6 - Frees 6 chunks of memory obtained with PetscMalloc6()
1072
1073 Synopsis:
1074 #include <petscsys.h>
1075 PetscErrorCode PetscFree6(void *m1,void *m2,void *m3,void *m4,void *m5,void *m6)
1076
1077 Not Collective
1078
1079 Input Parameters:
1080+ m1 - memory to free
1081. m2 - 2nd memory to free
1082. m3 - 3rd memory to free
1083. m4 - 4th memory to free
1084. m5 - 5th memory to free
1085- m6 - 6th memory to free
1086
1087
1088 Level: developer
1089
1090 Notes:
1091 Memory must have been obtained with PetscMalloc6()
1092
1093.seealso: PetscNew(), PetscMalloc(), PetscMalloc2(), PetscFree(), PetscMalloc3(), PetscMalloc4(), PetscMalloc5(), PetscMalloc6()
1094
1095M*/
1096#define PetscFree6(m1,m2,m3,m4,m5,m6)PetscFreeA(6,1096,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,&(m1),&(m2),&(m3),&(m4),&(m5),&(m6))
PetscFreeA(6,__LINE__1096,PETSC_FUNCTION_NAME__func__,__FILE__"/sandbox/petsc/petsc.next/include/petscsys.h",&(m1),&(m2),&(m3),&(m4),&(m5),&(m6))
1097
1098/*MC
1099 PetscFree7 - Frees 7 chunks of memory obtained with PetscMalloc7()
1100
1101 Synopsis:
1102 #include <petscsys.h>
1103 PetscErrorCode PetscFree7(void *m1,void *m2,void *m3,void *m4,void *m5,void *m6,void *m7)
1104
1105 Not Collective
1106
1107 Input Parameters:
1108+ m1 - memory to free
1109. m2 - 2nd memory to free
1110. m3 - 3rd memory to free
1111. m4 - 4th memory to free
1112. m5 - 5th memory to free
1113. m6 - 6th memory to free
1114- m7 - 7th memory to free
1115
1116
1117 Level: developer
1118
1119 Notes:
1120 Memory must have been obtained with PetscMalloc7()
1121
1122.seealso: PetscNew(), PetscMalloc(), PetscMalloc2(), PetscFree(), PetscMalloc3(), PetscMalloc4(), PetscMalloc5(), PetscMalloc6(),
1123 PetscMalloc7()
1124
1125M*/
1126#define PetscFree7(m1,m2,m3,m4,m5,m6,m7)PetscFreeA(7,1126,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,&(m1),&(m2),&(m3),&(m4),&(m5),&(m6),
&(m7))
PetscFreeA(7,__LINE__1126,PETSC_FUNCTION_NAME__func__,__FILE__"/sandbox/petsc/petsc.next/include/petscsys.h",&(m1),&(m2),&(m3),&(m4),&(m5),&(m6),&(m7))
1127
1128PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscMallocA(int,PetscBool,int,const char *,const char *,size_t,void *,...);
1129PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscFreeA(int,int,const char *,const char *,void *,...);
1130PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode (*PetscTrMalloc)(size_t,PetscBool,int,const char[],const char[],void**);
1131PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode (*PetscTrFree)(void*,int,const char[],const char[]);
1132PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode (*PetscTrRealloc)(size_t,int,const char[],const char[],void**);
1133PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscMallocSetCoalesce(PetscBool);
1134PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscMallocSet(PetscErrorCode (*)(size_t,PetscBool,int,const char[],const char[],void**),PetscErrorCode (*)(void*,int,const char[],const char[]));
1135PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscMallocClear(void);
1136
1137/*
1138 Unlike PetscMallocSet and PetscMallocClear which overwrite the existing settings, these two functions save the previous choice of allocator, and should be used in pair.
1139*/
1140PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscMallocSetDRAM(void);
1141PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscMallocResetDRAM(void);
1142
1143#define MPIU_PETSCLOGDOUBLE((MPI_Datatype)0x4c00080b) MPI_DOUBLE((MPI_Datatype)0x4c00080b)
1144#define MPIU_2PETSCLOGDOUBLE((MPI_Datatype)1275072547) MPI_2DOUBLE_PRECISION((MPI_Datatype)1275072547)
1145
1146/*
1147 Routines for tracing memory corruption/bleeding with default PETSc memory allocation
1148*/
1149PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscMallocDump(FILE *);
1150PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscMallocDumpLog(FILE *);
1151PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscMallocGetCurrentUsage(PetscLogDouble *);
1152PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscMallocGetMaximumUsage(PetscLogDouble *);
1153PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscMallocPushMaximumUsage(int);
1154PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscMallocPopMaximumUsage(int,PetscLogDouble*);
1155PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscMallocDebug(PetscBool);
1156PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscMallocGetDebug(PetscBool*);
1157PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscMallocValidate(int,const char[],const char[]);
1158PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscMallocSetDumpLog(void);
1159PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscMallocSetDumpLogThreshold(PetscLogDouble);
1160PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscMallocGetDumpLog(PetscBool*);
1161
1162PETSC_EXTERNextern __attribute__((visibility ("default"))) const char *const PetscDataTypes[];
1163PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscDataTypeToMPIDataType(PetscDataType,MPI_Datatype*);
1164PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscMPIDataTypeToPetscDataType(MPI_Datatype,PetscDataType*);
1165PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscDataTypeGetSize(PetscDataType,size_t*);
1166PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscDataTypeFromString(const char*,PetscDataType*,PetscBool*);
1167
1168/*
1169 Basic memory and string operations. These are usually simple wrappers
1170 around the basic Unix system calls, but a few of them have additional
1171 functionality and/or error checking.
1172*/
1173PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscBitMemcpy(void*,PetscInt,const void*,PetscInt,PetscInt,PetscDataType);
1174PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscMemcmp(const void*,const void*,size_t,PetscBool *);
1175PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscStrlen(const char[],size_t*);
1176PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscStrToArray(const char[],char,int*,char ***);
1177PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscStrToArrayDestroy(int,char **);
1178PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscStrcmp(const char[],const char[],PetscBool *);
1179PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscStrgrt(const char[],const char[],PetscBool *);
1180PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscStrcasecmp(const char[],const char[],PetscBool *);
1181PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscStrncmp(const char[],const char[],size_t,PetscBool *);
1182PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscStrcpy(char[],const char[]);
1183PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscStrcat(char[],const char[]);
1184PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscStrlcat(char[],const char[],size_t);
1185PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscStrncpy(char[],const char[],size_t);
1186PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscStrchr(const char[],char,char *[]);
1187PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscStrtolower(char[]);
1188PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscStrtoupper(char[]);
1189PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscStrrchr(const char[],char,char *[]);
1190PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscStrstr(const char[],const char[],char *[]);
1191PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscStrrstr(const char[],const char[],char *[]);
1192PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscStrendswith(const char[],const char[],PetscBool*);
1193PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscStrbeginswith(const char[],const char[],PetscBool*);
1194PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscStrendswithwhich(const char[],const char *const*,PetscInt*);
1195PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscStrallocpy(const char[],char *[]);
1196PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscStrArrayallocpy(const char *const*,char***);
1197PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscStrArrayDestroy(char***);
1198PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscStrNArrayallocpy(PetscInt,const char *const*,char***);
1199PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscStrNArrayDestroy(PetscInt,char***);
1200PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscStrreplace(MPI_Comm,const char[],char[],size_t);
1201
1202PETSC_EXTERNextern __attribute__((visibility ("default"))) void PetscStrcmpNoError(const char[],const char[],PetscBool *);
1203
1204PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscTokenCreate(const char[],const char,PetscToken*);
1205PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscTokenFind(PetscToken,char *[]);
1206PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscTokenDestroy(PetscToken*);
1207PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscStrInList(const char[],const char[],char,PetscBool*);
1208
1209PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscEListFind(PetscInt,const char *const*,const char*,PetscInt*,PetscBool*);
1210PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscEnumFind(const char *const*,const char*,PetscEnum*,PetscBool*);
1211
1212/*
1213 These are MPI operations for MPI_Allreduce() etc
1214*/
1215PETSC_EXTERNextern __attribute__((visibility ("default"))) MPI_Op MPIU_MAXSUM_OP;
1216#if (defined(PETSC_HAVE_COMPLEX1) && !defined(PETSC_HAVE_MPI_C_DOUBLE_COMPLEX1)) || defined(PETSC_USE_REAL___FLOAT128) || defined(PETSC_USE_REAL___FP16)
1217PETSC_EXTERNextern __attribute__((visibility ("default"))) MPI_Op MPIU_SUM(MPI_Op)(0x58000003);
1218#else
1219#define MPIU_SUM(MPI_Op)(0x58000003) MPI_SUM(MPI_Op)(0x58000003)
1220#endif
1221#if defined(PETSC_USE_REAL___FLOAT128) || defined(PETSC_USE_REAL___FP16)
1222PETSC_EXTERNextern __attribute__((visibility ("default"))) MPI_Op MPIU_MAX(MPI_Op)(0x58000001);
1223PETSC_EXTERNextern __attribute__((visibility ("default"))) MPI_Op MPIU_MIN(MPI_Op)(0x58000002);
1224#else
1225#define MPIU_MAX(MPI_Op)(0x58000001) MPI_MAX(MPI_Op)(0x58000001)
1226#define MPIU_MIN(MPI_Op)(0x58000002) MPI_MIN(MPI_Op)(0x58000002)
1227#endif
1228PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscMaxSum(MPI_Comm,const PetscInt[],PetscInt*,PetscInt*);
1229
1230PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode MPIULong_Send(void*,PetscInt,MPI_Datatype,PetscMPIInt,PetscMPIInt,MPI_Comm);
1231PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode MPIULong_Recv(void*,PetscInt,MPI_Datatype,PetscMPIInt,PetscMPIInt,MPI_Comm);
1232
1233PETSC_EXTERNextern __attribute__((visibility ("default"))) const char *const PetscFileModes[];
1234
1235/*
1236 Defines PETSc error handling.
1237*/
1238#include <petscerror.h>
1239
1240#define PETSC_SMALLEST_CLASSID1211211 1211211
1241PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscClassId PETSC_LARGEST_CLASSID;
1242PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscClassId PETSC_OBJECT_CLASSID;
1243PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscClassIdRegister(const char[],PetscClassId *);
1244
1245/*
1246 Routines that get memory usage information from the OS
1247*/
1248PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscMemoryGetCurrentUsage(PetscLogDouble *);
1249PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscMemoryGetMaximumUsage(PetscLogDouble *);
1250PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscMemorySetGetMaximumUsage(void);
1251PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscMemoryTrace(const char[]);
1252
1253PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscInfoAllow(PetscBool ,const char []);
1254PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSleep(PetscReal);
1255
1256/*
1257 Initialization of PETSc
1258*/
1259PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscInitialize(int*,char***,const char[],const char[]);
1260PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscInitializeNoPointers(int,char**,const char[],const char[]);
1261PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscInitializeNoArguments(void);
1262PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscInitialized(PetscBool *);
1263PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscFinalized(PetscBool *);
1264PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscFinalize(void);
1265PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscInitializeFortran(void);
1266PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscGetArgs(int*,char ***);
1267PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscGetArguments(char ***);
1268PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscFreeArguments(char **);
1269
1270PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscEnd(void);
1271PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSysInitializePackage(void);
1272
1273PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscPythonInitialize(const char[],const char[]);
1274PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscPythonFinalize(void);
1275PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscPythonPrintError(void);
1276PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscPythonMonitorSet(PetscObject,const char[]);
1277
1278/*
1279 These are so that in extern C code we can caste function pointers to non-extern C
1280 function pointers. Since the regular C++ code expects its function pointers to be C++
1281*/
1282PETSC_EXTERN_TYPEDEF typedef void (**PetscVoidStarFunction)(void);
1283PETSC_EXTERN_TYPEDEF typedef void (*PetscVoidFunction)(void);
1284PETSC_EXTERN_TYPEDEF typedef PetscErrorCode (*PetscErrorCodeFunction)(void);
1285
1286/*
1287 Functions that can act on any PETSc object.
1288*/
1289PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectDestroy(PetscObject*);
1290PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectGetComm(PetscObject,MPI_Comm *);
1291PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectGetClassId(PetscObject,PetscClassId *);
1292PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectGetClassName(PetscObject,const char *[]);
1293PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectSetType(PetscObject,const char []);
1294PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectGetType(PetscObject,const char *[]);
1295PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectSetName(PetscObject,const char[]);
1296PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectGetName(PetscObject,const char*[]);
1297PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectSetTabLevel(PetscObject,PetscInt);
1298PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectGetTabLevel(PetscObject,PetscInt*);
1299PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectIncrementTabLevel(PetscObject,PetscObject,PetscInt);
1300PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectReference(PetscObject);
1301PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectGetReference(PetscObject,PetscInt*);
1302PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectDereference(PetscObject);
1303PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectGetNewTag(PetscObject,PetscMPIInt *);
1304PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectCompose(PetscObject,const char[],PetscObject);
1305PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectRemoveReference(PetscObject,const char[]);
1306PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectQuery(PetscObject,const char[],PetscObject *);
1307PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectComposeFunction_Private(PetscObject,const char[],void (*)(void));
1308#define PetscObjectComposeFunction(a,b,d)PetscObjectComposeFunction_Private(a,b,(PetscVoidFunction)(d)
)
PetscObjectComposeFunction_Private(a,b,(PetscVoidFunction)(d))
1309PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectSetFromOptions(PetscObject);
1310PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectSetUp(PetscObject);
1311PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectSetPrintedOptions(PetscObject);
1312PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectInheritPrintedOptions(PetscObject,PetscObject);
1313PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscCommGetNewTag(MPI_Comm,PetscMPIInt *);
1314
1315#include <petscviewertypes.h>
1316#include <petscoptions.h>
1317
1318PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectsListGetGlobalNumbering(MPI_Comm,PetscInt,PetscObject*,PetscInt*,PetscInt*);
1319
1320PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscMemoryShowUsage(PetscViewer,const char[]);
1321PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscMemoryView(PetscViewer,const char[]);
1322PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectPrintClassNamePrefixType(PetscObject,PetscViewer);
1323PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectView(PetscObject,PetscViewer);
1324#define PetscObjectQueryFunction(obj,name,fptr)PetscObjectQueryFunction_Private((obj),(name),(PetscVoidFunction
*)(fptr))
PetscObjectQueryFunction_Private((obj),(name),(PetscVoidFunction*)(fptr))
1325PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectQueryFunction_Private(PetscObject,const char[],void (**)(void));
1326PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectSetOptionsPrefix(PetscObject,const char[]);
1327PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectAppendOptionsPrefix(PetscObject,const char[]);
1328PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectPrependOptionsPrefix(PetscObject,const char[]);
1329PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectGetOptionsPrefix(PetscObject,const char*[]);
1330PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectChangeTypeName(PetscObject,const char[]);
1331PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectRegisterDestroy(PetscObject);
1332PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectRegisterDestroyAll(void);
1333PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectViewFromOptions(PetscObject,PetscObject,const char[]);
1334PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectName(PetscObject);
1335PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectTypeCompare(PetscObject,const char[],PetscBool *);
1336PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectBaseTypeCompare(PetscObject,const char[],PetscBool *);
1337PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectTypeCompareAny(PetscObject,PetscBool*,const char[],...);
1338PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectBaseTypeCompareAny(PetscObject,PetscBool*,const char[],...);
1339PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscRegisterFinalize(PetscErrorCode (*)(void));
1340PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscRegisterFinalizeAll(void);
1341
1342#if defined(PETSC_HAVE_SAWS)
1343PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSAWsBlock(void);
1344PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectSAWsViewOff(PetscObject)0;
1345PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectSAWsSetBlock(PetscObject,PetscBool)0;
1346PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectSAWsBlock(PetscObject)0;
1347PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectSAWsGrantAccess(PetscObject)0;
1348PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectSAWsTakeAccess(PetscObject)0;
1349PETSC_EXTERNextern __attribute__((visibility ("default"))) void PetscStackSAWsGrantAccess(void);
1350PETSC_EXTERNextern __attribute__((visibility ("default"))) void PetscStackSAWsTakeAccess(void);
1351PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscStackViewSAWs(void);
1352PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscStackSAWsViewOff(void);
1353
1354#else
1355#define PetscSAWsBlock()0 0
1356#define PetscObjectSAWsViewOff(obj)0 0
1357#define PetscObjectSAWsSetBlock(obj,flg)0 0
1358#define PetscObjectSAWsBlock(obj)0 0
1359#define PetscObjectSAWsGrantAccess(obj)0 0
1360#define PetscObjectSAWsTakeAccess(obj)0 0
1361#define PetscStackViewSAWs()0 0
1362#define PetscStackSAWsViewOff()0 0
1363#define PetscStackSAWsTakeAccess()
1364#define PetscStackSAWsGrantAccess()
1365
1366#endif
1367
1368PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscDLOpen(const char[],PetscDLMode,PetscDLHandle *);
1369PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscDLClose(PetscDLHandle *);
1370PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscDLSym(PetscDLHandle,const char[],void **);
1371
1372#if defined(PETSC_USE_DEBUG1)
1373PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscMallocGetStack(void*,PetscStack**);
1374#endif
1375PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectsDump(FILE*,PetscBool);
1376
1377PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectListDestroy(PetscObjectList*);
1378PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectListFind(PetscObjectList,const char[],PetscObject*);
1379PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectListReverseFind(PetscObjectList,PetscObject,char**,PetscBool*);
1380PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectListAdd(PetscObjectList *,const char[],PetscObject);
1381PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectListRemoveReference(PetscObjectList *,const char[]);
1382PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectListDuplicate(PetscObjectList,PetscObjectList *);
1383
1384/*
1385 Dynamic library lists. Lists of names of routines in objects or in dynamic
1386 link libraries that will be loaded as needed.
1387*/
1388
1389#define PetscFunctionListAdd(list,name,fptr)PetscFunctionListAdd_Private((list),(name),(PetscVoidFunction
)(fptr))
PetscFunctionListAdd_Private((list),(name),(PetscVoidFunction)(fptr))
1390PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscFunctionListAdd_Private(PetscFunctionList*,const char[],void (*)(void));
1391PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscFunctionListDestroy(PetscFunctionList*);
1392#define PetscFunctionListFind(list,name,fptr)PetscFunctionListFind_Private((list),(name),(PetscVoidFunction
*)(fptr))
PetscFunctionListFind_Private((list),(name),(PetscVoidFunction*)(fptr))
1393PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscFunctionListFind_Private(PetscFunctionList,const char[],void (**)(void));
1394PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscFunctionListPrintTypes(MPI_Comm,FILE*,const char[],const char[],const char[],const char[],PetscFunctionList,const char[],const char[]);
1395PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscFunctionListDuplicate(PetscFunctionList,PetscFunctionList *);
1396PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscFunctionListView(PetscFunctionList,PetscViewer);
1397PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscFunctionListGet(PetscFunctionList,const char ***,int*);
1398
1399PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscDLLibrary PetscDLLibrariesLoaded;
1400PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscDLLibraryAppend(MPI_Comm,PetscDLLibrary *,const char[]);
1401PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscDLLibraryPrepend(MPI_Comm,PetscDLLibrary *,const char[]);
1402PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscDLLibrarySym(MPI_Comm,PetscDLLibrary *,const char[],const char[],void **);
1403PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscDLLibraryPrintPath(PetscDLLibrary);
1404PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscDLLibraryRetrieve(MPI_Comm,const char[],char *,size_t,PetscBool *);
1405PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscDLLibraryOpen(MPI_Comm,const char[],PetscDLLibrary *);
1406PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscDLLibraryClose(PetscDLLibrary);
1407
1408/*
1409 Useful utility routines
1410*/
1411PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSplitOwnership(MPI_Comm,PetscInt*,PetscInt*);
1412PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSplitOwnershipBlock(MPI_Comm,PetscInt,PetscInt*,PetscInt*);
1413PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSequentialPhaseBegin(MPI_Comm,PetscMPIInt);
1414PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSequentialPhaseEnd(MPI_Comm,PetscMPIInt);
1415PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscBarrier(PetscObject);
1416PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscMPIDump(FILE*);
1417PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscGlobalMinMaxInt(MPI_Comm,PetscInt[],PetscInt[]);
1418PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscGlobalMinMaxReal(MPI_Comm,PetscReal[],PetscReal[]);
1419
1420/*MC
1421 PetscNot - negates a logical type value and returns result as a PetscBool
1422
1423 Notes:
1424 This is useful in cases like
1425$ int *a;
1426$ PetscBool flag = PetscNot(a)
1427 where !a would not return a PetscBool because we cannot provide a cast from int to PetscBool in C.
1428
1429 Level: beginner
1430
1431 .seealso : PetscBool, PETSC_TRUE, PETSC_FALSE
1432M*/
1433#define PetscNot(a)((a) ? PETSC_FALSE : PETSC_TRUE) ((a) ? PETSC_FALSE : PETSC_TRUE)
1434
1435/*MC
1436 PetscHelpPrintf - Prints help messages.
1437
1438 Synopsis:
1439 #include <petscsys.h>
1440 PetscErrorCode (*PetscHelpPrintf)(const char format[],...);
1441
1442 Not Collective
1443
1444 Input Parameters:
1445. format - the usual printf() format string
1446
1447 Level: developer
1448
1449 Fortran Note:
1450 This routine is not supported in Fortran.
1451
1452
1453.seealso: PetscFPrintf(), PetscSynchronizedPrintf(), PetscErrorPrintf()
1454M*/
1455PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode (*PetscHelpPrintf)(MPI_Comm,const char[],...);
1456
1457/*
1458 Defines PETSc profiling.
1459*/
1460#include <petsclog.h>
1461
1462/*
1463 Simple PETSc parallel IO for ASCII printing
1464*/
1465PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscFixFilename(const char[],char[]);
1466PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscFOpen(MPI_Comm,const char[],const char[],FILE**);
1467PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscFClose(MPI_Comm,FILE*);
1468PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscFPrintf(MPI_Comm,FILE*,const char[],...);
1469PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscPrintf(MPI_Comm,const char[],...);
1470PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSNPrintf(char*,size_t,const char [],...);
1471PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSNPrintfCount(char*,size_t,const char [],size_t*,...);
1472PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscFormatRealArray(char[],size_t,const char*,PetscInt,const PetscReal[]);
1473
1474PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscErrorPrintfDefault(const char [],...);
1475PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscErrorPrintfNone(const char [],...);
1476PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscHelpPrintfDefault(MPI_Comm,const char [],...);
1477
1478PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscFormatConvertGetSize(const char*,size_t*);
1479PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscFormatConvert(const char*,char *);
1480
1481#if defined(PETSC_HAVE_POPEN1)
1482PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscPOpen(MPI_Comm,const char[],const char[],const char[],FILE **);
1483PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscPClose(MPI_Comm,FILE*);
1484PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscPOpenSetMachine(const char[]);
1485#endif
1486
1487PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSynchronizedPrintf(MPI_Comm,const char[],...);
1488PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSynchronizedFPrintf(MPI_Comm,FILE*,const char[],...);
1489PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSynchronizedFlush(MPI_Comm,FILE*);
1490PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSynchronizedFGets(MPI_Comm,FILE*,size_t,char[]);
1491PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscStartMatlab(MPI_Comm,const char[],const char[],FILE**);
1492PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscStartJava(MPI_Comm,const char[],const char[],FILE**);
1493PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscGetPetscDir(const char*[]);
1494
1495PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscClassId PETSC_CONTAINER_CLASSID;
1496PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscContainerGetPointer(PetscContainer,void **);
1497PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscContainerSetPointer(PetscContainer,void *);
1498PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscContainerDestroy(PetscContainer*);
1499PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscContainerCreate(MPI_Comm,PetscContainer *);
1500PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscContainerSetUserDestroy(PetscContainer, PetscErrorCode (*)(void*));
1501PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscContainerUserDestroyDefault(void*);
1502
1503/*
1504 For use in debuggers
1505*/
1506PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscMPIInt PetscGlobalRank;
1507PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscMPIInt PetscGlobalSize;
1508PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscIntView(PetscInt,const PetscInt[],PetscViewer);
1509PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscRealView(PetscInt,const PetscReal[],PetscViewer);
1510PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscScalarView(PetscInt,const PetscScalar[],PetscViewer);
1511
1512#include <stddef.h>
1513#include <string.h> /* for memcpy, memset */
1514#include <stdlib.h>
1515
1516#if defined(PETSC_HAVE_XMMINTRIN_H1) && !defined(__CUDACC__)
1517#include <xmmintrin.h>
1518#endif
1519
1520/*@C
1521 PetscMemmove - Copies n bytes, beginning at location b, to the space
1522 beginning at location a. Copying between regions that overlap will
1523 take place correctly. Use PetscMemcpy() if the locations do not overlap
1524
1525 Not Collective
1526
1527 Input Parameters:
1528+ b - pointer to initial memory space
1529. a - pointer to copy space
1530- n - length (in bytes) of space to copy
1531
1532 Level: intermediate
1533
1534 Note:
1535 PetscArraymove() is preferred
1536 This routine is analogous to memmove().
1537
1538 Developers Note: This is inlined for performance
1539
1540.seealso: PetscMemcpy(), PetscMemcmp(), PetscArrayzero(), PetscMemzero(), PetscArraycmp(), PetscArraycpy(), PetscStrallocpy(),
1541 PetscArraymove()
1542@*/
1543PETSC_STATIC_INLINEstatic inline PetscErrorCode PetscMemmove(void *a,const void *b,size_t n)
1544{
1545 PetscFunctionBegindo { do { ; if (petscstack && (petscstack->currentsize
< 64)) { petscstack->function[petscstack->currentsize
] = __func__; petscstack->file[petscstack->currentsize]
= "/sandbox/petsc/petsc.next/include/petscsys.h"; petscstack
->line[petscstack->currentsize] = 1545; petscstack->
petscroutine[petscstack->currentsize] = PETSC_TRUE; petscstack
->currentsize++; } if (petscstack) { petscstack->hotdepth
+= (PETSC_FALSE || petscstack->hotdepth); } ; } while (0)
; ; } while (0)
;
1546 if (n > 0 && !a) SETERRQ(PETSC_COMM_SELF,PETSC_ERR_ARG_NULL,"Trying to copy to null pointer")return PetscError(((MPI_Comm)0x44000001),1546,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,85,PETSC_ERROR_INITIAL,"Trying to copy to null pointer")
;
1547 if (n > 0 && !b) SETERRQ(PETSC_COMM_SELF,PETSC_ERR_ARG_NULL,"Trying to copy from a null pointer")return PetscError(((MPI_Comm)0x44000001),1547,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,85,PETSC_ERROR_INITIAL,"Trying to copy from a null pointer")
;
1548#if !defined(PETSC_HAVE_MEMMOVE1)
1549 if (a < b) {
1550 if (a <= b - n) memcpy(a,b,n);
1551 else {
1552 memcpy(a,b,(int)(b - a));
1553 PetscMemmove(b,b + (int)(b - a),n - (int)(b - a));
1554 }
1555 } else {
1556 if (b <= a - n) memcpy(a,b,n);
1557 else {
1558 memcpy(b + n,b + (n - (int)(a - b)),(int)(a - b));
1559 PetscMemmove(a,b,n - (int)(a - b));
1560 }
1561 }
1562#else
1563 memmove((char*)(a),(char*)(b),n);
1564#endif
1565 PetscFunctionReturn(0)do { do { ; if (petscstack && petscstack->currentsize
> 0) { petscstack->currentsize--; petscstack->function
[petscstack->currentsize] = 0; petscstack->file[petscstack
->currentsize] = 0; petscstack->line[petscstack->currentsize
] = 0; petscstack->petscroutine[petscstack->currentsize
] = PETSC_FALSE; } if (petscstack) { petscstack->hotdepth =
(((petscstack->hotdepth-1)<(0)) ? (0) : (petscstack->
hotdepth-1)); } ; } while (0); return(0);} while (0)
;
1566}
1567
1568/*@C
1569 PetscMemcpy - Copies n bytes, beginning at location b, to the space
1570 beginning at location a. The two memory regions CANNOT overlap, use
1571 PetscMemmove() in that case.
1572
1573 Not Collective
1574
1575 Input Parameters:
1576+ b - pointer to initial memory space
1577- n - length (in bytes) of space to copy
1578
1579 Output Parameter:
1580. a - pointer to copy space
1581
1582 Level: intermediate
1583
1584 Compile Option:
1585 PETSC_PREFER_DCOPY_FOR_MEMCPY will cause the BLAS dcopy() routine to be used
1586 for memory copies on double precision values.
1587 PETSC_PREFER_COPY_FOR_MEMCPY will cause C code to be used
1588 for memory copies on double precision values.
1589 PETSC_PREFER_FORTRAN_FORMEMCPY will cause Fortran code to be used
1590 for memory copies on double precision values.
1591
1592 Note:
1593 Prefer PetscArraycpy()
1594 This routine is analogous to memcpy().
1595 Not available from Fortran
1596
1597 Developer Note: this is inlined for fastest performance
1598
1599.seealso: PetscMemzero(), PetscMemcmp(), PetscArrayzero(), PetscArraycmp(), PetscArraycpy(), PetscMemmove(), PetscStrallocpy()
1600
1601@*/
1602PETSC_STATIC_INLINEstatic inline PetscErrorCode PetscMemcpy(void *a,const void *b,size_t n)
1603{
1604#if defined(PETSC_USE_DEBUG1)
1605 size_t al = (size_t) a,bl = (size_t) b;
1606 size_t nl = (size_t) n;
1607 PetscFunctionBegindo { do { ; if (petscstack && (petscstack->currentsize
< 64)) { petscstack->function[petscstack->currentsize
] = __func__; petscstack->file[petscstack->currentsize]
= "/sandbox/petsc/petsc.next/include/petscsys.h"; petscstack
->line[petscstack->currentsize] = 1607; petscstack->
petscroutine[petscstack->currentsize] = PETSC_TRUE; petscstack
->currentsize++; } if (petscstack) { petscstack->hotdepth
+= (PETSC_FALSE || petscstack->hotdepth); } ; } while (0)
; ; } while (0)
;
1608 if (n > 0 && !b) SETERRQ(PETSC_COMM_SELF,PETSC_ERR_ARG_NULL,"Trying to copy from a null pointer")return PetscError(((MPI_Comm)0x44000001),1608,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,85,PETSC_ERROR_INITIAL,"Trying to copy from a null pointer")
;
1609 if (n > 0 && !a) SETERRQ(PETSC_COMM_SELF,PETSC_ERR_ARG_NULL,"Trying to copy to a null pointer")return PetscError(((MPI_Comm)0x44000001),1609,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,85,PETSC_ERROR_INITIAL,"Trying to copy to a null pointer")
;
1610#else
1611 PetscFunctionBegindo { do { ; if (petscstack && (petscstack->currentsize
< 64)) { petscstack->function[petscstack->currentsize
] = __func__; petscstack->file[petscstack->currentsize]
= "/sandbox/petsc/petsc.next/include/petscsys.h"; petscstack
->line[petscstack->currentsize] = 1611; petscstack->
petscroutine[petscstack->currentsize] = PETSC_TRUE; petscstack
->currentsize++; } if (petscstack) { petscstack->hotdepth
+= (PETSC_FALSE || petscstack->hotdepth); } ; } while (0)
; ; } while (0)
;
1612#endif
1613 if (a != b && n > 0) {
1614#if defined(PETSC_USE_DEBUG1)
1615 if ((al > bl && (al - bl) < nl) || (bl - al) < nl) SETERRQ3(PETSC_COMM_SELF,PETSC_ERR_ARG_INCOMP,"Memory regions overlap: either use PetscMemmov()\n\return PetscError(((MPI_Comm)0x44000001),1617,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,75,PETSC_ERROR_INITIAL,"Memory regions overlap: either use PetscMemmov()\n or make sure your copy regions and lengths are correct. \n Length (bytes) %ld first address %ld second address %ld"
,nl,al,bl)
1616 or make sure your copy regions and lengths are correct. \n\return PetscError(((MPI_Comm)0x44000001),1617,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,75,PETSC_ERROR_INITIAL,"Memory regions overlap: either use PetscMemmov()\n or make sure your copy regions and lengths are correct. \n Length (bytes) %ld first address %ld second address %ld"
,nl,al,bl)
1617 Length (bytes) %ld first address %ld second address %ld",nl,al,bl)return PetscError(((MPI_Comm)0x44000001),1617,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,75,PETSC_ERROR_INITIAL,"Memory regions overlap: either use PetscMemmov()\n or make sure your copy regions and lengths are correct. \n Length (bytes) %ld first address %ld second address %ld"
,nl,al,bl)
;
1618#endif
1619#if (defined(PETSC_PREFER_DCOPY_FOR_MEMCPY) || defined(PETSC_PREFER_COPY_FOR_MEMCPY) || defined(PETSC_PREFER_FORTRAN_FORMEMCPY))
1620 if (!(a % sizeof(PetscScalar)) && !(n % sizeof(PetscScalar))) {
1621 size_t len = n/sizeof(PetscScalar);
1622#if defined(PETSC_PREFER_DCOPY_FOR_MEMCPY)
1623 PetscBLASInt one = 1,blen;
1624 PetscErrorCode ierr;
1625 ierr = PetscBLASIntCast(len,&blen);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),1625,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
1626 PetscStackCallBLAS("BLAScopy",BLAScopy_(&blen,(PetscScalar *)b,&one,(PetscScalar *)a,&one));
1627#elif defined(PETSC_PREFER_FORTRAN_FORMEMCPY)
1628 fortrancopy_(&len,(PetscScalar*)b,(PetscScalar*)a);
1629#else
1630 size_t i;
1631 PetscScalar *x = (PetscScalar*)b, *y = (PetscScalar*)a;
1632 for (i=0; i<len; i++) y[i] = x[i];
1633#endif
1634 } else {
1635 memcpy((char*)(a),(char*)(b),n);
1636 }
1637#else
1638 memcpy((char*)(a),(char*)(b),n);
1639#endif
1640 }
1641 PetscFunctionReturn(0)do { do { ; if (petscstack && petscstack->currentsize
> 0) { petscstack->currentsize--; petscstack->function
[petscstack->currentsize] = 0; petscstack->file[petscstack
->currentsize] = 0; petscstack->line[petscstack->currentsize
] = 0; petscstack->petscroutine[petscstack->currentsize
] = PETSC_FALSE; } if (petscstack) { petscstack->hotdepth =
(((petscstack->hotdepth-1)<(0)) ? (0) : (petscstack->
hotdepth-1)); } ; } while (0); return(0);} while (0)
;
1642}
1643
1644/*@C
1645 PetscMemzero - Zeros the specified memory.
1646
1647 Not Collective
1648
1649 Input Parameters:
1650+ a - pointer to beginning memory location
1651- n - length (in bytes) of memory to initialize
1652
1653 Level: intermediate
1654
1655 Compile Option:
1656 PETSC_PREFER_BZERO - on certain machines (the IBM RS6000) the bzero() routine happens
1657 to be faster than the memset() routine. This flag causes the bzero() routine to be used.
1658
1659 Not available from Fortran
1660 Prefer PetscArrayzero()
1661
1662 Developer Note: this is inlined for fastest performance
1663
1664.seealso: PetscMemcpy(), PetscMemcmp(), PetscArrayzero(), PetscArraycmp(), PetscArraycpy(), PetscMemmove(), PetscStrallocpy()
1665@*/
1666PETSC_STATIC_INLINEstatic inline PetscErrorCode PetscMemzero(void *a,size_t n)
1667{
1668 if (n > 0) {
1669#if defined(PETSC_USE_DEBUG1)
1670 if (!a) SETERRQ(PETSC_COMM_SELF,PETSC_ERR_ARG_NULL,"Trying to zero at a null pointer")return PetscError(((MPI_Comm)0x44000001),1670,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,85,PETSC_ERROR_INITIAL,"Trying to zero at a null pointer")
;
1671#endif
1672#if defined(PETSC_PREFER_ZERO_FOR_MEMZERO)
1673 if (!(((long) a) % sizeof(PetscScalar)) && !(n % sizeof(PetscScalar))) {
1674 size_t i,len = n/sizeof(PetscScalar);
1675 PetscScalar *x = (PetscScalar*)a;
1676 for (i=0; i<len; i++) x[i] = 0.0;
1677 } else {
1678#elif defined(PETSC_PREFER_FORTRAN_FOR_MEMZERO)
1679 if (!(((long) a) % sizeof(PetscScalar)) && !(n % sizeof(PetscScalar))) {
1680 PetscInt len = n/sizeof(PetscScalar);
1681 fortranzero_(&len,(PetscScalar*)a);
1682 } else {
1683#endif
1684#if defined(PETSC_PREFER_BZERO)
1685 bzero((char *)a,n);
1686#else
1687 memset((char*)a,0,n);
1688#endif
1689#if defined(PETSC_PREFER_ZERO_FOR_MEMZERO) || defined(PETSC_PREFER_FORTRAN_FOR_MEMZERO)
1690 }
1691#endif
1692 }
1693 return 0;
1694}
1695
1696/*MC
1697 PetscArraycmp - Compares two arrays in memory.
1698
1699 Synopsis:
1700 #include <petscsys.h>
1701 PetscErrorCode PetscArraycmp(const anytype *str1,const anytype *str2,size_t cnt,PetscBool *e)
1702
1703 Not Collective
1704
1705 Input Parameters:
1706+ str1 - First array
1707. str2 - Second array
1708- cnt - Count of the array, not in bytes, but number of entries in the arrays
1709
1710 Output Parameters:
1711. e - PETSC_TRUE if equal else PETSC_FALSE.
1712
1713 Level: intermediate
1714
1715 Note:
1716 This routine is a preferred replacement to PetscMemcmp()
1717 The arrays must be of the same type
1718
1719.seealso: PetscMemcpy(), PetscMemcmp(), PetscArrayzero(), PetscMemzero(), PetscArraycpy(), PetscMemmove(), PetscStrallocpy(),
1720 PetscArraymove()
1721M*/
1722#define PetscArraycmp(str1,str2,cnt,e)((sizeof(*(str1)) != sizeof(*(str2))) || PetscMemcmp(str1,str2
,(cnt)*sizeof(*(str1)),e))
((sizeof(*(str1)) != sizeof(*(str2))) || PetscMemcmp(str1,str2,(cnt)*sizeof(*(str1)),e))
1723
1724/*MC
1725 PetscArraymove - Copies from one array in memory to another, the arrays may overlap. Use PetscArraycpy() when the arrays
1726 do not overlap
1727
1728 Synopsis:
1729 #include <petscsys.h>
1730 PetscErrorCode PetscArraymove(anytype *str1,const anytype *str2,size_t cnt)
1731
1732 Not Collective
1733
1734 Input Parameters:
1735+ str1 - First array
1736. str2 - Second array
1737- cnt - Count of the array, not in bytes, but number of entries in the arrays
1738
1739 Level: intermediate
1740
1741 Note:
1742 This routine is a preferred replacement to PetscMemmove()
1743 The arrays must be of the same type
1744
1745.seealso: PetscMemcpy(), PetscMemcmp(), PetscArrayzero(), PetscMemzero(), PetscArraycpy(), PetscMemmove(), PetscArraycmp(), PetscStrallocpy()
1746M*/
1747#define PetscArraymove(str1,str2,cnt)((sizeof(*(str1)) != sizeof(*(str2))) || PetscMemmove(str1,str2
,(cnt)*sizeof(*(str1))))
((sizeof(*(str1)) != sizeof(*(str2))) || PetscMemmove(str1,str2,(cnt)*sizeof(*(str1))))
1748
1749/*MC
1750 PetscArraycpy - Copies from one array in memory to another
1751
1752 Synopsis:
1753 #include <petscsys.h>
1754 PetscErrorCode PetscArraycpy(anytype *str1,const anytype *str2,size_t cnt)
1755
1756 Not Collective
1757
1758 Input Parameters:
1759+ str1 - First array
1760. str2 - Second array
1761- cnt - Count of the array, not in bytes, but number of entries in the arrays
1762
1763 Level: intermediate
1764
1765 Note:
1766 This routine is a preferred replacement to PetscMemcpy()
1767 The arrays must be of the same type
1768
1769.seealso: PetscMemcpy(), PetscMemcmp(), PetscArrayzero(), PetscMemzero(), PetscArraymove(), PetscMemmove(), PetscArraycmp(), PetscStrallocpy()
1770M*/
1771#define PetscArraycpy(str1,str2,cnt)((sizeof(*(str1)) != sizeof(*(str2))) || PetscMemcpy(str1,str2
,(cnt)*sizeof(*(str1))))
((sizeof(*(str1)) != sizeof(*(str2))) || PetscMemcpy(str1,str2,(cnt)*sizeof(*(str1))))
1772
1773/*MC
1774 PetscArrayzero - Zeros an array in memory.
1775
1776 Synopsis:
1777 #include <petscsys.h>
1778 PetscErrorCode PetscArrayzero(anytype *str1,size_t cnt)
1779
1780 Not Collective
1781
1782 Input Parameters:
1783+ str1 - array
1784- cnt - Count of the array, not in bytes, but number of entries in the array
1785
1786 Level: intermediate
1787
1788 Note:
1789 This routine is a preferred replacement to PetscMemzero()
1790
1791.seealso: PetscMemcpy(), PetscMemcmp(), PetscMemzero(), PetscArraycmp(), PetscArraycpy(), PetscMemmove(), PetscStrallocpy(), PetscArraymove()
1792M*/
1793#define PetscArrayzero(str1,cnt)PetscMemzero(str1,(cnt)*sizeof(*(str1))) PetscMemzero(str1,(cnt)*sizeof(*(str1)))
1794
1795/*MC
1796 PetscPrefetchBlock - Prefetches a block of memory
1797
1798 Synopsis:
1799 #include <petscsys.h>
1800 void PetscPrefetchBlock(const anytype *a,size_t n,int rw,int t)
1801
1802 Not Collective
1803
1804 Input Parameters:
1805+ a - pointer to first element to fetch (any type but usually PetscInt or PetscScalar)
1806. n - number of elements to fetch
1807. rw - 1 if the memory will be written to, otherwise 0 (ignored by many processors)
1808- t - temporal locality (PETSC_PREFETCH_HINT_{NTA,T0,T1,T2}), see note
1809
1810 Level: developer
1811
1812 Notes:
1813 The last two arguments (rw and t) must be compile-time constants.
1814
1815 Adopting Intel's x86/x86-64 conventions, there are four levels of temporal locality. Not all architectures offer
1816 equivalent locality hints, but the following macros are always defined to their closest analogue.
1817+ PETSC_PREFETCH_HINT_NTA - Non-temporal. Prefetches directly to L1, evicts to memory (skips higher level cache unless it was already there when prefetched).
1818. PETSC_PREFETCH_HINT_T0 - Fetch to all levels of cache and evict to the closest level. Use this when the memory will be reused regularly despite necessary eviction from L1.
1819. PETSC_PREFETCH_HINT_T1 - Fetch to level 2 and higher (not L1).
1820- PETSC_PREFETCH_HINT_T2 - Fetch to high-level cache only. (On many systems, T0 and T1 are equivalent.)
1821
1822 This function does nothing on architectures that do not support prefetch and never errors (even if passed an invalid
1823 address).
1824
1825M*/
1826#define PetscPrefetchBlock(a,n,rw,t)do { const char *_p = (const char*)(a),*_end = (const char*)(
(a)+(n)); for ( ; _p < _end; _p += 64) (__builtin_prefetch
((void *)((const char*)(_p)), ((((t))) >> 2) & 1, (
((t))) & 0x3)); } while (0)
do { \
1827 const char *_p = (const char*)(a),*_end = (const char*)((a)+(n)); \
1828 for ( ; _p < _end; _p += PETSC_LEVEL1_DCACHE_LINESIZE64) PETSC_Prefetch(_p,(rw),(t))(__builtin_prefetch((void *)((const char*)(_p)), ((((t))) >>
2) & 1, (((t))) & 0x3))
; \
1829 } while (0)
1830
1831/*
1832 Determine if some of the kernel computation routines use
1833 Fortran (rather than C) for the numerical calculations. On some machines
1834 and compilers (like complex numbers) the Fortran version of the routines
1835 is faster than the C/C++ versions. The flag --with-fortran-kernels
1836 should be used with ./configure to turn these on.
1837*/
1838#if defined(PETSC_USE_FORTRAN_KERNELS)
1839
1840#if !defined(PETSC_USE_FORTRAN_KERNEL_MULTCRL)
1841#define PETSC_USE_FORTRAN_KERNEL_MULTCRL
1842#endif
1843
1844#if !defined(PETSC_USE_FORTRAN_KERNEL_MULTAIJPERM)
1845#define PETSC_USE_FORTRAN_KERNEL_MULTAIJPERM
1846#endif
1847
1848#if !defined(PETSC_USE_FORTRAN_KERNEL_MULTAIJ)
1849#define PETSC_USE_FORTRAN_KERNEL_MULTAIJ
1850#endif
1851
1852#if !defined(PETSC_USE_FORTRAN_KERNEL_MULTTRANSPOSEAIJ)
1853#define PETSC_USE_FORTRAN_KERNEL_MULTTRANSPOSEAIJ
1854#endif
1855
1856#if !defined(PETSC_USE_FORTRAN_KERNEL_NORM)
1857#define PETSC_USE_FORTRAN_KERNEL_NORM
1858#endif
1859
1860#if !defined(PETSC_USE_FORTRAN_KERNEL_MAXPY)
1861#define PETSC_USE_FORTRAN_KERNEL_MAXPY
1862#endif
1863
1864#if !defined(PETSC_USE_FORTRAN_KERNEL_SOLVEAIJ)
1865#define PETSC_USE_FORTRAN_KERNEL_SOLVEAIJ
1866#endif
1867
1868#if !defined(PETSC_USE_FORTRAN_KERNEL_RELAXAIJ)
1869#define PETSC_USE_FORTRAN_KERNEL_RELAXAIJ
1870#endif
1871
1872#if !defined(PETSC_USE_FORTRAN_KERNEL_SOLVEBAIJ)
1873#define PETSC_USE_FORTRAN_KERNEL_SOLVEBAIJ
1874#endif
1875
1876#if !defined(PETSC_USE_FORTRAN_KERNEL_MULTADDAIJ)
1877#define PETSC_USE_FORTRAN_KERNEL_MULTADDAIJ
1878#endif
1879
1880#if !defined(PETSC_USE_FORTRAN_KERNEL_MDOT)
1881#define PETSC_USE_FORTRAN_KERNEL_MDOT
1882#endif
1883
1884#if !defined(PETSC_USE_FORTRAN_KERNEL_XTIMESY)
1885#define PETSC_USE_FORTRAN_KERNEL_XTIMESY
1886#endif
1887
1888#if !defined(PETSC_USE_FORTRAN_KERNEL_AYPX)
1889#define PETSC_USE_FORTRAN_KERNEL_AYPX
1890#endif
1891
1892#if !defined(PETSC_USE_FORTRAN_KERNEL_WAXPY)
1893#define PETSC_USE_FORTRAN_KERNEL_WAXPY
1894#endif
1895
1896#endif
1897
1898/*
1899 Macros for indicating code that should be compiled with a C interface,
1900 rather than a C++ interface. Any routines that are dynamically loaded
1901 (such as the PCCreate_XXX() routines) must be wrapped so that the name
1902 mangler does not change the functions symbol name. This just hides the
1903 ugly extern "C" {} wrappers.
1904*/
1905#if defined(__cplusplus)
1906# define EXTERN_C_BEGIN extern "C" {
1907# define EXTERN_C_END }
1908#else
1909# define EXTERN_C_BEGIN
1910# define EXTERN_C_END
1911#endif
1912
1913/* --------------------------------------------------------------------*/
1914
1915/*MC
1916 MPI_Comm - the basic object used by MPI to determine which processes are involved in a
1917 communication
1918
1919 Level: beginner
1920
1921 Note: This manual page is a place-holder because MPICH does not have a manual page for MPI_Comm
1922
1923.seealso: PETSC_COMM_WORLD, PETSC_COMM_SELF
1924M*/
1925
1926#if defined(PETSC_HAVE_MPIIO1)
1927PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode MPIU_File_write_all(MPI_File,void*,PetscMPIInt,MPI_Datatype,MPI_Status*);
1928PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode MPIU_File_read_all(MPI_File,void*,PetscMPIInt,MPI_Datatype,MPI_Status*);
1929#endif
1930
1931/* the following petsc_static_inline require petscerror.h */
1932
1933/* Limit MPI to 32-bits */
1934#define PETSC_MPI_INT_MAX2147483647 2147483647
1935#define PETSC_MPI_INT_MIN-2147483647 -2147483647
1936/* Limit BLAS to 32-bits */
1937#define PETSC_BLAS_INT_MAX2147483647 2147483647
1938#define PETSC_BLAS_INT_MIN-2147483647 -2147483647
1939
1940/*@C
1941 PetscBLASIntCast - casts a PetscInt (which may be 64 bits in size) to a PetscBLASInt (which may be 32 bits in size), generates an
1942 error if the PetscBLASInt is not large enough to hold the number.
1943
1944 Not Collective
1945
1946 Input Parameter:
1947. a - the PetscInt value
1948
1949 Output Parameter:
1950. b - the resulting PetscBLASInt value
1951
1952 Level: advanced
1953
1954 Not available from Fortran
1955
1956.seealso: PetscBLASInt, PetscMPIInt, PetscInt, PetscMPIIntCast()
1957@*/
1958PETSC_STATIC_INLINEstatic inline PetscErrorCode PetscBLASIntCast(PetscInt a,PetscBLASInt *b)
1959{
1960 PetscFunctionBegindo { do { ; if (petscstack && (petscstack->currentsize
< 64)) { petscstack->function[petscstack->currentsize
] = __func__; petscstack->file[petscstack->currentsize]
= "/sandbox/petsc/petsc.next/include/petscsys.h"; petscstack
->line[petscstack->currentsize] = 1960; petscstack->
petscroutine[petscstack->currentsize] = PETSC_TRUE; petscstack
->currentsize++; } if (petscstack) { petscstack->hotdepth
+= (PETSC_FALSE || petscstack->hotdepth); } ; } while (0)
; ; } while (0)
;
1961 *b = (PetscBLASInt)(a);
1962#if defined(PETSC_USE_64BIT_INDICES) && !defined(PETSC_HAVE_64BIT_BLAS_INDICES)
1963 if ((a) > PETSC_BLAS_INT_MAX2147483647) SETERRQ(PETSC_COMM_SELF,PETSC_ERR_ARG_OUTOFRANGE,"Array too long for BLAS/LAPACK")return PetscError(((MPI_Comm)0x44000001),1963,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,63,PETSC_ERROR_INITIAL,"Array too long for BLAS/LAPACK")
;
1964#endif
1965 PetscFunctionReturn(0)do { do { ; if (petscstack && petscstack->currentsize
> 0) { petscstack->currentsize--; petscstack->function
[petscstack->currentsize] = 0; petscstack->file[petscstack
->currentsize] = 0; petscstack->line[petscstack->currentsize
] = 0; petscstack->petscroutine[petscstack->currentsize
] = PETSC_FALSE; } if (petscstack) { petscstack->hotdepth =
(((petscstack->hotdepth-1)<(0)) ? (0) : (petscstack->
hotdepth-1)); } ; } while (0); return(0);} while (0)
;
1966}
1967
1968/*@C
1969 PetscMPIIntCast - casts a PetscInt (which may be 64 bits in size) to a PetscMPIInt (which may be 32 bits in size), generates an
1970 error if the PetscMPIInt is not large enough to hold the number.
1971
1972 Not Collective
1973
1974 Input Parameter:
1975. a - the PetscInt value
1976
1977 Output Parameter:
1978. b - the resulting PetscMPIInt value
1979
1980 Level: advanced
1981
1982 Not available from Fortran
1983
1984.seealso: PetscBLASInt, PetscMPIInt, PetscInt, PetscBLASIntCast()
1985@*/
1986PETSC_STATIC_INLINEstatic inline PetscErrorCode PetscMPIIntCast(PetscInt a,PetscMPIInt *b)
1987{
1988 PetscFunctionBegindo { do { ; if (petscstack && (petscstack->currentsize
< 64)) { petscstack->function[petscstack->currentsize
] = __func__; petscstack->file[petscstack->currentsize]
= "/sandbox/petsc/petsc.next/include/petscsys.h"; petscstack
->line[petscstack->currentsize] = 1988; petscstack->
petscroutine[petscstack->currentsize] = PETSC_TRUE; petscstack
->currentsize++; } if (petscstack) { petscstack->hotdepth
+= (PETSC_FALSE || petscstack->hotdepth); } ; } while (0)
; ; } while (0)
;
1989 *b = (PetscMPIInt)(a);
1990#if defined(PETSC_USE_64BIT_INDICES)
1991 if ((a) > PETSC_MPI_INT_MAX2147483647) SETERRQ(PETSC_COMM_SELF,PETSC_ERR_ARG_OUTOFRANGE,"Array too long for MPI")return PetscError(((MPI_Comm)0x44000001),1991,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,63,PETSC_ERROR_INITIAL,"Array too long for MPI")
;
1992#endif
1993 PetscFunctionReturn(0)do { do { ; if (petscstack && petscstack->currentsize
> 0) { petscstack->currentsize--; petscstack->function
[petscstack->currentsize] = 0; petscstack->file[petscstack
->currentsize] = 0; petscstack->line[petscstack->currentsize
] = 0; petscstack->petscroutine[petscstack->currentsize
] = PETSC_FALSE; } if (petscstack) { petscstack->hotdepth =
(((petscstack->hotdepth-1)<(0)) ? (0) : (petscstack->
hotdepth-1)); } ; } while (0); return(0);} while (0)
;
1994}
1995
1996#define PetscInt64Mult(a,b)((PetscInt64)(a))*((PetscInt64)(b)) ((PetscInt64)(a))*((PetscInt64)(b))
1997
1998/*@C
1999
2000 PetscRealIntMultTruncate - Computes the product of a positive PetscReal and a positive PetscInt and truncates the value to slightly less than the maximal possible value
2001
2002 Not Collective
2003
2004 Input Parameter:
2005+ a - the PetscReal value
2006- b - the second value
2007
2008 Returns:
2009 the result as a PetscInt value
2010
2011 Use PetscInt64Mult() to compute the product of two PetscInt as a PetscInt64
2012 Use PetscIntMultTruncate() to compute the product of two positive PetscInt and truncate to fit a PetscInt
2013 Use PetscIntMultError() to compute the product of two PetscInt if you wish to generate an error if the result will not fit in a PetscInt
2014
2015 Developers Note:
2016 We currently assume that PetscInt addition can never overflow, this is obviously wrong but requires many more checks.
2017
2018 This is used where we compute approximate sizes for workspace and need to insure the workspace is index-able.
2019
2020 Not available from Fortran
2021
2022 Level: advanced
2023
2024.seealso: PetscBLASInt, PetscMPIInt, PetscInt, PetscBLASIntCast(), PetscInt64Mult()
2025@*/
2026PETSC_STATIC_INLINEstatic inline PetscInt PetscRealIntMultTruncate(PetscReal a,PetscInt b)
2027{
2028 PetscInt64 r;
2029
2030 r = (PetscInt64) (a*(PetscReal)b);
2031 if (r > PETSC_MAX_INT2147483647 - 100) r = PETSC_MAX_INT2147483647 - 100;
2032 return (PetscInt) r;
2033}
2034
2035/*@C
2036
2037 PetscIntMultTruncate - Computes the product of two positive PetscInt and truncates the value to slightly less than the maximal possible value
2038
2039 Not Collective
2040
2041 Input Parameter:
2042+ a - the PetscInt value
2043- b - the second value
2044
2045 Returns:
2046 the result as a PetscInt value
2047
2048 Use PetscInt64Mult() to compute the product of two PetscInt as a PetscInt64
2049 Use PetscRealIntMultTruncate() to compute the product of a PetscReal and a PetscInt and truncate to fit a PetscInt
2050 Use PetscIntMultError() to compute the product of two PetscInt if you wish to generate an error if the result will not fit in a PetscInt
2051
2052 Not available from Fortran
2053
2054 Developers Note: We currently assume that PetscInt addition can never overflow, this is obviously wrong but requires many more checks.
2055
2056 This is used where we compute approximate sizes for workspace and need to insure the workspace is index-able.
2057
2058 Level: advanced
2059
2060.seealso: PetscBLASInt, PetscMPIInt, PetscInt, PetscBLASIntCast(), PetscInt64Mult()
2061@*/
2062PETSC_STATIC_INLINEstatic inline PetscInt PetscIntMultTruncate(PetscInt a,PetscInt b)
2063{
2064 PetscInt64 r;
2065
2066 r = PetscInt64Mult(a,b)((PetscInt64)(a))*((PetscInt64)(b));
2067 if (r > PETSC_MAX_INT2147483647 - 100) r = PETSC_MAX_INT2147483647 - 100;
2068 return (PetscInt) r;
2069}
2070
2071/*@C
2072
2073 PetscIntSumTruncate - Computes the sum of two positive PetscInt and truncates the value to slightly less than the maximal possible value
2074
2075 Not Collective
2076
2077 Input Parameter:
2078+ a - the PetscInt value
2079- b - the second value
2080
2081 Returns:
2082 the result as a PetscInt value
2083
2084 Use PetscInt64Mult() to compute the product of two PetscInt as a PetscInt64
2085 Use PetscRealIntMultTruncate() to compute the product of a PetscReal and a PetscInt and truncate to fit a PetscInt
2086 Use PetscIntMultError() to compute the product of two PetscInt if you wish to generate an error if the result will not fit in a PetscInt
2087
2088 This is used where we compute approximate sizes for workspace and need to insure the workspace is index-able.
2089
2090 Not available from Fortran
2091
2092 Level: advanced
2093
2094.seealso: PetscBLASInt, PetscMPIInt, PetscInt, PetscBLASIntCast(), PetscInt64Mult()
2095@*/
2096PETSC_STATIC_INLINEstatic inline PetscInt PetscIntSumTruncate(PetscInt a,PetscInt b)
2097{
2098 PetscInt64 r;
2099
2100 r = ((PetscInt64)a) + ((PetscInt64)b);
2101 if (r > PETSC_MAX_INT2147483647 - 100) r = PETSC_MAX_INT2147483647 - 100;
2102 return (PetscInt) r;
2103}
2104
2105/*@C
2106
2107 PetscIntMultError - Computes the product of two positive PetscInt and generates an error with overflow.
2108
2109 Not Collective
2110
2111 Input Parameter:
2112+ a - the PetscInt value
2113- b - the second value
2114
2115 Output Parameter:ma
2116. result - the result as a PetscInt value, or NULL if you do not want the result, you just want to check if it overflows
2117
2118 Use PetscInt64Mult() to compute the product of two 32 bit PetscInt and store in a PetscInt64
2119 Use PetscIntMultTruncate() to compute the product of two PetscInt and truncate it to fit in a PetscInt
2120
2121 Not available from Fortran
2122
2123 Developers Note: We currently assume that PetscInt addition does not overflow, this is obviously wrong but requires many more checks.
2124
2125 Level: advanced
2126
2127.seealso: PetscBLASInt, PetscMPIInt, PetscInt, PetscBLASIntCast(), PetscIntMult64(), PetscIntSumError()
2128@*/
2129PETSC_STATIC_INLINEstatic inline PetscErrorCode PetscIntMultError(PetscInt a,PetscInt b,PetscInt *result)
2130{
2131 PetscInt64 r;
2132
2133 PetscFunctionBegindo { do { ; if (petscstack && (petscstack->currentsize
< 64)) { petscstack->function[petscstack->currentsize
] = __func__; petscstack->file[petscstack->currentsize]
= "/sandbox/petsc/petsc.next/include/petscsys.h"; petscstack
->line[petscstack->currentsize] = 2133; petscstack->
petscroutine[petscstack->currentsize] = PETSC_TRUE; petscstack
->currentsize++; } if (petscstack) { petscstack->hotdepth
+= (PETSC_FALSE || petscstack->hotdepth); } ; } while (0)
; ; } while (0)
;
2134 r = PetscInt64Mult(a,b)((PetscInt64)(a))*((PetscInt64)(b));
2135#if !defined(PETSC_USE_64BIT_INDICES)
2136 if (r > PETSC_MAX_INT2147483647) SETERRQ2(PETSC_COMM_SELF,PETSC_ERR_SUP,"Product of two integer %d %d overflow, you must ./configure PETSc with --with-64-bit-indices for the case you are running",a,b)return PetscError(((MPI_Comm)0x44000001),2136,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,56,PETSC_ERROR_INITIAL,"Product of two integer %d %d overflow, you must ./configure PETSc with --with-64-bit-indices for the case you are running"
,a,b)
;
2137#endif
2138 if (result) *result = (PetscInt) r;
2139 PetscFunctionReturn(0)do { do { ; if (petscstack && petscstack->currentsize
> 0) { petscstack->currentsize--; petscstack->function
[petscstack->currentsize] = 0; petscstack->file[petscstack
->currentsize] = 0; petscstack->line[petscstack->currentsize
] = 0; petscstack->petscroutine[petscstack->currentsize
] = PETSC_FALSE; } if (petscstack) { petscstack->hotdepth =
(((petscstack->hotdepth-1)<(0)) ? (0) : (petscstack->
hotdepth-1)); } ; } while (0); return(0);} while (0)
;
2140}
2141
2142/*@C
2143
2144 PetscIntSumError - Computes the sum of two positive PetscInt and generates an error with overflow.
2145
2146 Not Collective
2147
2148 Input Parameter:
2149+ a - the PetscInt value
2150- b - the second value
2151
2152 Output Parameter:ma
2153. c - the result as a PetscInt value, or NULL if you do not want the result, you just want to check if it overflows
2154
2155 Use PetscInt64Mult() to compute the product of two 32 bit PetscInt and store in a PetscInt64
2156 Use PetscIntMultTruncate() to compute the product of two PetscInt and truncate it to fit in a PetscInt
2157
2158 Not available from Fortran
2159
2160 Level: advanced
2161
2162.seealso: PetscBLASInt, PetscMPIInt, PetscInt, PetscBLASIntCast(), PetscInt64Mult()
2163@*/
2164PETSC_STATIC_INLINEstatic inline PetscErrorCode PetscIntSumError(PetscInt a,PetscInt b,PetscInt *result)
2165{
2166 PetscInt64 r;
2167
2168 PetscFunctionBegindo { do { ; if (petscstack && (petscstack->currentsize
< 64)) { petscstack->function[petscstack->currentsize
] = __func__; petscstack->file[petscstack->currentsize]
= "/sandbox/petsc/petsc.next/include/petscsys.h"; petscstack
->line[petscstack->currentsize] = 2168; petscstack->
petscroutine[petscstack->currentsize] = PETSC_TRUE; petscstack
->currentsize++; } if (petscstack) { petscstack->hotdepth
+= (PETSC_FALSE || petscstack->hotdepth); } ; } while (0)
; ; } while (0)
;
2169 r = ((PetscInt64)a) + ((PetscInt64)b);
2170#if !defined(PETSC_USE_64BIT_INDICES)
2171 if (r > PETSC_MAX_INT2147483647) SETERRQ2(PETSC_COMM_SELF,PETSC_ERR_SUP,"Sum of two integer %d %d overflow, you must ./configure PETSc with --with-64-bit-indices for the case you are running",a,b)return PetscError(((MPI_Comm)0x44000001),2171,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,56,PETSC_ERROR_INITIAL,"Sum of two integer %d %d overflow, you must ./configure PETSc with --with-64-bit-indices for the case you are running"
,a,b)
;
2172#endif
2173 if (result) *result = (PetscInt) r;
2174 PetscFunctionReturn(0)do { do { ; if (petscstack && petscstack->currentsize
> 0) { petscstack->currentsize--; petscstack->function
[petscstack->currentsize] = 0; petscstack->file[petscstack
->currentsize] = 0; petscstack->line[petscstack->currentsize
] = 0; petscstack->petscroutine[petscstack->currentsize
] = PETSC_FALSE; } if (petscstack) { petscstack->hotdepth =
(((petscstack->hotdepth-1)<(0)) ? (0) : (petscstack->
hotdepth-1)); } ; } while (0); return(0);} while (0)
;
2175}
2176
2177/*
2178 The IBM include files define hz, here we hide it so that it may be used as a regular user variable.
2179*/
2180#if defined(hz)
2181# undef hz
2182#endif
2183
2184#include <limits.h>
2185
2186/* The number of bits in a byte */
2187
2188#define PETSC_BITS_PER_BYTE8 CHAR_BIT8
2189
2190/* For arrays that contain filenames or paths */
2191
2192#if defined(PETSC_HAVE_SYS_PARAM_H1)
2193# include <sys/param.h>
2194#endif
2195#if defined(PETSC_HAVE_SYS_TYPES_H1)
2196# include <sys/types.h>
2197#endif
2198#if defined(MAXPATHLEN4096)
2199# define PETSC_MAX_PATH_LEN4096 MAXPATHLEN4096
2200#elif defined(MAX_PATH)
2201# define PETSC_MAX_PATH_LEN4096 MAX_PATH
2202#elif defined(_MAX_PATH)
2203# define PETSC_MAX_PATH_LEN4096 _MAX_PATH
2204#else
2205# define PETSC_MAX_PATH_LEN4096 4096
2206#endif
2207
2208/*MC
2209
2210 PETSC_VERSION - This manual page provides information about how PETSc documents and uses its version information. This information is available to both C/C++
2211 and Fortran compilers when petscsys.h is included.
2212
2213
2214 The current PETSc version and the API for accessing it are defined in petscversion.h
2215
2216 The complete version number is given as the triple PETSC_VERSION_MAJOR.PETSC_VERSION_MINOR.PETSC_VERSION_SUBMINOR (in short hand x.y.z)
2217
2218 A change in the minor version number (y) indicates possible/likely changes in the PETSc API. Note this is different than with the semantic versioning convention
2219 where only a change in the major version number (x) indicates a change in the API.
2220
2221 A subminor greater than zero indicates a patch release. Version x.y.z maintains source and binary compatibility with version x.y.w for all z and w
2222
2223 Use the macros PETSC_VERSION_EQ(x,y,z), PETSC_VERSION_LT(x,y,z), PETSC_VERSION_LE(x,y,z), PETSC_VERSION_GT(x,y,z),
2224 PETSC_VERSION_GE(x,y,z) to determine if the current version is equal to, less than, less than or equal to, greater than or greater than or equal to a given
2225 version number (x.y.z).
2226
2227 PETSC_RELEASE_DATE is the date the x.y version was released (i.e. the version before any patch releases)
2228
2229 PETSC_VERSION_DATE is the date the x.y.z version was released
2230
2231 PETSC_VERSION_GIT is the last git commit to the repository given in the form vx.y.z-wwwww
2232
2233 PETSC_VERSION_DATE_GIT is the date of the last git commit to the repository
2234
2235 Level: intermediate
2236
2237 PETSC_VERSION_() and PETSC_VERSION_PATCH are deprecated and will eventually be removed. For several releases PETSC_VERSION_PATCH is always 0
2238
2239M*/
2240
2241/*MC
2242
2243 UsingFortran - To use PETSc with Fortran you must use both PETSc include files and modules. At the beginning
2244 of every function and module definition you need something like
2245
2246$
2247$#include "petsc/finclude/petscXXX.h"
2248$ use petscXXX
2249
2250 You can declare PETSc variables using either of the following.
2251
2252$ XXX variablename
2253$ type(tXXX) variablename
2254
2255 For example,
2256
2257$#include "petsc/finclude/petscvec.h"
2258$ use petscvec
2259$
2260$ Vec b
2261$ type(tVec) x
2262
2263 Level: beginner
2264
2265M*/
2266
2267PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscGetArchType(char[],size_t);
2268PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscGetHostName(char[],size_t);
2269PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscGetUserName(char[],size_t);
2270PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscGetProgramName(char[],size_t);
2271PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSetProgramName(const char[]);
2272PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscGetDate(char[],size_t);
2273PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscGetVersion(char[], size_t);
2274PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscGetVersionNumber(PetscInt*,PetscInt*,PetscInt*,PetscInt*);
2275
2276PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSortInt(PetscInt,PetscInt[]);
2277PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSortedRemoveDupsInt(PetscInt*,PetscInt[]);
2278PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSortRemoveDupsInt(PetscInt*,PetscInt[]);
2279PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscFindInt(PetscInt, PetscInt, const PetscInt[], PetscInt*);
2280PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscFindMPIInt(PetscMPIInt, PetscInt, const PetscMPIInt[], PetscInt*);
2281PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSortIntWithPermutation(PetscInt,const PetscInt[],PetscInt[]);
2282PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSortStrWithPermutation(PetscInt,const char*[],PetscInt[]);
2283PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSortIntWithArray(PetscInt,PetscInt[],PetscInt[]);
2284PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSortIntWithArrayPair(PetscInt,PetscInt[],PetscInt[],PetscInt[]);
2285PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSortMPIInt(PetscInt,PetscMPIInt[]);
2286PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSortRemoveDupsMPIInt(PetscInt*,PetscMPIInt[]);
2287PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSortMPIIntWithArray(PetscMPIInt,PetscMPIInt[],PetscMPIInt[]);
2288PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSortMPIIntWithIntArray(PetscMPIInt,PetscMPIInt[],PetscInt[]);
2289PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSortIntWithScalarArray(PetscInt,PetscInt[],PetscScalar[]);
2290PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSortIntWithDataArray(PetscInt,PetscInt[],void*,size_t,void*);
2291PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSortReal(PetscInt,PetscReal[]);
2292PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSortRealWithArrayInt(PetscInt,PetscReal[],PetscInt[]);
2293PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSortRealWithPermutation(PetscInt,const PetscReal[],PetscInt[]);
2294PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSortRemoveDupsReal(PetscInt*,PetscReal[]);
2295PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscFindReal(PetscReal,PetscInt,const PetscReal[],PetscReal,PetscInt*);
2296PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSortSplit(PetscInt,PetscInt,PetscScalar[],PetscInt[]);
2297PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSortSplitReal(PetscInt,PetscInt,PetscReal[],PetscInt[]);
2298PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscProcessTree(PetscInt,const PetscBool [],const PetscInt[],PetscInt*,PetscInt**,PetscInt**,PetscInt**,PetscInt**);
2299PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscMergeIntArrayPair(PetscInt,const PetscInt[],const PetscInt[],PetscInt,const PetscInt[],const PetscInt[],PetscInt*,PetscInt**,PetscInt**);
2300PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscMergeIntArray(PetscInt,const PetscInt[],PetscInt,const PetscInt[],PetscInt*,PetscInt**);
2301PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscMergeMPIIntArray(PetscInt,const PetscMPIInt[],PetscInt,const PetscMPIInt[],PetscInt*,PetscMPIInt**);
2302
2303PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSetDisplay(void);
2304PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscGetDisplay(char[],size_t);
2305
2306/*J
2307 PetscRandomType - String with the name of a PETSc randomizer
2308
2309 Level: beginner
2310
2311 Notes:
2312 To use SPRNG or RANDOM123 you must have ./configure PETSc
2313 with the option --download-sprng or --download-random123
2314
2315.seealso: PetscRandomSetType(), PetscRandom, PetscRandomCreate()
2316J*/
2317typedef const char* PetscRandomType;
2318#define PETSCRAND"rand" "rand"
2319#define PETSCRAND48"rand48" "rand48"
2320#define PETSCSPRNG"sprng" "sprng"
2321#define PETSCRANDER48"rander48" "rander48"
2322#define PETSCRANDOM123"random123" "random123"
2323
2324/* Logging support */
2325PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscClassId PETSC_RANDOM_CLASSID;
2326
2327PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscRandomInitializePackage(void);
2328
2329/* Dynamic creation and loading functions */
2330PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscFunctionList PetscRandomList;
2331
2332PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscRandomRegister(const char[],PetscErrorCode (*)(PetscRandom));
2333PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscRandomSetType(PetscRandom, PetscRandomType);
2334PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscRandomSetFromOptions(PetscRandom);
2335PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscRandomGetType(PetscRandom, PetscRandomType*);
2336PETSC_STATIC_INLINEstatic inline PetscErrorCode PetscRandomViewFromOptions(PetscRandom A,PetscObject obj,const char name[]) {return PetscObjectViewFromOptions((PetscObject)A,obj,name);}
2337PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscRandomView(PetscRandom,PetscViewer);
2338
2339PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscRandomCreate(MPI_Comm,PetscRandom*);
2340PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscRandomGetValue(PetscRandom,PetscScalar*);
2341PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscRandomGetValueReal(PetscRandom,PetscReal*);
2342PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscRandomGetInterval(PetscRandom,PetscScalar*,PetscScalar*);
2343PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscRandomSetInterval(PetscRandom,PetscScalar,PetscScalar);
2344PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscRandomSetSeed(PetscRandom,unsigned long);
2345PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscRandomGetSeed(PetscRandom,unsigned long *);
2346PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscRandomSeed(PetscRandom);
2347PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscRandomDestroy(PetscRandom*);
2348
2349PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscGetFullPath(const char[],char[],size_t);
2350PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscGetRelativePath(const char[],char[],size_t);
2351PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscGetWorkingDirectory(char[],size_t);
2352PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscGetRealPath(const char[],char[]);
2353PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscGetHomeDirectory(char[],size_t);
2354PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscTestFile(const char[],char,PetscBool *);
2355PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscTestDirectory(const char[],char,PetscBool *);
2356PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscMkdir(const char[]);
2357PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscMkdtemp(char[]);
2358PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscRMTree(const char[]);
2359
2360PETSC_STATIC_INLINEstatic inline PetscBool PetscBinaryBigEndian(void) {long _petsc_v = 1; return ((char*)&_petsc_v)[0] ? PETSC_FALSE : PETSC_TRUE;}
2361
2362PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscBinaryRead(int,void*,PetscInt,PetscInt*,PetscDataType);
2363PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscBinarySynchronizedRead(MPI_Comm,int,void*,PetscInt,PetscInt*,PetscDataType);
2364PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscBinaryWrite(int,void*,PetscInt,PetscDataType,PetscBool);
2365PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscBinarySynchronizedWrite(MPI_Comm,int,void*,PetscInt,PetscDataType,PetscBool);
2366PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscBinaryOpen(const char[],PetscFileMode,int *);
2367PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscBinaryClose(int);
2368PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSharedTmp(MPI_Comm,PetscBool *);
2369PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSharedWorkingDirectory(MPI_Comm,PetscBool *);
2370PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscGetTmp(MPI_Comm,char[],size_t);
2371PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscFileRetrieve(MPI_Comm,const char[],char[],size_t,PetscBool *);
2372PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscLs(MPI_Comm,const char[],char[],size_t,PetscBool *);
2373#if defined(PETSC_USE_SOCKET_VIEWER1)
2374PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscOpenSocket(const char[],int,int*);
2375#endif
2376
2377PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscBinarySeek(int,off_t,PetscBinarySeekType,off_t*);
2378PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscBinarySynchronizedSeek(MPI_Comm,int,off_t,PetscBinarySeekType,off_t*);
2379PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscByteSwap(void *,PetscDataType,PetscInt);
2380
2381PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSetDebugTerminal(const char[]);
2382PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSetDebugger(const char[],PetscBool );
2383PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSetDefaultDebugger(void);
2384PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSetDebuggerFromString(const char*);
2385PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscAttachDebugger(void);
2386PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscStopForDebugger(void);
2387
2388PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscGatherNumberOfMessages(MPI_Comm,const PetscMPIInt[],const PetscMPIInt[],PetscMPIInt*);
2389PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscGatherMessageLengths(MPI_Comm,PetscMPIInt,PetscMPIInt,const PetscMPIInt[],PetscMPIInt**,PetscMPIInt**);
2390PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscGatherMessageLengths2(MPI_Comm,PetscMPIInt,PetscMPIInt,const PetscMPIInt[],const PetscMPIInt[],PetscMPIInt**,PetscMPIInt**,PetscMPIInt**);
2391PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscPostIrecvInt(MPI_Comm,PetscMPIInt,PetscMPIInt,const PetscMPIInt[],const PetscMPIInt[],PetscInt***,MPI_Request**);
2392PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscPostIrecvScalar(MPI_Comm,PetscMPIInt,PetscMPIInt,const PetscMPIInt[],const PetscMPIInt[],PetscScalar***,MPI_Request**);
2393PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscCommBuildTwoSided(MPI_Comm,PetscMPIInt,MPI_Datatype,PetscMPIInt,const PetscMPIInt*,const void*,PetscMPIInt*,PetscMPIInt**,void*)
2394 PetscAttrMPIPointerWithType(6,3);
2395PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscCommBuildTwoSidedF(MPI_Comm,PetscMPIInt,MPI_Datatype,PetscMPIInt,const PetscMPIInt[],const void*,PetscMPIInt*,PetscMPIInt**,void*,PetscMPIInt,
2396 PetscErrorCode (*send)(MPI_Comm,const PetscMPIInt[],PetscMPIInt,PetscMPIInt,void*,MPI_Request[],void*),
2397 PetscErrorCode (*recv)(MPI_Comm,const PetscMPIInt[],PetscMPIInt,void*,MPI_Request[],void*),void *ctx)
2398 PetscAttrMPIPointerWithType(6,3);
2399PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscCommBuildTwoSidedFReq(MPI_Comm,PetscMPIInt,MPI_Datatype,PetscMPIInt,const PetscMPIInt[],const void*,PetscMPIInt*,PetscMPIInt**,void*,PetscMPIInt,
2400 MPI_Request**,MPI_Request**,
2401 PetscErrorCode (*send)(MPI_Comm,const PetscMPIInt[],PetscMPIInt,PetscMPIInt,void*,MPI_Request[],void*),
2402 PetscErrorCode (*recv)(MPI_Comm,const PetscMPIInt[],PetscMPIInt,void*,MPI_Request[],void*),void *ctx)
2403 PetscAttrMPIPointerWithType(6,3);
2404
2405PETSC_EXTERNextern __attribute__((visibility ("default"))) const char *const PetscBuildTwoSidedTypes[];
2406PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscCommBuildTwoSidedSetType(MPI_Comm,PetscBuildTwoSidedType);
2407PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscCommBuildTwoSidedGetType(MPI_Comm,PetscBuildTwoSidedType*);
2408
2409PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSSEIsEnabled(MPI_Comm,PetscBool*,PetscBool*);
2410
2411PETSC_EXTERNextern __attribute__((visibility ("default"))) MPI_Comm PetscObjectComm(PetscObject);
2412
2413PETSC_EXTERNextern __attribute__((visibility ("default"))) const char *const PetscSubcommTypes[];
2414
2415struct _n_PetscSubcomm {
2416 MPI_Comm parent; /* parent communicator */
2417 MPI_Comm dupparent; /* duplicate parent communicator, under which the processors of this subcomm have contiguous rank */
2418 MPI_Comm child; /* the sub-communicator */
2419 PetscMPIInt n; /* num of subcommunicators under the parent communicator */
2420 PetscMPIInt color; /* color of processors belong to this communicator */
2421 PetscMPIInt *subsize; /* size of subcommunicator[color] */
2422 PetscSubcommType type;
2423 char *subcommprefix;
2424};
2425
2426PETSC_STATIC_INLINEstatic inline MPI_Comm PetscSubcommParent(PetscSubcomm scomm) {return scomm->parent;}
2427PETSC_STATIC_INLINEstatic inline MPI_Comm PetscSubcommChild(PetscSubcomm scomm) {return scomm->child;}
9
Access to field 'child' results in a dereference of a null pointer (loaded from variable 'scomm')
2428PETSC_STATIC_INLINEstatic inline MPI_Comm PetscSubcommContiguousParent(PetscSubcomm scomm) {return scomm->dupparent;}
2429PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSubcommCreate(MPI_Comm,PetscSubcomm*);
2430PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSubcommDestroy(PetscSubcomm*);
2431PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSubcommSetNumber(PetscSubcomm,PetscInt);
2432PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSubcommSetType(PetscSubcomm,PetscSubcommType);
2433PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSubcommSetTypeGeneral(PetscSubcomm,PetscMPIInt,PetscMPIInt);
2434PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSubcommView(PetscSubcomm,PetscViewer);
2435PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSubcommSetFromOptions(PetscSubcomm);
2436PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSubcommSetOptionsPrefix(PetscSubcomm,const char[]);
2437
2438PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscHeapCreate(PetscInt,PetscHeap*);
2439PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscHeapAdd(PetscHeap,PetscInt,PetscInt);
2440PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscHeapPop(PetscHeap,PetscInt*,PetscInt*);
2441PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscHeapPeek(PetscHeap,PetscInt*,PetscInt*);
2442PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscHeapStash(PetscHeap,PetscInt,PetscInt);
2443PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscHeapUnstash(PetscHeap);
2444PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscHeapDestroy(PetscHeap*);
2445PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscHeapView(PetscHeap,PetscViewer);
2446
2447PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscProcessPlacementView(PetscViewer);
2448PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscShmCommGet(MPI_Comm,PetscShmComm*);
2449PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscShmCommGlobalToLocal(PetscShmComm,PetscMPIInt,PetscMPIInt*);
2450PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscShmCommLocalToGlobal(PetscShmComm,PetscMPIInt,PetscMPIInt*);
2451PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscShmCommGetMpiShmComm(PetscShmComm,MPI_Comm*);
2452
2453/* routines to better support OpenMP multithreading needs of some PETSc third party libraries */
2454PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscOmpCtrlCreate(MPI_Comm,PetscInt,PetscOmpCtrl*);
2455PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscOmpCtrlGetOmpComms(PetscOmpCtrl,MPI_Comm*,MPI_Comm*,PetscBool*);
2456PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscOmpCtrlDestroy(PetscOmpCtrl*);
2457PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscOmpCtrlBarrier(PetscOmpCtrl);
2458PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscOmpCtrlOmpRegionOnMasterBegin(PetscOmpCtrl);
2459PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscOmpCtrlOmpRegionOnMasterEnd(PetscOmpCtrl);
2460
2461PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSegBufferCreate(size_t,size_t,PetscSegBuffer*);
2462PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSegBufferDestroy(PetscSegBuffer*);
2463PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSegBufferGet(PetscSegBuffer,size_t,void*);
2464PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSegBufferExtractAlloc(PetscSegBuffer,void*);
2465PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSegBufferExtractTo(PetscSegBuffer,void*);
2466PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSegBufferExtractInPlace(PetscSegBuffer,void*);
2467PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSegBufferGetSize(PetscSegBuffer,size_t*);
2468PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSegBufferUnuse(PetscSegBuffer,size_t);
2469
2470
2471/* Type-safe wrapper to encourage use of PETSC_RESTRICT. Does not use PetscFunctionBegin because the error handling
2472 * prevents the compiler from completely erasing the stub. This is called in inner loops so it has to be as fast as
2473 * possible. */
2474PETSC_STATIC_INLINEstatic inline PetscErrorCode PetscSegBufferGetInts(PetscSegBuffer seg,size_t count,PetscInt *PETSC_RESTRICT__restrict *slot) {return PetscSegBufferGet(seg,count,(void**)slot);}
2475
2476extern PetscOptionsHelpPrinted PetscOptionsHelpPrintedSingleton;
2477PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscOptionsHelpPrintedDestroy(PetscOptionsHelpPrinted*);
2478PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscOptionsHelpPrintedCreate(PetscOptionsHelpPrinted*);
2479PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscOptionsHelpPrintedCheck(PetscOptionsHelpPrinted,const char*,const char*,PetscBool*);
2480
2481#include <stdarg.h>
2482PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscVSNPrintf(char*,size_t,const char[],size_t*,va_list);
2483PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode (*PetscVFPrintf)(FILE*,const char[],va_list);
2484
2485PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscSegBuffer PetscCitationsList;
2486
2487/*@C
2488 PetscCitationsRegister - Register a bibtex item to obtain credit for an implemented algorithm used in the code.
2489
2490 Not Collective - only what is registered on rank 0 of PETSC_COMM_WORLD will be printed
2491
2492 Input Parameters:
2493+ cite - the bibtex item, formated to displayed on multiple lines nicely
2494- set - a boolean variable initially set to PETSC_FALSE; this is used to insure only a single registration of the citation
2495
2496 Level: intermediate
2497
2498 Not available from Fortran
2499
2500 Options Database:
2501. -citations [filename] - print out the bibtex entries for the given computation
2502@*/
2503PETSC_STATIC_INLINEstatic inline PetscErrorCode PetscCitationsRegister(const char cit[],PetscBool *set)
2504{
2505 size_t len;
2506 char *vstring;
2507 PetscErrorCode ierr;
2508
2509 PetscFunctionBegindo { do { ; if (petscstack && (petscstack->currentsize
< 64)) { petscstack->function[petscstack->currentsize
] = __func__; petscstack->file[petscstack->currentsize]
= "/sandbox/petsc/petsc.next/include/petscsys.h"; petscstack
->line[petscstack->currentsize] = 2509; petscstack->
petscroutine[petscstack->currentsize] = PETSC_TRUE; petscstack
->currentsize++; } if (petscstack) { petscstack->hotdepth
+= (PETSC_FALSE || petscstack->hotdepth); } ; } while (0)
; ; } while (0)
;
2510 if (set && *set) PetscFunctionReturn(0)do { do { ; if (petscstack && petscstack->currentsize
> 0) { petscstack->currentsize--; petscstack->function
[petscstack->currentsize] = 0; petscstack->file[petscstack
->currentsize] = 0; petscstack->line[petscstack->currentsize
] = 0; petscstack->petscroutine[petscstack->currentsize
] = PETSC_FALSE; } if (petscstack) { petscstack->hotdepth =
(((petscstack->hotdepth-1)<(0)) ? (0) : (petscstack->
hotdepth-1)); } ; } while (0); return(0);} while (0)
;
2511 ierr = PetscStrlen(cit,&len);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),2511,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
2512 ierr = PetscSegBufferGet(PetscCitationsList,len,&vstring);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),2512,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
2513 ierr = PetscArraycpy(vstring,cit,len)((sizeof(*(vstring)) != sizeof(*(cit))) || PetscMemcpy(vstring
,cit,(len)*sizeof(*(vstring))))
;CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),2513,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
2514 if (set) *set = PETSC_TRUE;
2515 PetscFunctionReturn(0)do { do { ; if (petscstack && petscstack->currentsize
> 0) { petscstack->currentsize--; petscstack->function
[petscstack->currentsize] = 0; petscstack->file[petscstack
->currentsize] = 0; petscstack->line[petscstack->currentsize
] = 0; petscstack->petscroutine[petscstack->currentsize
] = PETSC_FALSE; } if (petscstack) { petscstack->hotdepth =
(((petscstack->hotdepth-1)<(0)) ? (0) : (petscstack->
hotdepth-1)); } ; } while (0); return(0);} while (0)
;
2516}
2517
2518PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscURLShorten(const char[],char[],size_t);
2519PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscGoogleDriveAuthorize(MPI_Comm,char[],char[],size_t);
2520PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscGoogleDriveRefresh(MPI_Comm,const char[],char[],size_t);
2521PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscGoogleDriveUpload(MPI_Comm,const char[],const char []);
2522
2523PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscBoxAuthorize(MPI_Comm,char[],char[],size_t);
2524PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscBoxRefresh(MPI_Comm,const char[],char[],char[],size_t);
2525
2526PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscGlobusGetTransfers(MPI_Comm,const char[],char[],size_t);
2527
2528PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscTextBelt(MPI_Comm,const char[],const char[],PetscBool*);
2529PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscTellMyCell(MPI_Comm,const char[],const char[],PetscBool*);
2530
2531PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscPullJSONValue(const char[],const char[],char[],size_t,PetscBool*);
2532PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscPushJSONValue(char[],const char[],const char[],size_t);
2533
2534
2535#if defined(PETSC_USE_DEBUG1)
2536/*
2537 Verify that all processes in the communicator have called this from the same line of code
2538 */
2539PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscAllreduceBarrierCheck(MPI_Comm,PetscMPIInt,int,const char*,const char *);
2540
2541/*MC
2542 MPIU_Allreduce - a PETSc replacement for MPI_Allreduce() that tries to determine if the call from all the MPI processes occur from the
2543 same place in the PETSc code. This helps to detect bugs where different MPI processes follow different code paths
2544 resulting in inconsistent and incorrect calls to MPI_Allreduce().
2545
2546 Collective
2547
2548 Synopsis:
2549 #include <petscsys.h>
2550 PetscErrorCode MPIU_Allreduce(void *indata,void *outdata,PetscMPIInt count,MPI_Datatype datatype, MPI_Op op, MPI_Comm comm);
2551
2552 Input Parameters:
2553+ indata - pointer to the input data to be reduced
2554. count - the number of MPI data items in indata and outdata
2555. datatype - the MPI datatype, for example MPI_INT
2556. op - the MPI operation, for example MPI_SUM
2557- comm - the MPI communicator on which the operation occurs
2558
2559 Output Parameter:
2560. outdata - the reduced values
2561
2562 Notes:
2563 In optimized mode this directly calls MPI_Allreduce()
2564
2565 Level: developer
2566
2567.seealso: MPI_Allreduce()
2568M*/
2569#define MPIU_Allreduce(a,b,c,d,e,fcomm)(PetscAllreduceBarrierCheck(fcomm,c,2569,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
) || ((petsc_allreduce_ct += PetscMPIParallelComm((fcomm)),0)
|| MPI_Allreduce((a),(b),(c),(d),(e),(fcomm))))
(PetscAllreduceBarrierCheck(fcomm,c,__LINE__2569,PETSC_FUNCTION_NAME__func__,__FILE__"/sandbox/petsc/petsc.next/include/petscsys.h") || MPI_Allreduce(a,b,c,d,e,fcomm)((petsc_allreduce_ct += PetscMPIParallelComm((fcomm)),0) || MPI_Allreduce
((a),(b),(c),(d),(e),(fcomm)))
)
2570#else
2571#define MPIU_Allreduce(a,b,c,d,e,fcomm)(PetscAllreduceBarrierCheck(fcomm,c,2571,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
) || ((petsc_allreduce_ct += PetscMPIParallelComm((fcomm)),0)
|| MPI_Allreduce((a),(b),(c),(d),(e),(fcomm))))
MPI_Allreduce(a,b,c,d,e,fcomm)((petsc_allreduce_ct += PetscMPIParallelComm((fcomm)),0) || MPI_Allreduce
((a),(b),(c),(d),(e),(fcomm)))
2572#endif
2573
2574#if defined(PETSC_HAVE_MPI_WIN_CREATE_FEATURE)
2575PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode MPIU_Win_allocate_shared(MPI_Aint,PetscMPIInt,MPI_Info,MPI_Comm,void*,MPI_Win*);
2576PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode MPIU_Win_shared_query(MPI_Win,PetscMPIInt,MPI_Aint*,PetscMPIInt*,void*);
2577#endif
2578
2579/*
2580 Returned from PETSc functions that are called from MPI, such as related to attributes
2581*/
2582PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscMPIInt PETSC_MPI_ERROR_CLASS;
2583PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscMPIInt PETSC_MPI_ERROR_CODE;
2584
2585#endif