Bug Summary

File:ksp/pc/impls/redundant/redundant.c
Warning:line 2459, column 76
Access to field 'child' results in a dereference of a null pointer (loaded from variable 'scomm')

Annotated Source Code

[?] Use j/k keys for keyboard navigation

/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c

1
2/*
3 This file defines a "solve the problem redundantly on each subgroup of processor" preconditioner.
4*/
5#include <petsc/private/pcimpl.h>
6#include <petscksp.h> /*I "petscksp.h" I*/
7
8typedef struct {
9 KSP ksp;
10 PC pc; /* actual preconditioner used on each processor */
11 Vec xsub,ysub; /* vectors of a subcommunicator to hold parallel vectors of PetscObjectComm((PetscObject)pc) */
12 Vec xdup,ydup; /* parallel vector that congregates xsub or ysub facilitating vector scattering */
13 Mat pmats; /* matrix and optional preconditioner matrix belong to a subcommunicator */
14 VecScatter scatterin,scatterout; /* scatter used to move all values to each processor group (subcommunicator) */
15 PetscBool useparallelmat;
16 PetscSubcomm psubcomm;
17 PetscInt nsubcomm; /* num of data structure PetscSubcomm */
18 PetscBool shifttypeset;
19 MatFactorShiftType shifttype;
20} PC_Redundant;
21
22PetscErrorCode PCFactorSetShiftType_Redundant(PC pc,MatFactorShiftType shifttype)
23{
24 PC_Redundant *red = (PC_Redundant*)pc->data;
25 PetscErrorCode ierr;
26
27 PetscFunctionBegindo { do { ; if (petscstack && (petscstack->currentsize
< 64)) { petscstack->function[petscstack->currentsize
] = __func__; petscstack->file[petscstack->currentsize]
= "/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
; petscstack->line[petscstack->currentsize] = 27; petscstack
->petscroutine[petscstack->currentsize] = PETSC_TRUE; petscstack
->currentsize++; } if (petscstack) { petscstack->hotdepth
+= (PETSC_FALSE || petscstack->hotdepth); } ; } while (0)
; ; } while (0)
;
28 if (red->ksp) {
29 PC pc;
30 ierr = KSPGetPC(red->ksp,&pc);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),30,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
31 ierr = PCFactorSetShiftType(pc,shifttype);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),31,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
32 } else {
33 red->shifttypeset = PETSC_TRUE;
34 red->shifttype = shifttype;
35 }
36 PetscFunctionReturn(0)do { do { ; if (petscstack && petscstack->currentsize
> 0) { petscstack->currentsize--; petscstack->function
[petscstack->currentsize] = 0; petscstack->file[petscstack
->currentsize] = 0; petscstack->line[petscstack->currentsize
] = 0; petscstack->petscroutine[petscstack->currentsize
] = PETSC_FALSE; } if (petscstack) { petscstack->hotdepth =
(((petscstack->hotdepth-1)<(0)) ? (0) : (petscstack->
hotdepth-1)); } ; } while (0); return(0);} while (0)
;
37}
38
39static PetscErrorCode PCView_Redundant(PC pc,PetscViewer viewer)
40{
41 PC_Redundant *red = (PC_Redundant*)pc->data;
42 PetscErrorCode ierr;
43 PetscBool iascii,isstring;
44 PetscViewer subviewer;
45
46 PetscFunctionBegindo { do { ; if (petscstack && (petscstack->currentsize
< 64)) { petscstack->function[petscstack->currentsize
] = __func__; petscstack->file[petscstack->currentsize]
= "/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
; petscstack->line[petscstack->currentsize] = 46; petscstack
->petscroutine[petscstack->currentsize] = PETSC_TRUE; petscstack
->currentsize++; } if (petscstack) { petscstack->hotdepth
+= (PETSC_FALSE || petscstack->hotdepth); } ; } while (0)
; ; } while (0)
;
47 ierr = PetscObjectTypeCompare((PetscObject)viewer,PETSCVIEWERASCII"ascii",&iascii);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),47,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
48 ierr = PetscObjectTypeCompare((PetscObject)viewer,PETSCVIEWERSTRING"string",&isstring);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),48,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
49 if (iascii) {
50 if (!red->psubcomm) {
51 ierr = PetscViewerASCIIPrintf(viewer," Not yet setup\n");CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),51,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
52 } else {
53 ierr = PetscViewerASCIIPrintf(viewer," First (color=0) of %D PCs follows\n",red->nsubcomm);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),53,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
54 ierr = PetscViewerGetSubViewer(viewer,((PetscObject)red->pc)->comm,&subviewer);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),54,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
55 if (!red->psubcomm->color) { /* only view first redundant pc */
56 ierr = PetscViewerASCIIPushTab(subviewer);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),56,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
57 ierr = KSPView(red->ksp,subviewer);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),57,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
58 ierr = PetscViewerASCIIPopTab(subviewer);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),58,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
59 }
60 ierr = PetscViewerRestoreSubViewer(viewer,((PetscObject)red->pc)->comm,&subviewer);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),60,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
61 }
62 } else if (isstring) {
63 ierr = PetscViewerStringSPrintf(viewer," Redundant solver preconditioner");CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),63,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
64 }
65 PetscFunctionReturn(0)do { do { ; if (petscstack && petscstack->currentsize
> 0) { petscstack->currentsize--; petscstack->function
[petscstack->currentsize] = 0; petscstack->file[petscstack
->currentsize] = 0; petscstack->line[petscstack->currentsize
] = 0; petscstack->petscroutine[petscstack->currentsize
] = PETSC_FALSE; } if (petscstack) { petscstack->hotdepth =
(((petscstack->hotdepth-1)<(0)) ? (0) : (petscstack->
hotdepth-1)); } ; } while (0); return(0);} while (0)
;
66}
67
68#include <../src/mat/impls/aij/mpi/mpiaij.h>
69static PetscErrorCode PCSetUp_Redundant(PC pc)
70{
71 PC_Redundant *red = (PC_Redundant*)pc->data;
72 PetscErrorCode ierr;
73 PetscInt mstart,mend,mlocal,M;
74 PetscMPIInt size;
75 MPI_Comm comm,subcomm;
76 Vec x;
77
78 PetscFunctionBegindo { do { ; if (petscstack && (petscstack->currentsize
< 64)) { petscstack->function[petscstack->currentsize
] = __func__; petscstack->file[petscstack->currentsize]
= "/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
; petscstack->line[petscstack->currentsize] = 78; petscstack
->petscroutine[petscstack->currentsize] = PETSC_TRUE; petscstack
->currentsize++; } if (petscstack) { petscstack->hotdepth
+= (PETSC_FALSE || petscstack->hotdepth); } ; } while (0)
; ; } while (0)
;
79 ierr = PetscObjectGetComm((PetscObject)pc,&comm);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),79,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
80
81 /* if pmatrix set by user is sequential then we do not need to gather the parallel matrix */
82 ierr = MPI_Comm_size(comm,&size);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),82,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
83 if (size == 1) red->useparallelmat = PETSC_FALSE;
1
Assuming 'size' is not equal to 1
2
Taking false branch
84
85 if (!pc->setupcalled) {
3
Assuming the condition is true
4
Taking true branch
86 PetscInt mloc_sub;
87 if (!red->psubcomm) { /* create red->psubcomm, new ksp and pc over subcomm */
5
Assuming pointer value is null
6
Taking true branch
88 KSP ksp;
89 ierr = PCRedundantGetKSP(pc,&ksp);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),89,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
90 }
91 subcomm = PetscSubcommChild(red->psubcomm);
7
Passing null pointer value via 1st parameter 'scomm'
8
Calling 'PetscSubcommChild'
92
93 if (red->useparallelmat) {
94 /* grab the parallel matrix and put it into processors of a subcomminicator */
95 ierr = MatCreateRedundantMatrix(pc->pmat,red->psubcomm->n,subcomm,MAT_INITIAL_MATRIX,&red->pmats);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),95,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
96
97 ierr = MPI_Comm_size(subcomm,&size);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),97,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
98 if (size > 1) {
99 PetscBool foundpack,issbaij;
100 ierr = PetscObjectTypeCompare((PetscObject)red->pmats,MATMPISBAIJ"mpisbaij",&issbaij);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),100,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
101 if (!issbaij) {
102 ierr = MatGetFactorAvailable(red->pmats,NULL((void*)0),MAT_FACTOR_LU,&foundpack);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),102,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
103 } else {
104 ierr = MatGetFactorAvailable(red->pmats,NULL((void*)0),MAT_FACTOR_CHOLESKY,&foundpack);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),104,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
105 }
106 if (!foundpack) { /* reset default ksp and pc */
107 ierr = KSPSetType(red->ksp,KSPGMRES"gmres");CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),107,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
108 ierr = PCSetType(red->pc,PCBJACOBI"bjacobi");CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),108,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
109 } else {
110 ierr = PCFactorSetMatSolverType(red->pc,NULL((void*)0));CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),110,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
111 }
112 }
113
114 ierr = KSPSetOperators(red->ksp,red->pmats,red->pmats);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),114,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
115
116 /* get working vectors xsub and ysub */
117 ierr = MatCreateVecs(red->pmats,&red->xsub,&red->ysub);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),117,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
118
119 /* create working vectors xdup and ydup.
120 xdup concatenates all xsub's contigously to form a mpi vector over dupcomm (see PetscSubcommCreate_interlaced())
121 ydup concatenates all ysub and has empty local arrays because ysub's arrays will be place into it.
122 Note: we use communicator dupcomm, not PetscObjectComm((PetscObject)pc)! */
123 ierr = MatGetLocalSize(red->pmats,&mloc_sub,NULL((void*)0));CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),123,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
124 ierr = VecCreateMPI(PetscSubcommContiguousParent(red->psubcomm),mloc_sub,PETSC_DECIDE-1,&red->xdup);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),124,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
125 ierr = VecCreateMPIWithArray(PetscSubcommContiguousParent(red->psubcomm),1,mloc_sub,PETSC_DECIDE-1,NULL((void*)0),&red->ydup);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),125,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
126
127 /* create vecscatters */
128 if (!red->scatterin) { /* efficiency of scatterin is independent from psubcomm_type! */
129 IS is1,is2;
130 PetscInt *idx1,*idx2,i,j,k;
131
132 ierr = MatCreateVecs(pc->pmat,&x,0);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),132,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
133 ierr = VecGetSize(x,&M);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),133,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
134 ierr = VecGetOwnershipRange(x,&mstart,&mend);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),134,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
135 mlocal = mend - mstart;
136 ierr = PetscMalloc2(red->psubcomm->n*mlocal,&idx1,red->psubcomm->n*mlocal,&idx2)PetscMallocA(2,PETSC_FALSE,136,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,(size_t)(red->psubcomm->n*mlocal)*sizeof(**(&idx1)
),(&idx1),(size_t)(red->psubcomm->n*mlocal)*sizeof(
**(&idx2)),(&idx2))
;CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),136,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
137 j = 0;
138 for (k=0; k<red->psubcomm->n; k++) {
139 for (i=mstart; i<mend; i++) {
140 idx1[j] = i;
141 idx2[j++] = i + M*k;
142 }
143 }
144 ierr = ISCreateGeneral(comm,red->psubcomm->n*mlocal,idx1,PETSC_COPY_VALUES,&is1);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),144,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
145 ierr = ISCreateGeneral(comm,red->psubcomm->n*mlocal,idx2,PETSC_COPY_VALUES,&is2);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),145,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
146 ierr = VecScatterCreate(x,is1,red->xdup,is2,&red->scatterin);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),146,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
147 ierr = ISDestroy(&is1);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),147,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
148 ierr = ISDestroy(&is2);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),148,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
149
150 /* Impl below is good for PETSC_SUBCOMM_INTERLACED (no inter-process communication) and PETSC_SUBCOMM_CONTIGUOUS (communication within subcomm) */
151 ierr = ISCreateStride(comm,mlocal,mstart+ red->psubcomm->color*M,1,&is1);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),151,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
152 ierr = ISCreateStride(comm,mlocal,mstart,1,&is2);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),152,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
153 ierr = VecScatterCreate(red->xdup,is1,x,is2,&red->scatterout);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),153,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
154 ierr = ISDestroy(&is1);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),154,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
155 ierr = ISDestroy(&is2);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),155,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
156 ierr = PetscFree2(idx1,idx2)PetscFreeA(2,156,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,&(idx1),&(idx2))
;CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),156,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
157 ierr = VecDestroy(&x);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),157,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
158 }
159 } else { /* !red->useparallelmat */
160 ierr = KSPSetOperators(red->ksp,pc->mat,pc->pmat);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),160,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
161 }
162 } else { /* pc->setupcalled */
163 if (red->useparallelmat) {
164 MatReuse reuse;
165 /* grab the parallel matrix and put it into processors of a subcomminicator */
166 /*--------------------------------------------------------------------------*/
167 if (pc->flag == DIFFERENT_NONZERO_PATTERN) {
168 /* destroy old matrices */
169 ierr = MatDestroy(&red->pmats);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),169,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
170 reuse = MAT_INITIAL_MATRIX;
171 } else {
172 reuse = MAT_REUSE_MATRIX;
173 }
174 ierr = MatCreateRedundantMatrix(pc->pmat,red->psubcomm->n,PetscSubcommChild(red->psubcomm),reuse,&red->pmats);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),174,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
175 ierr = KSPSetOperators(red->ksp,red->pmats,red->pmats);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),175,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
176 } else { /* !red->useparallelmat */
177 ierr = KSPSetOperators(red->ksp,pc->mat,pc->pmat);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),177,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
178 }
179 }
180
181 if (pc->setfromoptionscalled) {
182 ierr = KSPSetFromOptions(red->ksp);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),182,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
183 }
184 ierr = KSPSetUp(red->ksp);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),184,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
185 PetscFunctionReturn(0)do { do { ; if (petscstack && petscstack->currentsize
> 0) { petscstack->currentsize--; petscstack->function
[petscstack->currentsize] = 0; petscstack->file[petscstack
->currentsize] = 0; petscstack->line[petscstack->currentsize
] = 0; petscstack->petscroutine[petscstack->currentsize
] = PETSC_FALSE; } if (petscstack) { petscstack->hotdepth =
(((petscstack->hotdepth-1)<(0)) ? (0) : (petscstack->
hotdepth-1)); } ; } while (0); return(0);} while (0)
;
186}
187
188static PetscErrorCode PCApply_Redundant(PC pc,Vec x,Vec y)
189{
190 PC_Redundant *red = (PC_Redundant*)pc->data;
191 PetscErrorCode ierr;
192 PetscScalar *array;
193
194 PetscFunctionBegindo { do { ; if (petscstack && (petscstack->currentsize
< 64)) { petscstack->function[petscstack->currentsize
] = __func__; petscstack->file[petscstack->currentsize]
= "/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
; petscstack->line[petscstack->currentsize] = 194; petscstack
->petscroutine[petscstack->currentsize] = PETSC_TRUE; petscstack
->currentsize++; } if (petscstack) { petscstack->hotdepth
+= (PETSC_FALSE || petscstack->hotdepth); } ; } while (0)
; ; } while (0)
;
195 if (!red->useparallelmat) {
196 ierr = KSPSolve(red->ksp,x,y);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),196,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
197 ierr = KSPCheckSolve(red->ksp,pc,y);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),197,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
198 PetscFunctionReturn(0)do { do { ; if (petscstack && petscstack->currentsize
> 0) { petscstack->currentsize--; petscstack->function
[petscstack->currentsize] = 0; petscstack->file[petscstack
->currentsize] = 0; petscstack->line[petscstack->currentsize
] = 0; petscstack->petscroutine[petscstack->currentsize
] = PETSC_FALSE; } if (petscstack) { petscstack->hotdepth =
(((petscstack->hotdepth-1)<(0)) ? (0) : (petscstack->
hotdepth-1)); } ; } while (0); return(0);} while (0)
;
199 }
200
201 /* scatter x to xdup */
202 ierr = VecScatterBegin(red->scatterin,x,red->xdup,INSERT_VALUES,SCATTER_FORWARD);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),202,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
203 ierr = VecScatterEnd(red->scatterin,x,red->xdup,INSERT_VALUES,SCATTER_FORWARD);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),203,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
204
205 /* place xdup's local array into xsub */
206 ierr = VecGetArray(red->xdup,&array);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),206,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
207 ierr = VecPlaceArray(red->xsub,(const PetscScalar*)array);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),207,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
208
209 /* apply preconditioner on each processor */
210 ierr = KSPSolve(red->ksp,red->xsub,red->ysub);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),210,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
211 ierr = KSPCheckSolve(red->ksp,pc,red->ysub);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),211,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
212 ierr = VecResetArray(red->xsub);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),212,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
213 ierr = VecRestoreArray(red->xdup,&array);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),213,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
214
215 /* place ysub's local array into ydup */
216 ierr = VecGetArray(red->ysub,&array);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),216,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
217 ierr = VecPlaceArray(red->ydup,(const PetscScalar*)array);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),217,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
218
219 /* scatter ydup to y */
220 ierr = VecScatterBegin(red->scatterout,red->ydup,y,INSERT_VALUES,SCATTER_FORWARD);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),220,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
221 ierr = VecScatterEnd(red->scatterout,red->ydup,y,INSERT_VALUES,SCATTER_FORWARD);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),221,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
222 ierr = VecResetArray(red->ydup);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),222,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
223 ierr = VecRestoreArray(red->ysub,&array);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),223,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
224 PetscFunctionReturn(0)do { do { ; if (petscstack && petscstack->currentsize
> 0) { petscstack->currentsize--; petscstack->function
[petscstack->currentsize] = 0; petscstack->file[petscstack
->currentsize] = 0; petscstack->line[petscstack->currentsize
] = 0; petscstack->petscroutine[petscstack->currentsize
] = PETSC_FALSE; } if (petscstack) { petscstack->hotdepth =
(((petscstack->hotdepth-1)<(0)) ? (0) : (petscstack->
hotdepth-1)); } ; } while (0); return(0);} while (0)
;
225}
226
227static PetscErrorCode PCApplyTranspose_Redundant(PC pc,Vec x,Vec y)
228{
229 PC_Redundant *red = (PC_Redundant*)pc->data;
230 PetscErrorCode ierr;
231 PetscScalar *array;
232
233 PetscFunctionBegindo { do { ; if (petscstack && (petscstack->currentsize
< 64)) { petscstack->function[petscstack->currentsize
] = __func__; petscstack->file[petscstack->currentsize]
= "/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
; petscstack->line[petscstack->currentsize] = 233; petscstack
->petscroutine[petscstack->currentsize] = PETSC_TRUE; petscstack
->currentsize++; } if (petscstack) { petscstack->hotdepth
+= (PETSC_FALSE || petscstack->hotdepth); } ; } while (0)
; ; } while (0)
;
234 if (!red->useparallelmat) {
235 ierr = KSPSolveTranspose(red->ksp,x,y);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),235,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
236 ierr = KSPCheckSolve(red->ksp,pc,y);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),236,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
237 PetscFunctionReturn(0)do { do { ; if (petscstack && petscstack->currentsize
> 0) { petscstack->currentsize--; petscstack->function
[petscstack->currentsize] = 0; petscstack->file[petscstack
->currentsize] = 0; petscstack->line[petscstack->currentsize
] = 0; petscstack->petscroutine[petscstack->currentsize
] = PETSC_FALSE; } if (petscstack) { petscstack->hotdepth =
(((petscstack->hotdepth-1)<(0)) ? (0) : (petscstack->
hotdepth-1)); } ; } while (0); return(0);} while (0)
;
238 }
239
240 /* scatter x to xdup */
241 ierr = VecScatterBegin(red->scatterin,x,red->xdup,INSERT_VALUES,SCATTER_FORWARD);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),241,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
242 ierr = VecScatterEnd(red->scatterin,x,red->xdup,INSERT_VALUES,SCATTER_FORWARD);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),242,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
243
244 /* place xdup's local array into xsub */
245 ierr = VecGetArray(red->xdup,&array);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),245,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
246 ierr = VecPlaceArray(red->xsub,(const PetscScalar*)array);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),246,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
247
248 /* apply preconditioner on each processor */
249 ierr = KSPSolveTranspose(red->ksp,red->xsub,red->ysub);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),249,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
250 ierr = KSPCheckSolve(red->ksp,pc,red->ysub);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),250,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
251 ierr = VecResetArray(red->xsub);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),251,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
252 ierr = VecRestoreArray(red->xdup,&array);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),252,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
253
254 /* place ysub's local array into ydup */
255 ierr = VecGetArray(red->ysub,&array);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),255,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
256 ierr = VecPlaceArray(red->ydup,(const PetscScalar*)array);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),256,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
257
258 /* scatter ydup to y */
259 ierr = VecScatterBegin(red->scatterout,red->ydup,y,INSERT_VALUES,SCATTER_FORWARD);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),259,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
260 ierr = VecScatterEnd(red->scatterout,red->ydup,y,INSERT_VALUES,SCATTER_FORWARD);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),260,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
261 ierr = VecResetArray(red->ydup);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),261,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
262 ierr = VecRestoreArray(red->ysub,&array);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),262,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
263 PetscFunctionReturn(0)do { do { ; if (petscstack && petscstack->currentsize
> 0) { petscstack->currentsize--; petscstack->function
[petscstack->currentsize] = 0; petscstack->file[petscstack
->currentsize] = 0; petscstack->line[petscstack->currentsize
] = 0; petscstack->petscroutine[petscstack->currentsize
] = PETSC_FALSE; } if (petscstack) { petscstack->hotdepth =
(((petscstack->hotdepth-1)<(0)) ? (0) : (petscstack->
hotdepth-1)); } ; } while (0); return(0);} while (0)
;
264}
265
266static PetscErrorCode PCReset_Redundant(PC pc)
267{
268 PC_Redundant *red = (PC_Redundant*)pc->data;
269 PetscErrorCode ierr;
270
271 PetscFunctionBegindo { do { ; if (petscstack && (petscstack->currentsize
< 64)) { petscstack->function[petscstack->currentsize
] = __func__; petscstack->file[petscstack->currentsize]
= "/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
; petscstack->line[petscstack->currentsize] = 271; petscstack
->petscroutine[petscstack->currentsize] = PETSC_TRUE; petscstack
->currentsize++; } if (petscstack) { petscstack->hotdepth
+= (PETSC_FALSE || petscstack->hotdepth); } ; } while (0)
; ; } while (0)
;
272 if (red->useparallelmat) {
273 ierr = VecScatterDestroy(&red->scatterin);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),273,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
274 ierr = VecScatterDestroy(&red->scatterout);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),274,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
275 ierr = VecDestroy(&red->ysub);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),275,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
276 ierr = VecDestroy(&red->xsub);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),276,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
277 ierr = VecDestroy(&red->xdup);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),277,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
278 ierr = VecDestroy(&red->ydup);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),278,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
279 }
280 ierr = MatDestroy(&red->pmats);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),280,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
281 ierr = KSPReset(red->ksp);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),281,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
282 PetscFunctionReturn(0)do { do { ; if (petscstack && petscstack->currentsize
> 0) { petscstack->currentsize--; petscstack->function
[petscstack->currentsize] = 0; petscstack->file[petscstack
->currentsize] = 0; petscstack->line[petscstack->currentsize
] = 0; petscstack->petscroutine[petscstack->currentsize
] = PETSC_FALSE; } if (petscstack) { petscstack->hotdepth =
(((petscstack->hotdepth-1)<(0)) ? (0) : (petscstack->
hotdepth-1)); } ; } while (0); return(0);} while (0)
;
283}
284
285static PetscErrorCode PCDestroy_Redundant(PC pc)
286{
287 PC_Redundant *red = (PC_Redundant*)pc->data;
288 PetscErrorCode ierr;
289
290 PetscFunctionBegindo { do { ; if (petscstack && (petscstack->currentsize
< 64)) { petscstack->function[petscstack->currentsize
] = __func__; petscstack->file[petscstack->currentsize]
= "/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
; petscstack->line[petscstack->currentsize] = 290; petscstack
->petscroutine[petscstack->currentsize] = PETSC_TRUE; petscstack
->currentsize++; } if (petscstack) { petscstack->hotdepth
+= (PETSC_FALSE || petscstack->hotdepth); } ; } while (0)
; ; } while (0)
;
291 ierr = PCReset_Redundant(pc);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),291,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
292 ierr = KSPDestroy(&red->ksp);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),292,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
293 ierr = PetscSubcommDestroy(&red->psubcomm);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),293,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
294 ierr = PetscFree(pc->data)((*PetscTrFree)((void*)(pc->data),294,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
) || ((pc->data) = 0,0))
;CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),294,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
295 PetscFunctionReturn(0)do { do { ; if (petscstack && petscstack->currentsize
> 0) { petscstack->currentsize--; petscstack->function
[petscstack->currentsize] = 0; petscstack->file[petscstack
->currentsize] = 0; petscstack->line[petscstack->currentsize
] = 0; petscstack->petscroutine[petscstack->currentsize
] = PETSC_FALSE; } if (petscstack) { petscstack->hotdepth =
(((petscstack->hotdepth-1)<(0)) ? (0) : (petscstack->
hotdepth-1)); } ; } while (0); return(0);} while (0)
;
296}
297
298static PetscErrorCode PCSetFromOptions_Redundant(PetscOptionItems *PetscOptionsObject,PC pc)
299{
300 PetscErrorCode ierr;
301 PC_Redundant *red = (PC_Redundant*)pc->data;
302
303 PetscFunctionBegindo { do { ; if (petscstack && (petscstack->currentsize
< 64)) { petscstack->function[petscstack->currentsize
] = __func__; petscstack->file[petscstack->currentsize]
= "/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
; petscstack->line[petscstack->currentsize] = 303; petscstack
->petscroutine[petscstack->currentsize] = PETSC_TRUE; petscstack
->currentsize++; } if (petscstack) { petscstack->hotdepth
+= (PETSC_FALSE || petscstack->hotdepth); } ; } while (0)
; ; } while (0)
;
304 ierr = PetscOptionsHead(PetscOptionsObject,"Redundant options");CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),304,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
305 ierr = PetscOptionsInt("-pc_redundant_number","Number of redundant pc","PCRedundantSetNumber",red->nsubcomm,&red->nsubcomm,0)PetscOptionsInt_Private(PetscOptionsObject,"-pc_redundant_number"
,"Number of redundant pc","PCRedundantSetNumber",red->nsubcomm
,&red->nsubcomm,0,(-2147483647 - 1),2147483647)
;CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),305,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
306 ierr = PetscOptionsTail()0; do {if (PetscOptionsObject->count != 1) do { do { ; if (
petscstack && petscstack->currentsize > 0) { petscstack
->currentsize--; petscstack->function[petscstack->currentsize
] = 0; petscstack->file[petscstack->currentsize] = 0; petscstack
->line[petscstack->currentsize] = 0; petscstack->petscroutine
[petscstack->currentsize] = PETSC_FALSE; } if (petscstack)
{ petscstack->hotdepth = (((petscstack->hotdepth-1)<
(0)) ? (0) : (petscstack->hotdepth-1)); } ; } while (0); return
(0);} while (0);} while(0)
;CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),306,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
307 PetscFunctionReturn(0)do { do { ; if (petscstack && petscstack->currentsize
> 0) { petscstack->currentsize--; petscstack->function
[petscstack->currentsize] = 0; petscstack->file[petscstack
->currentsize] = 0; petscstack->line[petscstack->currentsize
] = 0; petscstack->petscroutine[petscstack->currentsize
] = PETSC_FALSE; } if (petscstack) { petscstack->hotdepth =
(((petscstack->hotdepth-1)<(0)) ? (0) : (petscstack->
hotdepth-1)); } ; } while (0); return(0);} while (0)
;
308}
309
310static PetscErrorCode PCRedundantSetNumber_Redundant(PC pc,PetscInt nreds)
311{
312 PC_Redundant *red = (PC_Redundant*)pc->data;
313
314 PetscFunctionBegindo { do { ; if (petscstack && (petscstack->currentsize
< 64)) { petscstack->function[petscstack->currentsize
] = __func__; petscstack->file[petscstack->currentsize]
= "/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
; petscstack->line[petscstack->currentsize] = 314; petscstack
->petscroutine[petscstack->currentsize] = PETSC_TRUE; petscstack
->currentsize++; } if (petscstack) { petscstack->hotdepth
+= (PETSC_FALSE || petscstack->hotdepth); } ; } while (0)
; ; } while (0)
;
315 red->nsubcomm = nreds;
316 PetscFunctionReturn(0)do { do { ; if (petscstack && petscstack->currentsize
> 0) { petscstack->currentsize--; petscstack->function
[petscstack->currentsize] = 0; petscstack->file[petscstack
->currentsize] = 0; petscstack->line[petscstack->currentsize
] = 0; petscstack->petscroutine[petscstack->currentsize
] = PETSC_FALSE; } if (petscstack) { petscstack->hotdepth =
(((petscstack->hotdepth-1)<(0)) ? (0) : (petscstack->
hotdepth-1)); } ; } while (0); return(0);} while (0)
;
317}
318
319/*@
320 PCRedundantSetNumber - Sets the number of redundant preconditioner contexts.
321
322 Logically Collective on PC
323
324 Input Parameters:
325+ pc - the preconditioner context
326- nredundant - number of redundant preconditioner contexts; for example if you are using 64 MPI processes and
327 use an nredundant of 4 there will be 4 parallel solves each on 16 = 64/4 processes.
328
329 Level: advanced
330
331@*/
332PetscErrorCode PCRedundantSetNumber(PC pc,PetscInt nredundant)
333{
334 PetscErrorCode ierr;
335
336 PetscFunctionBegindo { do { ; if (petscstack && (petscstack->currentsize
< 64)) { petscstack->function[petscstack->currentsize
] = __func__; petscstack->file[petscstack->currentsize]
= "/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
; petscstack->line[petscstack->currentsize] = 336; petscstack
->petscroutine[petscstack->currentsize] = PETSC_TRUE; petscstack
->currentsize++; } if (petscstack) { petscstack->hotdepth
+= (PETSC_FALSE || petscstack->hotdepth); } ; } while (0)
; ; } while (0)
;
337 PetscValidHeaderSpecific(pc,PC_CLASSID,1)do { if (!pc) return PetscError(((MPI_Comm)0x44000001),337,__func__
,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,85,PETSC_ERROR_INITIAL,"Null Object: Parameter # %d",1); if (
!PetscCheckPointer(pc,PETSC_OBJECT)) return PetscError(((MPI_Comm
)0x44000001),337,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,64,PETSC_ERROR_INITIAL,"Invalid Pointer to Object: Parameter # %d"
,1); if (((PetscObject)(pc))->classid != PC_CLASSID) { if (
((PetscObject)(pc))->classid == -1) return PetscError(((MPI_Comm
)0x44000001),337,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,64,PETSC_ERROR_INITIAL,"Object already free: Parameter # %d"
,1); else return PetscError(((MPI_Comm)0x44000001),337,__func__
,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,62,PETSC_ERROR_INITIAL,"Wrong type of object: Parameter # %d"
,1); } } while (0)
;
338 if (nredundant <= 0) SETERRQ1(PetscObjectComm((PetscObject)pc),PETSC_ERR_ARG_WRONG, "num of redundant pc %D must be positive",nredundant)return PetscError(PetscObjectComm((PetscObject)pc),338,__func__
,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,62,PETSC_ERROR_INITIAL,"num of redundant pc %D must be positive"
,nredundant)
;
339 ierr = PetscTryMethod(pc,"PCRedundantSetNumber_C",(PC,PetscInt),(pc,nredundant))0; do { PetscErrorCode (*_7_f)(PC,PetscInt), _7_ierr; _7_ierr
= PetscObjectQueryFunction_Private(((PetscObject)(pc)),("PCRedundantSetNumber_C"
),(PetscVoidFunction*)(&_7_f));do {if (__builtin_expect(!
!(_7_ierr),0)) return PetscError(((MPI_Comm)0x44000001),339,__func__
,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,_7_ierr,PETSC_ERROR_REPEAT," ");} while (0); if (_7_f) {_7_ierr
= (*_7_f)(pc,nredundant);do {if (__builtin_expect(!!(_7_ierr
),0)) return PetscError(((MPI_Comm)0x44000001),339,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,_7_ierr,PETSC_ERROR_REPEAT," ");} while (0);} } while(0)
;CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),339,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
340 PetscFunctionReturn(0)do { do { ; if (petscstack && petscstack->currentsize
> 0) { petscstack->currentsize--; petscstack->function
[petscstack->currentsize] = 0; petscstack->file[petscstack
->currentsize] = 0; petscstack->line[petscstack->currentsize
] = 0; petscstack->petscroutine[petscstack->currentsize
] = PETSC_FALSE; } if (petscstack) { petscstack->hotdepth =
(((petscstack->hotdepth-1)<(0)) ? (0) : (petscstack->
hotdepth-1)); } ; } while (0); return(0);} while (0)
;
341}
342
343static PetscErrorCode PCRedundantSetScatter_Redundant(PC pc,VecScatter in,VecScatter out)
344{
345 PC_Redundant *red = (PC_Redundant*)pc->data;
346 PetscErrorCode ierr;
347
348 PetscFunctionBegindo { do { ; if (petscstack && (petscstack->currentsize
< 64)) { petscstack->function[petscstack->currentsize
] = __func__; petscstack->file[petscstack->currentsize]
= "/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
; petscstack->line[petscstack->currentsize] = 348; petscstack
->petscroutine[petscstack->currentsize] = PETSC_TRUE; petscstack
->currentsize++; } if (petscstack) { petscstack->hotdepth
+= (PETSC_FALSE || petscstack->hotdepth); } ; } while (0)
; ; } while (0)
;
349 ierr = PetscObjectReference((PetscObject)in);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),349,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
350 ierr = VecScatterDestroy(&red->scatterin);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),350,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
351
352 red->scatterin = in;
353
354 ierr = PetscObjectReference((PetscObject)out);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),354,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
355 ierr = VecScatterDestroy(&red->scatterout);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),355,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
356 red->scatterout = out;
357 PetscFunctionReturn(0)do { do { ; if (petscstack && petscstack->currentsize
> 0) { petscstack->currentsize--; petscstack->function
[petscstack->currentsize] = 0; petscstack->file[petscstack
->currentsize] = 0; petscstack->line[petscstack->currentsize
] = 0; petscstack->petscroutine[petscstack->currentsize
] = PETSC_FALSE; } if (petscstack) { petscstack->hotdepth =
(((petscstack->hotdepth-1)<(0)) ? (0) : (petscstack->
hotdepth-1)); } ; } while (0); return(0);} while (0)
;
358}
359
360/*@
361 PCRedundantSetScatter - Sets the scatter used to copy values into the
362 redundant local solve and the scatter to move them back into the global
363 vector.
364
365 Logically Collective on PC
366
367 Input Parameters:
368+ pc - the preconditioner context
369. in - the scatter to move the values in
370- out - the scatter to move them out
371
372 Level: advanced
373
374@*/
375PetscErrorCode PCRedundantSetScatter(PC pc,VecScatter in,VecScatter out)
376{
377 PetscErrorCode ierr;
378
379 PetscFunctionBegindo { do { ; if (petscstack && (petscstack->currentsize
< 64)) { petscstack->function[petscstack->currentsize
] = __func__; petscstack->file[petscstack->currentsize]
= "/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
; petscstack->line[petscstack->currentsize] = 379; petscstack
->petscroutine[petscstack->currentsize] = PETSC_TRUE; petscstack
->currentsize++; } if (petscstack) { petscstack->hotdepth
+= (PETSC_FALSE || petscstack->hotdepth); } ; } while (0)
; ; } while (0)
;
380 PetscValidHeaderSpecific(pc,PC_CLASSID,1)do { if (!pc) return PetscError(((MPI_Comm)0x44000001),380,__func__
,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,85,PETSC_ERROR_INITIAL,"Null Object: Parameter # %d",1); if (
!PetscCheckPointer(pc,PETSC_OBJECT)) return PetscError(((MPI_Comm
)0x44000001),380,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,64,PETSC_ERROR_INITIAL,"Invalid Pointer to Object: Parameter # %d"
,1); if (((PetscObject)(pc))->classid != PC_CLASSID) { if (
((PetscObject)(pc))->classid == -1) return PetscError(((MPI_Comm
)0x44000001),380,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,64,PETSC_ERROR_INITIAL,"Object already free: Parameter # %d"
,1); else return PetscError(((MPI_Comm)0x44000001),380,__func__
,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,62,PETSC_ERROR_INITIAL,"Wrong type of object: Parameter # %d"
,1); } } while (0)
;
381 PetscValidHeaderSpecific(in,VEC_SCATTER_CLASSID,2)do { if (!in) return PetscError(((MPI_Comm)0x44000001),381,__func__
,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,85,PETSC_ERROR_INITIAL,"Null Object: Parameter # %d",2); if (
!PetscCheckPointer(in,PETSC_OBJECT)) return PetscError(((MPI_Comm
)0x44000001),381,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,64,PETSC_ERROR_INITIAL,"Invalid Pointer to Object: Parameter # %d"
,2); if (((PetscObject)(in))->classid != VEC_SCATTER_CLASSID
) { if (((PetscObject)(in))->classid == -1) return PetscError
(((MPI_Comm)0x44000001),381,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,64,PETSC_ERROR_INITIAL,"Object already free: Parameter # %d"
,2); else return PetscError(((MPI_Comm)0x44000001),381,__func__
,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,62,PETSC_ERROR_INITIAL,"Wrong type of object: Parameter # %d"
,2); } } while (0)
;
382 PetscValidHeaderSpecific(out,VEC_SCATTER_CLASSID,3)do { if (!out) return PetscError(((MPI_Comm)0x44000001),382,__func__
,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,85,PETSC_ERROR_INITIAL,"Null Object: Parameter # %d",3); if (
!PetscCheckPointer(out,PETSC_OBJECT)) return PetscError(((MPI_Comm
)0x44000001),382,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,64,PETSC_ERROR_INITIAL,"Invalid Pointer to Object: Parameter # %d"
,3); if (((PetscObject)(out))->classid != VEC_SCATTER_CLASSID
) { if (((PetscObject)(out))->classid == -1) return PetscError
(((MPI_Comm)0x44000001),382,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,64,PETSC_ERROR_INITIAL,"Object already free: Parameter # %d"
,3); else return PetscError(((MPI_Comm)0x44000001),382,__func__
,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,62,PETSC_ERROR_INITIAL,"Wrong type of object: Parameter # %d"
,3); } } while (0)
;
383 ierr = PetscTryMethod(pc,"PCRedundantSetScatter_C",(PC,VecScatter,VecScatter),(pc,in,out))0; do { PetscErrorCode (*_7_f)(PC,VecScatter,VecScatter), _7_ierr
; _7_ierr = PetscObjectQueryFunction_Private(((PetscObject)(pc
)),("PCRedundantSetScatter_C"),(PetscVoidFunction*)(&_7_f
));do {if (__builtin_expect(!!(_7_ierr),0)) return PetscError
(((MPI_Comm)0x44000001),383,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,_7_ierr,PETSC_ERROR_REPEAT," ");} while (0); if (_7_f) {_7_ierr
= (*_7_f)(pc,in,out);do {if (__builtin_expect(!!(_7_ierr),0)
) return PetscError(((MPI_Comm)0x44000001),383,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,_7_ierr,PETSC_ERROR_REPEAT," ");} while (0);} } while(0)
;CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),383,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
384 PetscFunctionReturn(0)do { do { ; if (petscstack && petscstack->currentsize
> 0) { petscstack->currentsize--; petscstack->function
[petscstack->currentsize] = 0; petscstack->file[petscstack
->currentsize] = 0; petscstack->line[petscstack->currentsize
] = 0; petscstack->petscroutine[petscstack->currentsize
] = PETSC_FALSE; } if (petscstack) { petscstack->hotdepth =
(((petscstack->hotdepth-1)<(0)) ? (0) : (petscstack->
hotdepth-1)); } ; } while (0); return(0);} while (0)
;
385}
386
387static PetscErrorCode PCRedundantGetKSP_Redundant(PC pc,KSP *innerksp)
388{
389 PetscErrorCode ierr;
390 PC_Redundant *red = (PC_Redundant*)pc->data;
391 MPI_Comm comm,subcomm;
392 const char *prefix;
393 PetscBool issbaij;
394
395 PetscFunctionBegindo { do { ; if (petscstack && (petscstack->currentsize
< 64)) { petscstack->function[petscstack->currentsize
] = __func__; petscstack->file[petscstack->currentsize]
= "/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
; petscstack->line[petscstack->currentsize] = 395; petscstack
->petscroutine[petscstack->currentsize] = PETSC_TRUE; petscstack
->currentsize++; } if (petscstack) { petscstack->hotdepth
+= (PETSC_FALSE || petscstack->hotdepth); } ; } while (0)
; ; } while (0)
;
396 if (!red->psubcomm) {
397 ierr = PCGetOptionsPrefix(pc,&prefix);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),397,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
398
399 ierr = PetscObjectGetComm((PetscObject)pc,&comm);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),399,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
400 ierr = PetscSubcommCreate(comm,&red->psubcomm);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),400,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
401 ierr = PetscSubcommSetNumber(red->psubcomm,red->nsubcomm);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),401,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
402 ierr = PetscSubcommSetType(red->psubcomm,PETSC_SUBCOMM_CONTIGUOUS);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),402,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
403
404 ierr = PetscSubcommSetOptionsPrefix(red->psubcomm,prefix);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),404,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
405 ierr = PetscSubcommSetFromOptions(red->psubcomm);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),405,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
406 ierr = PetscLogObjectMemory((PetscObject)pc,sizeof(PetscSubcomm));CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),406,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
407
408 /* create a new PC that processors in each subcomm have copy of */
409 subcomm = PetscSubcommChild(red->psubcomm);
410
411 ierr = KSPCreate(subcomm,&red->ksp);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),411,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
412 ierr = KSPSetErrorIfNotConverged(red->ksp,pc->erroriffailure);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),412,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
413 ierr = PetscObjectIncrementTabLevel((PetscObject)red->ksp,(PetscObject)pc,1);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),413,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
414 ierr = PetscLogObjectParent((PetscObject)pc,(PetscObject)red->ksp);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),414,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
415 ierr = KSPSetType(red->ksp,KSPPREONLY"preonly");CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),415,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
416 ierr = KSPGetPC(red->ksp,&red->pc);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),416,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
417 ierr = PetscObjectTypeCompare((PetscObject)pc->pmat,MATSEQSBAIJ"seqsbaij",&issbaij);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),417,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
418 if (!issbaij) {
419 ierr = PetscObjectTypeCompare((PetscObject)pc->pmat,MATMPISBAIJ"mpisbaij",&issbaij);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),419,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
420 }
421 if (!issbaij) {
422 ierr = PCSetType(red->pc,PCLU"lu");CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),422,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
423 } else {
424 ierr = PCSetType(red->pc,PCCHOLESKY"cholesky");CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),424,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
425 }
426 if (red->shifttypeset) {
427 ierr = PCFactorSetShiftType(red->pc,red->shifttype);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),427,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
428 red->shifttypeset = PETSC_FALSE;
429 }
430 ierr = KSPSetOptionsPrefix(red->ksp,prefix);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),430,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
431 ierr = KSPAppendOptionsPrefix(red->ksp,"redundant_");CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),431,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
432 }
433 *innerksp = red->ksp;
434 PetscFunctionReturn(0)do { do { ; if (petscstack && petscstack->currentsize
> 0) { petscstack->currentsize--; petscstack->function
[petscstack->currentsize] = 0; petscstack->file[petscstack
->currentsize] = 0; petscstack->line[petscstack->currentsize
] = 0; petscstack->petscroutine[petscstack->currentsize
] = PETSC_FALSE; } if (petscstack) { petscstack->hotdepth =
(((petscstack->hotdepth-1)<(0)) ? (0) : (petscstack->
hotdepth-1)); } ; } while (0); return(0);} while (0)
;
435}
436
437/*@
438 PCRedundantGetKSP - Gets the less parallel KSP created by the redundant PC.
439
440 Not Collective
441
442 Input Parameter:
443. pc - the preconditioner context
444
445 Output Parameter:
446. innerksp - the KSP on the smaller set of processes
447
448 Level: advanced
449
450@*/
451PetscErrorCode PCRedundantGetKSP(PC pc,KSP *innerksp)
452{
453 PetscErrorCode ierr;
454
455 PetscFunctionBegindo { do { ; if (petscstack && (petscstack->currentsize
< 64)) { petscstack->function[petscstack->currentsize
] = __func__; petscstack->file[petscstack->currentsize]
= "/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
; petscstack->line[petscstack->currentsize] = 455; petscstack
->petscroutine[petscstack->currentsize] = PETSC_TRUE; petscstack
->currentsize++; } if (petscstack) { petscstack->hotdepth
+= (PETSC_FALSE || petscstack->hotdepth); } ; } while (0)
; ; } while (0)
;
456 PetscValidHeaderSpecific(pc,PC_CLASSID,1)do { if (!pc) return PetscError(((MPI_Comm)0x44000001),456,__func__
,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,85,PETSC_ERROR_INITIAL,"Null Object: Parameter # %d",1); if (
!PetscCheckPointer(pc,PETSC_OBJECT)) return PetscError(((MPI_Comm
)0x44000001),456,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,64,PETSC_ERROR_INITIAL,"Invalid Pointer to Object: Parameter # %d"
,1); if (((PetscObject)(pc))->classid != PC_CLASSID) { if (
((PetscObject)(pc))->classid == -1) return PetscError(((MPI_Comm
)0x44000001),456,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,64,PETSC_ERROR_INITIAL,"Object already free: Parameter # %d"
,1); else return PetscError(((MPI_Comm)0x44000001),456,__func__
,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,62,PETSC_ERROR_INITIAL,"Wrong type of object: Parameter # %d"
,1); } } while (0)
;
457 PetscValidPointer(innerksp,2)do { if (!innerksp) return PetscError(((MPI_Comm)0x44000001),
457,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,85,PETSC_ERROR_INITIAL,"Null Pointer: Parameter # %d",2); if
(!PetscCheckPointer(innerksp,PETSC_CHAR)) return PetscError(
((MPI_Comm)0x44000001),457,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,68,PETSC_ERROR_INITIAL,"Invalid Pointer: Parameter # %d",2);
} while (0)
;
458 ierr = PetscUseMethod(pc,"PCRedundantGetKSP_C",(PC,KSP*),(pc,innerksp))0; do { PetscErrorCode (*_7_f)(PC,KSP*), _7_ierr; _7_ierr = PetscObjectQueryFunction_Private
(((PetscObject)(pc)),("PCRedundantGetKSP_C"),(PetscVoidFunction
*)(&_7_f));do {if (__builtin_expect(!!(_7_ierr),0)) return
PetscError(((MPI_Comm)0x44000001),458,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,_7_ierr,PETSC_ERROR_REPEAT," ");} while (0); if (_7_f) {_7_ierr
= (*_7_f)(pc,innerksp);do {if (__builtin_expect(!!(_7_ierr),
0)) return PetscError(((MPI_Comm)0x44000001),458,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,_7_ierr,PETSC_ERROR_REPEAT," ");} while (0);} else return PetscError
(PetscObjectComm((PetscObject)(pc)),458,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,56,PETSC_ERROR_INITIAL,"Cannot locate function %s in object"
,"PCRedundantGetKSP_C"); } while(0)
;CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),458,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
459 PetscFunctionReturn(0)do { do { ; if (petscstack && petscstack->currentsize
> 0) { petscstack->currentsize--; petscstack->function
[petscstack->currentsize] = 0; petscstack->file[petscstack
->currentsize] = 0; petscstack->line[petscstack->currentsize
] = 0; petscstack->petscroutine[petscstack->currentsize
] = PETSC_FALSE; } if (petscstack) { petscstack->hotdepth =
(((petscstack->hotdepth-1)<(0)) ? (0) : (petscstack->
hotdepth-1)); } ; } while (0); return(0);} while (0)
;
460}
461
462static PetscErrorCode PCRedundantGetOperators_Redundant(PC pc,Mat *mat,Mat *pmat)
463{
464 PC_Redundant *red = (PC_Redundant*)pc->data;
465
466 PetscFunctionBegindo { do { ; if (petscstack && (petscstack->currentsize
< 64)) { petscstack->function[petscstack->currentsize
] = __func__; petscstack->file[petscstack->currentsize]
= "/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
; petscstack->line[petscstack->currentsize] = 466; petscstack
->petscroutine[petscstack->currentsize] = PETSC_TRUE; petscstack
->currentsize++; } if (petscstack) { petscstack->hotdepth
+= (PETSC_FALSE || petscstack->hotdepth); } ; } while (0)
; ; } while (0)
;
467 if (mat) *mat = red->pmats;
468 if (pmat) *pmat = red->pmats;
469 PetscFunctionReturn(0)do { do { ; if (petscstack && petscstack->currentsize
> 0) { petscstack->currentsize--; petscstack->function
[petscstack->currentsize] = 0; petscstack->file[petscstack
->currentsize] = 0; petscstack->line[petscstack->currentsize
] = 0; petscstack->petscroutine[petscstack->currentsize
] = PETSC_FALSE; } if (petscstack) { petscstack->hotdepth =
(((petscstack->hotdepth-1)<(0)) ? (0) : (petscstack->
hotdepth-1)); } ; } while (0); return(0);} while (0)
;
470}
471
472/*@
473 PCRedundantGetOperators - gets the sequential matrix and preconditioner matrix
474
475 Not Collective
476
477 Input Parameter:
478. pc - the preconditioner context
479
480 Output Parameters:
481+ mat - the matrix
482- pmat - the (possibly different) preconditioner matrix
483
484 Level: advanced
485
486@*/
487PetscErrorCode PCRedundantGetOperators(PC pc,Mat *mat,Mat *pmat)
488{
489 PetscErrorCode ierr;
490
491 PetscFunctionBegindo { do { ; if (petscstack && (petscstack->currentsize
< 64)) { petscstack->function[petscstack->currentsize
] = __func__; petscstack->file[petscstack->currentsize]
= "/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
; petscstack->line[petscstack->currentsize] = 491; petscstack
->petscroutine[petscstack->currentsize] = PETSC_TRUE; petscstack
->currentsize++; } if (petscstack) { petscstack->hotdepth
+= (PETSC_FALSE || petscstack->hotdepth); } ; } while (0)
; ; } while (0)
;
492 PetscValidHeaderSpecific(pc,PC_CLASSID,1)do { if (!pc) return PetscError(((MPI_Comm)0x44000001),492,__func__
,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,85,PETSC_ERROR_INITIAL,"Null Object: Parameter # %d",1); if (
!PetscCheckPointer(pc,PETSC_OBJECT)) return PetscError(((MPI_Comm
)0x44000001),492,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,64,PETSC_ERROR_INITIAL,"Invalid Pointer to Object: Parameter # %d"
,1); if (((PetscObject)(pc))->classid != PC_CLASSID) { if (
((PetscObject)(pc))->classid == -1) return PetscError(((MPI_Comm
)0x44000001),492,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,64,PETSC_ERROR_INITIAL,"Object already free: Parameter # %d"
,1); else return PetscError(((MPI_Comm)0x44000001),492,__func__
,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,62,PETSC_ERROR_INITIAL,"Wrong type of object: Parameter # %d"
,1); } } while (0)
;
493 if (mat) PetscValidPointer(mat,2)do { if (!mat) return PetscError(((MPI_Comm)0x44000001),493,__func__
,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,85,PETSC_ERROR_INITIAL,"Null Pointer: Parameter # %d",2); if
(!PetscCheckPointer(mat,PETSC_CHAR)) return PetscError(((MPI_Comm
)0x44000001),493,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,68,PETSC_ERROR_INITIAL,"Invalid Pointer: Parameter # %d",2);
} while (0)
;
494 if (pmat) PetscValidPointer(pmat,3)do { if (!pmat) return PetscError(((MPI_Comm)0x44000001),494,
__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,85,PETSC_ERROR_INITIAL,"Null Pointer: Parameter # %d",3); if
(!PetscCheckPointer(pmat,PETSC_CHAR)) return PetscError(((MPI_Comm
)0x44000001),494,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,68,PETSC_ERROR_INITIAL,"Invalid Pointer: Parameter # %d",3);
} while (0)
;
495 ierr = PetscUseMethod(pc,"PCRedundantGetOperators_C",(PC,Mat*,Mat*),(pc,mat,pmat))0; do { PetscErrorCode (*_7_f)(PC,Mat*,Mat*), _7_ierr; _7_ierr
= PetscObjectQueryFunction_Private(((PetscObject)(pc)),("PCRedundantGetOperators_C"
),(PetscVoidFunction*)(&_7_f));do {if (__builtin_expect(!
!(_7_ierr),0)) return PetscError(((MPI_Comm)0x44000001),495,__func__
,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,_7_ierr,PETSC_ERROR_REPEAT," ");} while (0); if (_7_f) {_7_ierr
= (*_7_f)(pc,mat,pmat);do {if (__builtin_expect(!!(_7_ierr),
0)) return PetscError(((MPI_Comm)0x44000001),495,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,_7_ierr,PETSC_ERROR_REPEAT," ");} while (0);} else return PetscError
(PetscObjectComm((PetscObject)(pc)),495,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,56,PETSC_ERROR_INITIAL,"Cannot locate function %s in object"
,"PCRedundantGetOperators_C"); } while(0)
;CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),495,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
496 PetscFunctionReturn(0)do { do { ; if (petscstack && petscstack->currentsize
> 0) { petscstack->currentsize--; petscstack->function
[petscstack->currentsize] = 0; petscstack->file[petscstack
->currentsize] = 0; petscstack->line[petscstack->currentsize
] = 0; petscstack->petscroutine[petscstack->currentsize
] = PETSC_FALSE; } if (petscstack) { petscstack->hotdepth =
(((petscstack->hotdepth-1)<(0)) ? (0) : (petscstack->
hotdepth-1)); } ; } while (0); return(0);} while (0)
;
497}
498
499/* -------------------------------------------------------------------------------------*/
500/*MC
501 PCREDUNDANT - Runs a KSP solver with preconditioner for the entire problem on subgroups of processors
502
503 Options for the redundant preconditioners can be set with -redundant_pc_xxx for the redundant KSP with -redundant_ksp_xxx
504
505 Options Database:
506. -pc_redundant_number <n> - number of redundant solves, for example if you are using 64 MPI processes and
507 use an n of 4 there will be 4 parallel solves each on 16 = 64/4 processes.
508
509 Level: intermediate
510
511 Notes:
512 The default KSP is preonly and the default PC is LU or CHOLESKY if Pmat is of type MATSBAIJ.
513
514 PCFactorSetShiftType() applied to this PC will convey they shift type into the inner PC if it is factorization based.
515
516 Developer Notes:
517 Note that PCSetInitialGuessNonzero() is not used by this class but likely should be.
518
519.seealso: PCCreate(), PCSetType(), PCType (for list of available types), PCRedundantSetScatter(),
520 PCRedundantGetKSP(), PCRedundantGetOperators(), PCRedundantSetNumber()
521M*/
522
523PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PCCreate_Redundant(PC pc)
524{
525 PetscErrorCode ierr;
526 PC_Redundant *red;
527 PetscMPIInt size;
528
529 PetscFunctionBegindo { do { ; if (petscstack && (petscstack->currentsize
< 64)) { petscstack->function[petscstack->currentsize
] = __func__; petscstack->file[petscstack->currentsize]
= "/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
; petscstack->line[petscstack->currentsize] = 529; petscstack
->petscroutine[petscstack->currentsize] = PETSC_TRUE; petscstack
->currentsize++; } if (petscstack) { petscstack->hotdepth
+= (PETSC_FALSE || petscstack->hotdepth); } ; } while (0)
; ; } while (0)
;
530 ierr = PetscNewLog(pc,&red)(PetscMallocA(1,PETSC_TRUE,530,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,(size_t)(1)*sizeof(**(((&red)))),(((&red)))) || PetscLogObjectMemory
((PetscObject)pc,sizeof(**(&red))))
;CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),530,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
531 ierr = MPI_Comm_size(PetscObjectComm((PetscObject)pc),&size);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),531,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
532
533 red->nsubcomm = size;
534 red->useparallelmat = PETSC_TRUE;
535 pc->data = (void*)red;
536
537 pc->ops->apply = PCApply_Redundant;
538 pc->ops->applytranspose = PCApplyTranspose_Redundant;
539 pc->ops->setup = PCSetUp_Redundant;
540 pc->ops->destroy = PCDestroy_Redundant;
541 pc->ops->reset = PCReset_Redundant;
542 pc->ops->setfromoptions = PCSetFromOptions_Redundant;
543 pc->ops->view = PCView_Redundant;
544
545 ierr = PetscObjectComposeFunction((PetscObject)pc,"PCRedundantSetScatter_C",PCRedundantSetScatter_Redundant)PetscObjectComposeFunction_Private((PetscObject)pc,"PCRedundantSetScatter_C"
,(PetscVoidFunction)(PCRedundantSetScatter_Redundant))
;CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),545,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
546 ierr = PetscObjectComposeFunction((PetscObject)pc,"PCRedundantSetNumber_C",PCRedundantSetNumber_Redundant)PetscObjectComposeFunction_Private((PetscObject)pc,"PCRedundantSetNumber_C"
,(PetscVoidFunction)(PCRedundantSetNumber_Redundant))
;CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),546,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
547 ierr = PetscObjectComposeFunction((PetscObject)pc,"PCRedundantGetKSP_C",PCRedundantGetKSP_Redundant)PetscObjectComposeFunction_Private((PetscObject)pc,"PCRedundantGetKSP_C"
,(PetscVoidFunction)(PCRedundantGetKSP_Redundant))
;CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),547,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
548 ierr = PetscObjectComposeFunction((PetscObject)pc,"PCRedundantGetOperators_C",PCRedundantGetOperators_Redundant)PetscObjectComposeFunction_Private((PetscObject)pc,"PCRedundantGetOperators_C"
,(PetscVoidFunction)(PCRedundantGetOperators_Redundant))
;CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),548,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
549 ierr = PetscObjectComposeFunction((PetscObject)pc,"PCFactorSetShiftType_C",PCFactorSetShiftType_Redundant)PetscObjectComposeFunction_Private((PetscObject)pc,"PCFactorSetShiftType_C"
,(PetscVoidFunction)(PCFactorSetShiftType_Redundant))
;CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),549,__func__,"/sandbox/petsc/petsc.next/src/ksp/pc/impls/redundant/redundant.c"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
550 PetscFunctionReturn(0)do { do { ; if (petscstack && petscstack->currentsize
> 0) { petscstack->currentsize--; petscstack->function
[petscstack->currentsize] = 0; petscstack->file[petscstack
->currentsize] = 0; petscstack->line[petscstack->currentsize
] = 0; petscstack->petscroutine[petscstack->currentsize
] = PETSC_FALSE; } if (petscstack) { petscstack->hotdepth =
(((petscstack->hotdepth-1)<(0)) ? (0) : (petscstack->
hotdepth-1)); } ; } while (0); return(0);} while (0)
;
551}
552

/sandbox/petsc/petsc.next/include/petscsys.h

1/*
2 This is the main PETSc include file (for C and C++). It is included by all
3 other PETSc include files, so it almost never has to be specifically included.
4*/
5#if !defined(PETSCSYS_H)
6#define PETSCSYS_H
7/* ========================================================================== */
8/*
9 petscconf.h is contained in ${PETSC_ARCH}/include/petscconf.h it is
10 found automatically by the compiler due to the -I${PETSC_DIR}/${PETSC_ARCH}/include.
11 For --prefix installs the ${PETSC_ARCH}/ does not exist and petscconf.h is in the same
12 directory as the other PETSc include files.
13*/
14#include <petscconf.h>
15#include <petscfix.h>
16
17#if defined(PETSC_DESIRE_FEATURE_TEST_MACROS)
18/*
19 Feature test macros must be included before headers defined by IEEE Std 1003.1-2001
20 We only turn these in PETSc source files that require them by setting PETSC_DESIRE_FEATURE_TEST_MACROS
21*/
22# if defined(PETSC__POSIX_C_SOURCE_200112L) && !defined(_POSIX_C_SOURCE200809L)
23# define _POSIX_C_SOURCE200809L 200112L
24# endif
25# if defined(PETSC__BSD_SOURCE1) && !defined(_BSD_SOURCE1)
26# define _BSD_SOURCE1
27# endif
28# if defined(PETSC__DEFAULT_SOURCE1) && !defined(_DEFAULT_SOURCE1)
29# define _DEFAULT_SOURCE1
30# endif
31# if defined(PETSC__GNU_SOURCE1) && !defined(_GNU_SOURCE)
32# define _GNU_SOURCE
33# endif
34#endif
35
36#include <petscsystypes.h>
37
38/* ========================================================================== */
39/*
40 This facilitates using the C version of PETSc from C++ and the C++ version from C.
41*/
42#if defined(__cplusplus)
43# define PETSC_FUNCTION_NAME__func__ PETSC_FUNCTION_NAME_CXX__func__
44#else
45# define PETSC_FUNCTION_NAME__func__ PETSC_FUNCTION_NAME_C__func__
46#endif
47
48/* ========================================================================== */
49/*
50 Since PETSc manages its own extern "C" handling users should never include PETSc include
51 files within extern "C". This will generate a compiler error if a user does put the include
52 file within an extern "C".
53*/
54#if defined(__cplusplus)
55void assert_never_put_petsc_headers_inside_an_extern_c(int); void assert_never_put_petsc_headers_inside_an_extern_c(double);
56#endif
57
58#if defined(__cplusplus)
59# define PETSC_RESTRICT__restrict PETSC_CXX_RESTRICT__restrict
60#else
61# define PETSC_RESTRICT__restrict PETSC_C_RESTRICT__restrict
62#endif
63
64#if defined(__cplusplus)
65# define PETSC_INLINEinline PETSC_CXX_INLINEinline
66#else
67# define PETSC_INLINEinline PETSC_C_INLINEinline
68#endif
69
70#define PETSC_STATIC_INLINEstatic inline static PETSC_INLINEinline
71
72#if defined(_WIN32) && defined(PETSC_USE_SHARED_LIBRARIES1) /* For Win32 shared libraries */
73# define PETSC_DLLEXPORT__attribute__((visibility ("default"))) __declspec(dllexport)
74# define PETSC_DLLIMPORT__attribute__((visibility ("default"))) __declspec(dllimport)
75# define PETSC_VISIBILITY_INTERNAL__attribute__((visibility ("hidden")))
76#elif defined(PETSC_USE_VISIBILITY_CXX1) && defined(__cplusplus)
77# define PETSC_DLLEXPORT__attribute__((visibility ("default"))) __attribute__((visibility ("default")))
78# define PETSC_DLLIMPORT__attribute__((visibility ("default"))) __attribute__((visibility ("default")))
79# define PETSC_VISIBILITY_INTERNAL__attribute__((visibility ("hidden"))) __attribute__((visibility ("hidden")))
80#elif defined(PETSC_USE_VISIBILITY_C1) && !defined(__cplusplus)
81# define PETSC_DLLEXPORT__attribute__((visibility ("default"))) __attribute__((visibility ("default")))
82# define PETSC_DLLIMPORT__attribute__((visibility ("default"))) __attribute__((visibility ("default")))
83# define PETSC_VISIBILITY_INTERNAL__attribute__((visibility ("hidden"))) __attribute__((visibility ("hidden")))
84#else
85# define PETSC_DLLEXPORT__attribute__((visibility ("default")))
86# define PETSC_DLLIMPORT__attribute__((visibility ("default")))
87# define PETSC_VISIBILITY_INTERNAL__attribute__((visibility ("hidden")))
88#endif
89
90#if defined(petsc_EXPORTS) /* CMake defines this when building the shared library */
91# define PETSC_VISIBILITY_PUBLIC__attribute__((visibility ("default"))) PETSC_DLLEXPORT__attribute__((visibility ("default")))
92#else /* Win32 users need this to import symbols from petsc.dll */
93# define PETSC_VISIBILITY_PUBLIC__attribute__((visibility ("default"))) PETSC_DLLIMPORT__attribute__((visibility ("default")))
94#endif
95
96/*
97 Functions tagged with PETSC_EXTERN in the header files are
98 always defined as extern "C" when compiled with C++ so they may be
99 used from C and are always visible in the shared libraries
100*/
101#if defined(__cplusplus)
102# define PETSC_EXTERNextern __attribute__((visibility ("default"))) extern "C" PETSC_VISIBILITY_PUBLIC__attribute__((visibility ("default")))
103# define PETSC_EXTERN_TYPEDEF extern "C"
104# define PETSC_INTERNextern __attribute__((visibility ("hidden"))) extern "C" PETSC_VISIBILITY_INTERNAL__attribute__((visibility ("hidden")))
105#else
106# define PETSC_EXTERNextern __attribute__((visibility ("default"))) extern PETSC_VISIBILITY_PUBLIC__attribute__((visibility ("default")))
107# define PETSC_EXTERN_TYPEDEF
108# define PETSC_INTERNextern __attribute__((visibility ("hidden"))) extern PETSC_VISIBILITY_INTERNAL__attribute__((visibility ("hidden")))
109#endif
110
111#include <petscversion.h>
112#define PETSC_AUTHOR_INFO" The PETSc Team\n [email protected]\n https://www.mcs.anl.gov/petsc/\n" " The PETSc Team\n [email protected]\n https://www.mcs.anl.gov/petsc/\n"
113
114/* ========================================================================== */
115
116/*
117 Defines the interface to MPI allowing the use of all MPI functions.
118
119 PETSc does not use the C++ binding of MPI at ALL. The following flag
120 makes sure the C++ bindings are not included. The C++ bindings REQUIRE
121 putting mpi.h before ANY C++ include files, we cannot control this
122 with all PETSc users. Users who want to use the MPI C++ bindings can include
123 mpicxx.h directly in their code
124*/
125#if !defined(MPICH_SKIP_MPICXX1)
126# define MPICH_SKIP_MPICXX1 1
127#endif
128#if !defined(OMPI_SKIP_MPICXX1)
129# define OMPI_SKIP_MPICXX1 1
130#endif
131#if defined(PETSC_HAVE_MPIUNI)
132# include <petsc/mpiuni/mpi.h>
133#else
134# include <mpi.h>
135#endif
136
137/*
138 Perform various sanity checks that the correct mpi.h is being included at compile time.
139 This usually happens because
140 * either an unexpected mpi.h is in the default compiler path (i.e. in /usr/include) or
141 * an extra include path -I/something (which contains the unexpected mpi.h) is being passed to the compiler
142*/
143#if defined(PETSC_HAVE_MPIUNI)
144# if !defined(MPIUNI_H)
145# error "PETSc was configured with --with-mpi=0 but now appears to be compiling using a different mpi.h"
146# endif
147#elif defined(PETSC_HAVE_I_MPI_NUMVERSION)
148# if !defined(I_MPI_NUMVERSION)
149# error "PETSc was configured with I_MPI but now appears to be compiling using a non-I_MPI mpi.h"
150# elif I_MPI_NUMVERSION != PETSC_HAVE_I_MPI_NUMVERSION
151# error "PETSc was configured with one I_MPI mpi.h version but now appears to be compiling using a different I_MPI mpi.h version"
152# endif
153#elif defined(PETSC_HAVE_MVAPICH2_NUMVERSION)
154# if !defined(MVAPICH2_NUMVERSION)
155# error "PETSc was configured with MVAPICH2 but now appears to be compiling using a non-MVAPICH2 mpi.h"
156# elif MVAPICH2_NUMVERSION != PETSC_HAVE_MVAPICH2_NUMVERSION
157# error "PETSc was configured with one MVAPICH2 mpi.h version but now appears to be compiling using a different MVAPICH2 mpi.h version"
158# endif
159#elif defined(PETSC_HAVE_MPICH_NUMVERSION30301300)
160# if !defined(MPICH_NUMVERSION30301300) || defined(MVAPICH2_NUMVERSION) || defined(I_MPI_NUMVERSION)
161# error "PETSc was configured with MPICH but now appears to be compiling using a non-MPICH mpi.h"
162# elif (MPICH_NUMVERSION30301300/100000 != PETSC_HAVE_MPICH_NUMVERSION30301300/100000) || (MPICH_NUMVERSION30301300%100000/1000 < PETSC_HAVE_MPICH_NUMVERSION30301300%100000/1000)
163# error "PETSc was configured with one MPICH mpi.h version but now appears to be compiling using a different MPICH mpi.h version"
164# endif
165#elif defined(PETSC_HAVE_OMPI_MAJOR_VERSION)
166# if !defined(OMPI_MAJOR_VERSION)
167# error "PETSc was configured with OpenMPI but now appears to be compiling using a non-OpenMPI mpi.h"
168# elif (OMPI_MAJOR_VERSION != PETSC_HAVE_OMPI_MAJOR_VERSION) || (OMPI_MINOR_VERSION != PETSC_HAVE_OMPI_MINOR_VERSION) || (OMPI_RELEASE_VERSION < PETSC_HAVE_OMPI_RELEASE_VERSION)
169# error "PETSc was configured with one OpenMPI mpi.h version but now appears to be compiling using a different OpenMPI mpi.h version"
170# endif
171#elif defined(OMPI_MAJOR_VERSION) || defined(MPICH_NUMVERSION30301300)
172# error "PETSc was configured with undetermined MPI - but now appears to be compiling using either of OpenMPI or a MPICH variant"
173#endif
174
175/*
176 Need to put stdio.h AFTER mpi.h for MPICH2 with C++ compiler
177 see the top of mpicxx.h in the MPICH2 distribution.
178*/
179#include <stdio.h>
180
181/* MSMPI on 32bit windows requires this yukky hack - that breaks MPI standard compliance */
182#if !defined(MPIAPI)
183#define MPIAPI
184#endif
185
186/*
187 Support for Clang (>=3.2) matching type tag arguments with void* buffer types.
188 This allows the compiler to detect cases where the MPI datatype argument passed to a MPI routine
189 does not match the actual type of the argument being passed in
190*/
191#if defined(__has_attribute)0 && defined(works_with_const_which_is_not_true)
192# if __has_attribute(argument_with_type_tag)1 && __has_attribute(pointer_with_type_tag)1 && __has_attribute(type_tag_for_datatype)1
193# define PetscAttrMPIPointerWithType(bufno,typeno) __attribute__((pointer_with_type_tag(MPI,bufno,typeno)))
194# define PetscAttrMPITypeTag(type) __attribute__((type_tag_for_datatype(MPI,type)))
195# define PetscAttrMPITypeTagLayoutCompatible(type) __attribute__((type_tag_for_datatype(MPI,type,layout_compatible)))
196# endif
197#endif
198#if !defined(PetscAttrMPIPointerWithType)
199# define PetscAttrMPIPointerWithType(bufno,typeno)
200# define PetscAttrMPITypeTag(type)
201# define PetscAttrMPITypeTagLayoutCompatible(type)
202#endif
203
204PETSC_EXTERNextern __attribute__((visibility ("default"))) MPI_Datatype MPIU_ENUM PetscAttrMPITypeTag(PetscEnum);
205PETSC_EXTERNextern __attribute__((visibility ("default"))) MPI_Datatype MPIU_BOOL PetscAttrMPITypeTag(PetscBool);
206
207/*MC
208 MPIU_INT - MPI datatype corresponding to PetscInt
209
210 Notes:
211 In MPI calls that require an MPI datatype that matches a PetscInt or array of PetscInt values, pass this value.
212
213 Level: beginner
214
215.seealso: PetscReal, PetscScalar, PetscComplex, PetscInt, MPIU_REAL, MPIU_SCALAR, MPIU_COMPLEX
216M*/
217
218#if defined(PETSC_HAVE_STDINT_H1) && defined(PETSC_HAVE_INTTYPES_H1) && defined(PETSC_HAVE_MPI_INT64_T1) /* MPI_INT64_T is not guaranteed to be a macro */
219# define MPIU_INT64((MPI_Datatype)0x4c00083a) MPI_INT64_T((MPI_Datatype)0x4c00083a)
220# define PetscInt64_FMT"l" "d" PRId64"l" "d"
221#elif (PETSC_SIZEOF_LONG_LONG8 == 8)
222# define MPIU_INT64((MPI_Datatype)0x4c00083a) MPI_LONG_LONG_INT((MPI_Datatype)0x4c000809)
223# define PetscInt64_FMT"l" "d" "lld"
224#elif defined(PETSC_HAVE___INT64)
225# define MPIU_INT64((MPI_Datatype)0x4c00083a) MPI_INT64_T((MPI_Datatype)0x4c00083a)
226# define PetscInt64_FMT"l" "d" "ld"
227#else
228# error "cannot determine PetscInt64 type"
229#endif
230
231PETSC_EXTERNextern __attribute__((visibility ("default"))) MPI_Datatype MPIU_FORTRANADDR;
232
233#if defined(PETSC_USE_64BIT_INDICES)
234# define MPIU_INT((MPI_Datatype)0x4c000405) MPIU_INT64((MPI_Datatype)0x4c00083a)
235# define PetscInt_FMT"d" PetscInt64_FMT"l" "d"
236#else
237# define MPIU_INT((MPI_Datatype)0x4c000405) MPI_INT((MPI_Datatype)0x4c000405)
238# define PetscInt_FMT"d" "d"
239#endif
240
241/*
242 For the rare cases when one needs to send a size_t object with MPI
243*/
244PETSC_EXTERNextern __attribute__((visibility ("default"))) MPI_Datatype MPIU_SIZE_T;
245
246/*
247 You can use PETSC_STDOUT as a replacement of stdout. You can also change
248 the value of PETSC_STDOUT to redirect all standard output elsewhere
249*/
250PETSC_EXTERNextern __attribute__((visibility ("default"))) FILE* PETSC_STDOUT;
251
252/*
253 You can use PETSC_STDERR as a replacement of stderr. You can also change
254 the value of PETSC_STDERR to redirect all standard error elsewhere
255*/
256PETSC_EXTERNextern __attribute__((visibility ("default"))) FILE* PETSC_STDERR;
257
258/*MC
259 PetscUnlikely - hints the compiler that the given condition is usually FALSE
260
261 Synopsis:
262 #include <petscsys.h>
263 PetscBool PetscUnlikely(PetscBool cond)
264
265 Not Collective
266
267 Input Parameters:
268. cond - condition or expression
269
270 Notes:
271 This returns the same truth value, it is only a hint to compilers that the resulting
272 branch is unlikely.
273
274 Level: advanced
275
276.seealso: PetscLikely(), CHKERRQ
277M*/
278
279/*MC
280 PetscLikely - hints the compiler that the given condition is usually TRUE
281
282 Synopsis:
283 #include <petscsys.h>
284 PetscBool PetscLikely(PetscBool cond)
285
286 Not Collective
287
288 Input Parameters:
289. cond - condition or expression
290
291 Notes:
292 This returns the same truth value, it is only a hint to compilers that the resulting
293 branch is likely.
294
295 Level: advanced
296
297.seealso: PetscUnlikely()
298M*/
299#if defined(PETSC_HAVE_BUILTIN_EXPECT1)
300# define PetscUnlikely(cond)__builtin_expect(!!(cond),0) __builtin_expect(!!(cond),0)
301# define PetscLikely(cond)__builtin_expect(!!(cond),1) __builtin_expect(!!(cond),1)
302#else
303# define PetscUnlikely(cond)__builtin_expect(!!(cond),0) (cond)
304# define PetscLikely(cond)__builtin_expect(!!(cond),1) (cond)
305#endif
306
307/*
308 Declare extern C stuff after including external header files
309*/
310
311PETSC_EXTERNextern __attribute__((visibility ("default"))) const char *const PetscBools[];
312
313/*
314 Defines elementary mathematics functions and constants.
315*/
316#include <petscmath.h>
317
318PETSC_EXTERNextern __attribute__((visibility ("default"))) const char *const PetscCopyModes[];
319
320/*MC
321 PETSC_IGNORE - same as NULL, means PETSc will ignore this argument
322
323 Level: beginner
324
325 Note:
326 Accepted by many PETSc functions to not set a parameter and instead use
327 some default
328
329 Fortran Notes:
330 This macro does not exist in Fortran; you must use PETSC_NULL_INTEGER,
331 PETSC_NULL_DOUBLE_PRECISION etc
332
333.seealso: PETSC_DECIDE, PETSC_DEFAULT, PETSC_DETERMINE
334
335M*/
336#define PETSC_IGNORE((void*)0) NULL((void*)0)
337
338/* This is deprecated */
339#define PETSC_NULL((void*)0) NULL((void*)0)
340
341/*MC
342 PETSC_DECIDE - standard way of passing in integer or floating point parameter
343 where you wish PETSc to use the default.
344
345 Level: beginner
346
347.seealso: PETSC_DEFAULT, PETSC_IGNORE, PETSC_DETERMINE
348
349M*/
350#define PETSC_DECIDE-1 -1
351
352/*MC
353 PETSC_DETERMINE - standard way of passing in integer or floating point parameter
354 where you wish PETSc to compute the required value.
355
356 Level: beginner
357
358 Developer Note:
359 I would like to use const PetscInt PETSC_DETERMINE = PETSC_DECIDE; but for
360 some reason this is not allowed by the standard even though PETSC_DECIDE is a constant value.
361
362.seealso: PETSC_DECIDE, PETSC_DEFAULT, PETSC_IGNORE, VecSetSizes()
363
364M*/
365#define PETSC_DETERMINE-1 PETSC_DECIDE-1
366
367/*MC
368 PETSC_DEFAULT - standard way of passing in integer or floating point parameter
369 where you wish PETSc to use the default.
370
371 Level: beginner
372
373 Fortran Notes:
374 You need to use PETSC_DEFAULT_INTEGER or PETSC_DEFAULT_REAL.
375
376.seealso: PETSC_DECIDE, PETSC_IGNORE, PETSC_DETERMINE
377
378M*/
379#define PETSC_DEFAULT-2 -2
380
381/*MC
382 PETSC_COMM_WORLD - the equivalent of the MPI_COMM_WORLD communicator which represents
383 all the processes that PETSc knows about.
384
385 Level: beginner
386
387 Notes:
388 By default PETSC_COMM_WORLD and MPI_COMM_WORLD are identical unless you wish to
389 run PETSc on ONLY a subset of MPI_COMM_WORLD. In that case create your new (smaller)
390 communicator, call it, say comm, and set PETSC_COMM_WORLD = comm BEFORE calling
391 PetscInitialize(), but after MPI_Init() has been called.
392
393 The value of PETSC_COMM_WORLD should never be USED/accessed before PetscInitialize()
394 is called because it may not have a valid value yet.
395
396.seealso: PETSC_COMM_SELF
397
398M*/
399PETSC_EXTERNextern __attribute__((visibility ("default"))) MPI_Comm PETSC_COMM_WORLD;
400
401/*MC
402 PETSC_COMM_SELF - This is always MPI_COMM_SELF
403
404 Level: beginner
405
406 Notes:
407 Do not USE/access or set this variable before PetscInitialize() has been called.
408
409.seealso: PETSC_COMM_WORLD
410
411M*/
412#define PETSC_COMM_SELF((MPI_Comm)0x44000001) MPI_COMM_SELF((MPI_Comm)0x44000001)
413
414PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscBool PetscBeganMPI;
415PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscBool PetscInitializeCalled;
416PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscBool PetscFinalizeCalled;
417PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscBool PetscViennaCLSynchronize;
418PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscBool PetscCUDASynchronize;
419
420PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSetHelpVersionFunctions(PetscErrorCode (*)(MPI_Comm),PetscErrorCode (*)(MPI_Comm));
421PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscCommDuplicate(MPI_Comm,MPI_Comm*,int*);
422PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscCommDestroy(MPI_Comm*);
423
424#if defined(PETSC_HAVE_CUDA)
425PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscCUDAInitialize(MPI_Comm);
426#endif
427
428/*MC
429 PetscMalloc - Allocates memory, One should use PetscNew(), PetscMalloc1() or PetscCalloc1() usually instead of this
430
431 Synopsis:
432 #include <petscsys.h>
433 PetscErrorCode PetscMalloc(size_t m,void **result)
434
435 Not Collective
436
437 Input Parameter:
438. m - number of bytes to allocate
439
440 Output Parameter:
441. result - memory allocated
442
443 Level: beginner
444
445 Notes:
446 Memory is always allocated at least double aligned
447
448 It is safe to allocate size 0 and pass the resulting pointer (which may or may not be NULL) to PetscFree().
449
450.seealso: PetscFree(), PetscNew()
451
452M*/
453#define PetscMalloc(a,b)((*PetscTrMalloc)((a),PETSC_FALSE,453,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,(void**)(b)))
((*PetscTrMalloc)((a),PETSC_FALSE,__LINE__453,PETSC_FUNCTION_NAME__func__,__FILE__"/sandbox/petsc/petsc.next/include/petscsys.h",(void**)(b)))
454
455/*MC
456 PetscRealloc - Rellocates memory
457
458 Synopsis:
459 #include <petscsys.h>
460 PetscErrorCode PetscRealloc(size_t m,void **result)
461
462 Not Collective
463
464 Input Parameters:
465+ m - number of bytes to allocate
466- result - initial memory allocated
467
468 Output Parameter:
469. result - new memory allocated
470
471 Level: developer
472
473 Notes:
474 Memory is always allocated at least double aligned
475
476.seealso: PetscMalloc(), PetscFree(), PetscNew()
477
478M*/
479#define PetscRealloc(a,b)((*PetscTrRealloc)((a),479,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,(void**)(b)))
((*PetscTrRealloc)((a),__LINE__479,PETSC_FUNCTION_NAME__func__,__FILE__"/sandbox/petsc/petsc.next/include/petscsys.h",(void**)(b)))
480
481/*MC
482 PetscAddrAlign - Rounds up an address to PETSC_MEMALIGN alignment
483
484 Synopsis:
485 #include <petscsys.h>
486 void *PetscAddrAlign(void *addr)
487
488 Not Collective
489
490 Input Parameters:
491. addr - address to align (any pointer type)
492
493 Level: developer
494
495.seealso: PetscMallocAlign()
496
497M*/
498#define PetscAddrAlign(a)(void*)((((uintptr_t)(a))+(16 -1)) & ~(16 -1)) (void*)((((PETSC_UINTPTR_Tuintptr_t)(a))+(PETSC_MEMALIGN16-1)) & ~(PETSC_MEMALIGN16-1))
499
500/*MC
501 PetscMalloc1 - Allocates an array of memory aligned to PETSC_MEMALIGN
502
503 Synopsis:
504 #include <petscsys.h>
505 PetscErrorCode PetscMalloc1(size_t m1,type **r1)
506
507 Not Collective
508
509 Input Parameter:
510. m1 - number of elements to allocate (may be zero)
511
512 Output Parameter:
513. r1 - memory allocated in first chunk
514
515 Note:
516 This uses the sizeof() of the memory type requested to determine the total memory to be allocated, therefore you should not
517 multiply the number of elements requested by the sizeof() the type. For example use
518$ PetscInt *id;
519$ PetscMalloc1(10,&id);
520 not
521$ PetscInt *id;
522$ PetscMalloc1(10*sizeof(PetscInt),&id);
523
524 Does not zero the memory allocatd, used PetscCalloc1() to obtain memory that has been zeroed.
525
526 Level: beginner
527
528.seealso: PetscFree(), PetscNew(), PetscMalloc(), PetscCalloc1(), PetscMalloc2()
529
530M*/
531#define PetscMalloc1(m1,r1)PetscMallocA(1,PETSC_FALSE,531,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,(size_t)(m1)*sizeof(**(r1)),(r1))
PetscMallocA(1,PETSC_FALSE,__LINE__531,PETSC_FUNCTION_NAME__func__,__FILE__"/sandbox/petsc/petsc.next/include/petscsys.h",(size_t)(m1)*sizeof(**(r1)),(r1))
532
533/*MC
534 PetscCalloc1 - Allocates a cleared (zeroed) array of memory aligned to PETSC_MEMALIGN
535
536 Synopsis:
537 #include <petscsys.h>
538 PetscErrorCode PetscCalloc1(size_t m1,type **r1)
539
540 Not Collective
541
542 Input Parameter:
543. m1 - number of elements to allocate in 1st chunk (may be zero)
544
545 Output Parameter:
546. r1 - memory allocated in first chunk
547
548 Notes:
549 See PetsMalloc1() for more details on usage.
550
551 Level: beginner
552
553.seealso: PetscFree(), PetscNew(), PetscMalloc(), PetscMalloc1(), PetscCalloc2()
554
555M*/
556#define PetscCalloc1(m1,r1)PetscMallocA(1,PETSC_TRUE,556,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,(size_t)(m1)*sizeof(**(r1)),(r1))
PetscMallocA(1,PETSC_TRUE,__LINE__556,PETSC_FUNCTION_NAME__func__,__FILE__"/sandbox/petsc/petsc.next/include/petscsys.h",(size_t)(m1)*sizeof(**(r1)),(r1))
557
558/*MC
559 PetscMalloc2 - Allocates 2 arrays of memory both aligned to PETSC_MEMALIGN
560
561 Synopsis:
562 #include <petscsys.h>
563 PetscErrorCode PetscMalloc2(size_t m1,type **r1,size_t m2,type **r2)
564
565 Not Collective
566
567 Input Parameter:
568+ m1 - number of elements to allocate in 1st chunk (may be zero)
569- m2 - number of elements to allocate in 2nd chunk (may be zero)
570
571 Output Parameter:
572+ r1 - memory allocated in first chunk
573- r2 - memory allocated in second chunk
574
575 Level: developer
576
577.seealso: PetscFree(), PetscNew(), PetscMalloc(), PetscMalloc1(), PetscCalloc2()
578
579M*/
580#define PetscMalloc2(m1,r1,m2,r2)PetscMallocA(2,PETSC_FALSE,580,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,(size_t)(m1)*sizeof(**(r1)),(r1),(size_t)(m2)*sizeof(**(r2))
,(r2))
PetscMallocA(2,PETSC_FALSE,__LINE__580,PETSC_FUNCTION_NAME__func__,__FILE__"/sandbox/petsc/petsc.next/include/petscsys.h",(size_t)(m1)*sizeof(**(r1)),(r1),(size_t)(m2)*sizeof(**(r2)),(r2))
581
582/*MC
583 PetscCalloc2 - Allocates 2 cleared (zeroed) arrays of memory both aligned to PETSC_MEMALIGN
584
585 Synopsis:
586 #include <petscsys.h>
587 PetscErrorCode PetscCalloc2(size_t m1,type **r1,size_t m2,type **r2)
588
589 Not Collective
590
591 Input Parameter:
592+ m1 - number of elements to allocate in 1st chunk (may be zero)
593- m2 - number of elements to allocate in 2nd chunk (may be zero)
594
595 Output Parameter:
596+ r1 - memory allocated in first chunk
597- r2 - memory allocated in second chunk
598
599 Level: developer
600
601.seealso: PetscFree(), PetscNew(), PetscMalloc(), PetscCalloc1(), PetscMalloc2()
602
603M*/
604#define PetscCalloc2(m1,r1,m2,r2)PetscMallocA(2,PETSC_TRUE,604,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,(size_t)(m1)*sizeof(**(r1)),(r1),(size_t)(m2)*sizeof(**(r2))
,(r2))
PetscMallocA(2,PETSC_TRUE,__LINE__604,PETSC_FUNCTION_NAME__func__,__FILE__"/sandbox/petsc/petsc.next/include/petscsys.h",(size_t)(m1)*sizeof(**(r1)),(r1),(size_t)(m2)*sizeof(**(r2)),(r2))
605
606/*MC
607 PetscMalloc3 - Allocates 3 arrays of memory, all aligned to PETSC_MEMALIGN
608
609 Synopsis:
610 #include <petscsys.h>
611 PetscErrorCode PetscMalloc3(size_t m1,type **r1,size_t m2,type **r2,size_t m3,type **r3)
612
613 Not Collective
614
615 Input Parameter:
616+ m1 - number of elements to allocate in 1st chunk (may be zero)
617. m2 - number of elements to allocate in 2nd chunk (may be zero)
618- m3 - number of elements to allocate in 3rd chunk (may be zero)
619
620 Output Parameter:
621+ r1 - memory allocated in first chunk
622. r2 - memory allocated in second chunk
623- r3 - memory allocated in third chunk
624
625 Level: developer
626
627.seealso: PetscFree(), PetscNew(), PetscMalloc(), PetscMalloc2(), PetscCalloc3(), PetscFree3()
628
629M*/
630#define PetscMalloc3(m1,r1,m2,r2,m3,r3)PetscMallocA(3,PETSC_FALSE,630,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,(size_t)(m1)*sizeof(**(r1)),(r1),(size_t)(m2)*sizeof(**(r2))
,(r2),(size_t)(m3)*sizeof(**(r3)),(r3))
PetscMallocA(3,PETSC_FALSE,__LINE__630,PETSC_FUNCTION_NAME__func__,__FILE__"/sandbox/petsc/petsc.next/include/petscsys.h",(size_t)(m1)*sizeof(**(r1)),(r1),(size_t)(m2)*sizeof(**(r2)),(r2),(size_t)(m3)*sizeof(**(r3)),(r3))
631
632/*MC
633 PetscCalloc3 - Allocates 3 cleared (zeroed) arrays of memory, all aligned to PETSC_MEMALIGN
634
635 Synopsis:
636 #include <petscsys.h>
637 PetscErrorCode PetscCalloc3(size_t m1,type **r1,size_t m2,type **r2,size_t m3,type **r3)
638
639 Not Collective
640
641 Input Parameter:
642+ m1 - number of elements to allocate in 1st chunk (may be zero)
643. m2 - number of elements to allocate in 2nd chunk (may be zero)
644- m3 - number of elements to allocate in 3rd chunk (may be zero)
645
646 Output Parameter:
647+ r1 - memory allocated in first chunk
648. r2 - memory allocated in second chunk
649- r3 - memory allocated in third chunk
650
651 Level: developer
652
653.seealso: PetscFree(), PetscNew(), PetscMalloc(), PetscCalloc2(), PetscMalloc3(), PetscFree3()
654
655M*/
656#define PetscCalloc3(m1,r1,m2,r2,m3,r3)PetscMallocA(3,PETSC_TRUE,656,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,(size_t)(m1)*sizeof(**(r1)),(r1),(size_t)(m2)*sizeof(**(r2))
,(r2),(size_t)(m3)*sizeof(**(r3)),(r3))
PetscMallocA(3,PETSC_TRUE,__LINE__656,PETSC_FUNCTION_NAME__func__,__FILE__"/sandbox/petsc/petsc.next/include/petscsys.h",(size_t)(m1)*sizeof(**(r1)),(r1),(size_t)(m2)*sizeof(**(r2)),(r2),(size_t)(m3)*sizeof(**(r3)),(r3))
657
658/*MC
659 PetscMalloc4 - Allocates 4 arrays of memory, all aligned to PETSC_MEMALIGN
660
661 Synopsis:
662 #include <petscsys.h>
663 PetscErrorCode PetscMalloc4(size_t m1,type **r1,size_t m2,type **r2,size_t m3,type **r3,size_t m4,type **r4)
664
665 Not Collective
666
667 Input Parameter:
668+ m1 - number of elements to allocate in 1st chunk (may be zero)
669. m2 - number of elements to allocate in 2nd chunk (may be zero)
670. m3 - number of elements to allocate in 3rd chunk (may be zero)
671- m4 - number of elements to allocate in 4th chunk (may be zero)
672
673 Output Parameter:
674+ r1 - memory allocated in first chunk
675. r2 - memory allocated in second chunk
676. r3 - memory allocated in third chunk
677- r4 - memory allocated in fourth chunk
678
679 Level: developer
680
681.seealso: PetscFree(), PetscNew(), PetscMalloc(), PetscMalloc2(), PetscCalloc4(), PetscFree4()
682
683M*/
684#define PetscMalloc4(m1,r1,m2,r2,m3,r3,m4,r4)PetscMallocA(4,PETSC_FALSE,684,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,(size_t)(m1)*sizeof(**(r1)),(r1),(size_t)(m2)*sizeof(**(r2))
,(r2),(size_t)(m3)*sizeof(**(r3)),(r3),(size_t)(m4)*sizeof(**
(r4)),(r4))
PetscMallocA(4,PETSC_FALSE,__LINE__684,PETSC_FUNCTION_NAME__func__,__FILE__"/sandbox/petsc/petsc.next/include/petscsys.h",(size_t)(m1)*sizeof(**(r1)),(r1),(size_t)(m2)*sizeof(**(r2)),(r2),(size_t)(m3)*sizeof(**(r3)),(r3),(size_t)(m4)*sizeof(**(r4)),(r4))
685
686/*MC
687 PetscCalloc4 - Allocates 4 cleared (zeroed) arrays of memory, all aligned to PETSC_MEMALIGN
688
689 Synopsis:
690 #include <petscsys.h>
691 PetscErrorCode PetscCalloc4(size_t m1,type **r1,size_t m2,type **r2,size_t m3,type **r3,size_t m4,type **r4)
692
693 Not Collective
694
695 Input Parameters:
696+ m1 - number of elements to allocate in 1st chunk (may be zero)
697. m2 - number of elements to allocate in 2nd chunk (may be zero)
698. m3 - number of elements to allocate in 3rd chunk (may be zero)
699- m4 - number of elements to allocate in 4th chunk (may be zero)
700
701 Output Parameters:
702+ r1 - memory allocated in first chunk
703. r2 - memory allocated in second chunk
704. r3 - memory allocated in third chunk
705- r4 - memory allocated in fourth chunk
706
707 Level: developer
708
709.seealso: PetscFree(), PetscNew(), PetscMalloc(), PetscMalloc2(), PetscCalloc4(), PetscFree4()
710
711M*/
712#define PetscCalloc4(m1,r1,m2,r2,m3,r3,m4,r4)PetscMallocA(4,PETSC_TRUE,712,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,(size_t)(m1)*sizeof(**(r1)),(r1),(size_t)(m2)*sizeof(**(r2))
,(r2),(size_t)(m3)*sizeof(**(r3)),(r3),(size_t)(m4)*sizeof(**
(r4)),(r4))
PetscMallocA(4,PETSC_TRUE,__LINE__712,PETSC_FUNCTION_NAME__func__,__FILE__"/sandbox/petsc/petsc.next/include/petscsys.h",(size_t)(m1)*sizeof(**(r1)),(r1),(size_t)(m2)*sizeof(**(r2)),(r2),(size_t)(m3)*sizeof(**(r3)),(r3),(size_t)(m4)*sizeof(**(r4)),(r4))
713
714/*MC
715 PetscMalloc5 - Allocates 5 arrays of memory, all aligned to PETSC_MEMALIGN
716
717 Synopsis:
718 #include <petscsys.h>
719 PetscErrorCode PetscMalloc5(size_t m1,type **r1,size_t m2,type **r2,size_t m3,type **r3,size_t m4,type **r4,size_t m5,type **r5)
720
721 Not Collective
722
723 Input Parameters:
724+ m1 - number of elements to allocate in 1st chunk (may be zero)
725. m2 - number of elements to allocate in 2nd chunk (may be zero)
726. m3 - number of elements to allocate in 3rd chunk (may be zero)
727. m4 - number of elements to allocate in 4th chunk (may be zero)
728- m5 - number of elements to allocate in 5th chunk (may be zero)
729
730 Output Parameters:
731+ r1 - memory allocated in first chunk
732. r2 - memory allocated in second chunk
733. r3 - memory allocated in third chunk
734. r4 - memory allocated in fourth chunk
735- r5 - memory allocated in fifth chunk
736
737 Level: developer
738
739.seealso: PetscFree(), PetscNew(), PetscMalloc(), PetscMalloc2(), PetscCalloc5(), PetscFree5()
740
741M*/
742#define PetscMalloc5(m1,r1,m2,r2,m3,r3,m4,r4,m5,r5)PetscMallocA(5,PETSC_FALSE,742,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,(size_t)(m1)*sizeof(**(r1)),(r1),(size_t)(m2)*sizeof(**(r2))
,(r2),(size_t)(m3)*sizeof(**(r3)),(r3),(size_t)(m4)*sizeof(**
(r4)),(r4),(size_t)(m5)*sizeof(**(r5)),(r5))
PetscMallocA(5,PETSC_FALSE,__LINE__742,PETSC_FUNCTION_NAME__func__,__FILE__"/sandbox/petsc/petsc.next/include/petscsys.h",(size_t)(m1)*sizeof(**(r1)),(r1),(size_t)(m2)*sizeof(**(r2)),(r2),(size_t)(m3)*sizeof(**(r3)),(r3),(size_t)(m4)*sizeof(**(r4)),(r4),(size_t)(m5)*sizeof(**(r5)),(r5))
743
744/*MC
745 PetscCalloc5 - Allocates 5 cleared (zeroed) arrays of memory, all aligned to PETSC_MEMALIGN
746
747 Synopsis:
748 #include <petscsys.h>
749 PetscErrorCode PetscCalloc5(size_t m1,type **r1,size_t m2,type **r2,size_t m3,type **r3,size_t m4,type **r4,size_t m5,type **r5)
750
751 Not Collective
752
753 Input Parameters:
754+ m1 - number of elements to allocate in 1st chunk (may be zero)
755. m2 - number of elements to allocate in 2nd chunk (may be zero)
756. m3 - number of elements to allocate in 3rd chunk (may be zero)
757. m4 - number of elements to allocate in 4th chunk (may be zero)
758- m5 - number of elements to allocate in 5th chunk (may be zero)
759
760 Output Parameters:
761+ r1 - memory allocated in first chunk
762. r2 - memory allocated in second chunk
763. r3 - memory allocated in third chunk
764. r4 - memory allocated in fourth chunk
765- r5 - memory allocated in fifth chunk
766
767 Level: developer
768
769.seealso: PetscFree(), PetscNew(), PetscMalloc(), PetscMalloc5(), PetscFree5()
770
771M*/
772#define PetscCalloc5(m1,r1,m2,r2,m3,r3,m4,r4,m5,r5)PetscMallocA(5,PETSC_TRUE,772,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,(size_t)(m1)*sizeof(**(r1)),(r1),(size_t)(m2)*sizeof(**(r2))
,(r2),(size_t)(m3)*sizeof(**(r3)),(r3),(size_t)(m4)*sizeof(**
(r4)),(r4),(size_t)(m5)*sizeof(**(r5)),(r5))
PetscMallocA(5,PETSC_TRUE,__LINE__772,PETSC_FUNCTION_NAME__func__,__FILE__"/sandbox/petsc/petsc.next/include/petscsys.h",(size_t)(m1)*sizeof(**(r1)),(r1),(size_t)(m2)*sizeof(**(r2)),(r2),(size_t)(m3)*sizeof(**(r3)),(r3),(size_t)(m4)*sizeof(**(r4)),(r4),(size_t)(m5)*sizeof(**(r5)),(r5))
773
774/*MC
775 PetscMalloc6 - Allocates 6 arrays of memory, all aligned to PETSC_MEMALIGN
776
777 Synopsis:
778 #include <petscsys.h>
779 PetscErrorCode PetscMalloc6(size_t m1,type **r1,size_t m2,type **r2,size_t m3,type **r3,size_t m4,type **r4,size_t m5,type **r5,size_t m6,type **r6)
780
781 Not Collective
782
783 Input Parameters:
784+ m1 - number of elements to allocate in 1st chunk (may be zero)
785. m2 - number of elements to allocate in 2nd chunk (may be zero)
786. m3 - number of elements to allocate in 3rd chunk (may be zero)
787. m4 - number of elements to allocate in 4th chunk (may be zero)
788. m5 - number of elements to allocate in 5th chunk (may be zero)
789- m6 - number of elements to allocate in 6th chunk (may be zero)
790
791 Output Parameteasr:
792+ r1 - memory allocated in first chunk
793. r2 - memory allocated in second chunk
794. r3 - memory allocated in third chunk
795. r4 - memory allocated in fourth chunk
796. r5 - memory allocated in fifth chunk
797- r6 - memory allocated in sixth chunk
798
799 Level: developer
800
801.seealso: PetscFree(), PetscNew(), PetscMalloc(), PetscMalloc2(), PetscCalloc6(), PetscFree3(), PetscFree4(), PetscFree5(), PetscFree6()
802
803M*/
804#define PetscMalloc6(m1,r1,m2,r2,m3,r3,m4,r4,m5,r5,m6,r6)PetscMallocA(6,PETSC_FALSE,804,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,(size_t)(m1)*sizeof(**(r1)),(r1),(size_t)(m2)*sizeof(**(r2))
,(r2),(size_t)(m3)*sizeof(**(r3)),(r3),(size_t)(m4)*sizeof(**
(r4)),(r4),(size_t)(m5)*sizeof(**(r5)),(r5),(size_t)(m6)*sizeof
(**(r6)),(r6))
PetscMallocA(6,PETSC_FALSE,__LINE__804,PETSC_FUNCTION_NAME__func__,__FILE__"/sandbox/petsc/petsc.next/include/petscsys.h",(size_t)(m1)*sizeof(**(r1)),(r1),(size_t)(m2)*sizeof(**(r2)),(r2),(size_t)(m3)*sizeof(**(r3)),(r3),(size_t)(m4)*sizeof(**(r4)),(r4),(size_t)(m5)*sizeof(**(r5)),(r5),(size_t)(m6)*sizeof(**(r6)),(r6))
805
806/*MC
807 PetscCalloc6 - Allocates 6 cleared (zeroed) arrays of memory, all aligned to PETSC_MEMALIGN
808
809 Synopsis:
810 #include <petscsys.h>
811 PetscErrorCode PetscCalloc6(size_t m1,type **r1,size_t m2,type **r2,size_t m3,type **r3,size_t m4,type **r4,size_t m5,type **r5,size_t m6,type **r6)
812
813 Not Collective
814
815 Input Parameters:
816+ m1 - number of elements to allocate in 1st chunk (may be zero)
817. m2 - number of elements to allocate in 2nd chunk (may be zero)
818. m3 - number of elements to allocate in 3rd chunk (may be zero)
819. m4 - number of elements to allocate in 4th chunk (may be zero)
820. m5 - number of elements to allocate in 5th chunk (may be zero)
821- m6 - number of elements to allocate in 6th chunk (may be zero)
822
823 Output Parameters:
824+ r1 - memory allocated in first chunk
825. r2 - memory allocated in second chunk
826. r3 - memory allocated in third chunk
827. r4 - memory allocated in fourth chunk
828. r5 - memory allocated in fifth chunk
829- r6 - memory allocated in sixth chunk
830
831 Level: developer
832
833.seealso: PetscFree(), PetscNew(), PetscMalloc(), PetscMalloc2(), PetscMalloc6(), PetscFree6()
834
835M*/
836#define PetscCalloc6(m1,r1,m2,r2,m3,r3,m4,r4,m5,r5,m6,r6)PetscMallocA(6,PETSC_TRUE,836,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,(size_t)(m1)*sizeof(**(r1)),(r1),(size_t)(m2)*sizeof(**(r2))
,(r2),(size_t)(m3)*sizeof(**(r3)),(r3),(size_t)(m4)*sizeof(**
(r4)),(r4),(size_t)(m5)*sizeof(**(r5)),(r5),(size_t)(m6)*sizeof
(**(r6)),(r6))
PetscMallocA(6,PETSC_TRUE,__LINE__836,PETSC_FUNCTION_NAME__func__,__FILE__"/sandbox/petsc/petsc.next/include/petscsys.h",(size_t)(m1)*sizeof(**(r1)),(r1),(size_t)(m2)*sizeof(**(r2)),(r2),(size_t)(m3)*sizeof(**(r3)),(r3),(size_t)(m4)*sizeof(**(r4)),(r4),(size_t)(m5)*sizeof(**(r5)),(r5),(size_t)(m6)*sizeof(**(r6)),(r6))
837
838/*MC
839 PetscMalloc7 - Allocates 7 arrays of memory, all aligned to PETSC_MEMALIGN
840
841 Synopsis:
842 #include <petscsys.h>
843 PetscErrorCode PetscMalloc7(size_t m1,type **r1,size_t m2,type **r2,size_t m3,type **r3,size_t m4,type **r4,size_t m5,type **r5,size_t m6,type **r6,size_t m7,type **r7)
844
845 Not Collective
846
847 Input Parameters:
848+ m1 - number of elements to allocate in 1st chunk (may be zero)
849. m2 - number of elements to allocate in 2nd chunk (may be zero)
850. m3 - number of elements to allocate in 3rd chunk (may be zero)
851. m4 - number of elements to allocate in 4th chunk (may be zero)
852. m5 - number of elements to allocate in 5th chunk (may be zero)
853. m6 - number of elements to allocate in 6th chunk (may be zero)
854- m7 - number of elements to allocate in 7th chunk (may be zero)
855
856 Output Parameters:
857+ r1 - memory allocated in first chunk
858. r2 - memory allocated in second chunk
859. r3 - memory allocated in third chunk
860. r4 - memory allocated in fourth chunk
861. r5 - memory allocated in fifth chunk
862. r6 - memory allocated in sixth chunk
863- r7 - memory allocated in seventh chunk
864
865 Level: developer
866
867.seealso: PetscFree(), PetscNew(), PetscMalloc(), PetscMalloc2(), PetscCalloc7(), PetscFree7()
868
869M*/
870#define PetscMalloc7(m1,r1,m2,r2,m3,r3,m4,r4,m5,r5,m6,r6,m7,r7)PetscMallocA(7,PETSC_FALSE,870,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,(size_t)(m1)*sizeof(**(r1)),(r1),(size_t)(m2)*sizeof(**(r2))
,(r2),(size_t)(m3)*sizeof(**(r3)),(r3),(size_t)(m4)*sizeof(**
(r4)),(r4),(size_t)(m5)*sizeof(**(r5)),(r5),(size_t)(m6)*sizeof
(**(r6)),(r6),(size_t)(m7)*sizeof(**(r7)),(r7))
PetscMallocA(7,PETSC_FALSE,__LINE__870,PETSC_FUNCTION_NAME__func__,__FILE__"/sandbox/petsc/petsc.next/include/petscsys.h",(size_t)(m1)*sizeof(**(r1)),(r1),(size_t)(m2)*sizeof(**(r2)),(r2),(size_t)(m3)*sizeof(**(r3)),(r3),(size_t)(m4)*sizeof(**(r4)),(r4),(size_t)(m5)*sizeof(**(r5)),(r5),(size_t)(m6)*sizeof(**(r6)),(r6),(size_t)(m7)*sizeof(**(r7)),(r7))
871
872/*MC
873 PetscCalloc7 - Allocates 7 cleared (zeroed) arrays of memory, all aligned to PETSC_MEMALIGN
874
875 Synopsis:
876 #include <petscsys.h>
877 PetscErrorCode PetscCalloc7(size_t m1,type **r1,size_t m2,type **r2,size_t m3,type **r3,size_t m4,type **r4,size_t m5,type **r5,size_t m6,type **r6,size_t m7,type **r7)
878
879 Not Collective
880
881 Input Parameters:
882+ m1 - number of elements to allocate in 1st chunk (may be zero)
883. m2 - number of elements to allocate in 2nd chunk (may be zero)
884. m3 - number of elements to allocate in 3rd chunk (may be zero)
885. m4 - number of elements to allocate in 4th chunk (may be zero)
886. m5 - number of elements to allocate in 5th chunk (may be zero)
887. m6 - number of elements to allocate in 6th chunk (may be zero)
888- m7 - number of elements to allocate in 7th chunk (may be zero)
889
890 Output Parameters:
891+ r1 - memory allocated in first chunk
892. r2 - memory allocated in second chunk
893. r3 - memory allocated in third chunk
894. r4 - memory allocated in fourth chunk
895. r5 - memory allocated in fifth chunk
896. r6 - memory allocated in sixth chunk
897- r7 - memory allocated in seventh chunk
898
899 Level: developer
900
901.seealso: PetscFree(), PetscNew(), PetscMalloc(), PetscMalloc2(), PetscMalloc7(), PetscFree7()
902
903M*/
904#define PetscCalloc7(m1,r1,m2,r2,m3,r3,m4,r4,m5,r5,m6,r6,m7,r7)PetscMallocA(7,PETSC_TRUE,904,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,(size_t)(m1)*sizeof(**(r1)),(r1),(size_t)(m2)*sizeof(**(r2))
,(r2),(size_t)(m3)*sizeof(**(r3)),(r3),(size_t)(m4)*sizeof(**
(r4)),(r4),(size_t)(m5)*sizeof(**(r5)),(r5),(size_t)(m6)*sizeof
(**(r6)),(r6),(size_t)(m7)*sizeof(**(r7)),(r7))
PetscMallocA(7,PETSC_TRUE,__LINE__904,PETSC_FUNCTION_NAME__func__,__FILE__"/sandbox/petsc/petsc.next/include/petscsys.h",(size_t)(m1)*sizeof(**(r1)),(r1),(size_t)(m2)*sizeof(**(r2)),(r2),(size_t)(m3)*sizeof(**(r3)),(r3),(size_t)(m4)*sizeof(**(r4)),(r4),(size_t)(m5)*sizeof(**(r5)),(r5),(size_t)(m6)*sizeof(**(r6)),(r6),(size_t)(m7)*sizeof(**(r7)),(r7))
905
906/*MC
907 PetscNew - Allocates memory of a particular type, zeros the memory! Aligned to PETSC_MEMALIGN
908
909 Synopsis:
910 #include <petscsys.h>
911 PetscErrorCode PetscNew(type **result)
912
913 Not Collective
914
915 Output Parameter:
916. result - memory allocated, sized to match pointer type
917
918 Level: beginner
919
920.seealso: PetscFree(), PetscMalloc(), PetscNewLog(), PetscCalloc1(), PetscMalloc1()
921
922M*/
923#define PetscNew(b)PetscMallocA(1,PETSC_TRUE,923,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,(size_t)(1)*sizeof(**((b))),((b)))
PetscCalloc1(1,(b))PetscMallocA(1,PETSC_TRUE,923,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,(size_t)(1)*sizeof(**((b))),((b)))
924
925/*MC
926 PetscNewLog - Allocates memory of a type matching pointer, zeros the memory! Aligned to PETSC_MEMALIGN. Associates the memory allocated
927 with the given object using PetscLogObjectMemory().
928
929 Synopsis:
930 #include <petscsys.h>
931 PetscErrorCode PetscNewLog(PetscObject obj,type **result)
932
933 Not Collective
934
935 Input Parameter:
936. obj - object memory is logged to
937
938 Output Parameter:
939. result - memory allocated, sized to match pointer type
940
941 Level: developer
942
943.seealso: PetscFree(), PetscMalloc(), PetscNew(), PetscLogObjectMemory(), PetscCalloc1(), PetscMalloc1()
944
945M*/
946#define PetscNewLog(o,b)(PetscMallocA(1,PETSC_TRUE,946,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,(size_t)(1)*sizeof(**(((b)))),(((b)))) || PetscLogObjectMemory
((PetscObject)o,sizeof(**(b))))
(PetscNew((b))PetscMallocA(1,PETSC_TRUE,946,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,(size_t)(1)*sizeof(**(((b)))),(((b))))
|| PetscLogObjectMemory((PetscObject)o,sizeof(**(b))))
947
948/*MC
949 PetscFree - Frees memory
950
951 Synopsis:
952 #include <petscsys.h>
953 PetscErrorCode PetscFree(void *memory)
954
955 Not Collective
956
957 Input Parameter:
958. memory - memory to free (the pointer is ALWAYS set to NULL upon sucess)
959
960 Level: beginner
961
962 Notes:
963 Do not free memory obtained with PetscMalloc2(), PetscCalloc2() etc, they must be freed with PetscFree2() etc.
964
965 It is safe to call PetscFree() on a NULL pointer.
966
967.seealso: PetscNew(), PetscMalloc(), PetscNewLog(), PetscMalloc1(), PetscCalloc1()
968
969M*/
970#define PetscFree(a)((*PetscTrFree)((void*)(a),970,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
) || ((a) = 0,0))
((*PetscTrFree)((void*)(a),__LINE__970,PETSC_FUNCTION_NAME__func__,__FILE__"/sandbox/petsc/petsc.next/include/petscsys.h") || ((a) = 0,0))
971
972/*MC
973 PetscFree2 - Frees 2 chunks of memory obtained with PetscMalloc2()
974
975 Synopsis:
976 #include <petscsys.h>
977 PetscErrorCode PetscFree2(void *memory1,void *memory2)
978
979 Not Collective
980
981 Input Parameters:
982+ memory1 - memory to free
983- memory2 - 2nd memory to free
984
985 Level: developer
986
987 Notes:
988 Memory must have been obtained with PetscMalloc2()
989
990.seealso: PetscNew(), PetscMalloc(), PetscMalloc2(), PetscFree()
991
992M*/
993#define PetscFree2(m1,m2)PetscFreeA(2,993,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,&(m1),&(m2))
PetscFreeA(2,__LINE__993,PETSC_FUNCTION_NAME__func__,__FILE__"/sandbox/petsc/petsc.next/include/petscsys.h",&(m1),&(m2))
994
995/*MC
996 PetscFree3 - Frees 3 chunks of memory obtained with PetscMalloc3()
997
998 Synopsis:
999 #include <petscsys.h>
1000 PetscErrorCode PetscFree3(void *memory1,void *memory2,void *memory3)
1001
1002 Not Collective
1003
1004 Input Parameters:
1005+ memory1 - memory to free
1006. memory2 - 2nd memory to free
1007- memory3 - 3rd memory to free
1008
1009 Level: developer
1010
1011 Notes:
1012 Memory must have been obtained with PetscMalloc3()
1013
1014.seealso: PetscNew(), PetscMalloc(), PetscMalloc2(), PetscFree(), PetscMalloc3()
1015
1016M*/
1017#define PetscFree3(m1,m2,m3)PetscFreeA(3,1017,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,&(m1),&(m2),&(m3))
PetscFreeA(3,__LINE__1017,PETSC_FUNCTION_NAME__func__,__FILE__"/sandbox/petsc/petsc.next/include/petscsys.h",&(m1),&(m2),&(m3))
1018
1019/*MC
1020 PetscFree4 - Frees 4 chunks of memory obtained with PetscMalloc4()
1021
1022 Synopsis:
1023 #include <petscsys.h>
1024 PetscErrorCode PetscFree4(void *m1,void *m2,void *m3,void *m4)
1025
1026 Not Collective
1027
1028 Input Parameters:
1029+ m1 - memory to free
1030. m2 - 2nd memory to free
1031. m3 - 3rd memory to free
1032- m4 - 4th memory to free
1033
1034 Level: developer
1035
1036 Notes:
1037 Memory must have been obtained with PetscMalloc4()
1038
1039.seealso: PetscNew(), PetscMalloc(), PetscMalloc2(), PetscFree(), PetscMalloc3(), PetscMalloc4()
1040
1041M*/
1042#define PetscFree4(m1,m2,m3,m4)PetscFreeA(4,1042,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,&(m1),&(m2),&(m3),&(m4))
PetscFreeA(4,__LINE__1042,PETSC_FUNCTION_NAME__func__,__FILE__"/sandbox/petsc/petsc.next/include/petscsys.h",&(m1),&(m2),&(m3),&(m4))
1043
1044/*MC
1045 PetscFree5 - Frees 5 chunks of memory obtained with PetscMalloc5()
1046
1047 Synopsis:
1048 #include <petscsys.h>
1049 PetscErrorCode PetscFree5(void *m1,void *m2,void *m3,void *m4,void *m5)
1050
1051 Not Collective
1052
1053 Input Parameters:
1054+ m1 - memory to free
1055. m2 - 2nd memory to free
1056. m3 - 3rd memory to free
1057. m4 - 4th memory to free
1058- m5 - 5th memory to free
1059
1060 Level: developer
1061
1062 Notes:
1063 Memory must have been obtained with PetscMalloc5()
1064
1065.seealso: PetscNew(), PetscMalloc(), PetscMalloc2(), PetscFree(), PetscMalloc3(), PetscMalloc4(), PetscMalloc5()
1066
1067M*/
1068#define PetscFree5(m1,m2,m3,m4,m5)PetscFreeA(5,1068,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,&(m1),&(m2),&(m3),&(m4),&(m5))
PetscFreeA(5,__LINE__1068,PETSC_FUNCTION_NAME__func__,__FILE__"/sandbox/petsc/petsc.next/include/petscsys.h",&(m1),&(m2),&(m3),&(m4),&(m5))
1069
1070/*MC
1071 PetscFree6 - Frees 6 chunks of memory obtained with PetscMalloc6()
1072
1073 Synopsis:
1074 #include <petscsys.h>
1075 PetscErrorCode PetscFree6(void *m1,void *m2,void *m3,void *m4,void *m5,void *m6)
1076
1077 Not Collective
1078
1079 Input Parameters:
1080+ m1 - memory to free
1081. m2 - 2nd memory to free
1082. m3 - 3rd memory to free
1083. m4 - 4th memory to free
1084. m5 - 5th memory to free
1085- m6 - 6th memory to free
1086
1087
1088 Level: developer
1089
1090 Notes:
1091 Memory must have been obtained with PetscMalloc6()
1092
1093.seealso: PetscNew(), PetscMalloc(), PetscMalloc2(), PetscFree(), PetscMalloc3(), PetscMalloc4(), PetscMalloc5(), PetscMalloc6()
1094
1095M*/
1096#define PetscFree6(m1,m2,m3,m4,m5,m6)PetscFreeA(6,1096,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,&(m1),&(m2),&(m3),&(m4),&(m5),&(m6))
PetscFreeA(6,__LINE__1096,PETSC_FUNCTION_NAME__func__,__FILE__"/sandbox/petsc/petsc.next/include/petscsys.h",&(m1),&(m2),&(m3),&(m4),&(m5),&(m6))
1097
1098/*MC
1099 PetscFree7 - Frees 7 chunks of memory obtained with PetscMalloc7()
1100
1101 Synopsis:
1102 #include <petscsys.h>
1103 PetscErrorCode PetscFree7(void *m1,void *m2,void *m3,void *m4,void *m5,void *m6,void *m7)
1104
1105 Not Collective
1106
1107 Input Parameters:
1108+ m1 - memory to free
1109. m2 - 2nd memory to free
1110. m3 - 3rd memory to free
1111. m4 - 4th memory to free
1112. m5 - 5th memory to free
1113. m6 - 6th memory to free
1114- m7 - 7th memory to free
1115
1116
1117 Level: developer
1118
1119 Notes:
1120 Memory must have been obtained with PetscMalloc7()
1121
1122.seealso: PetscNew(), PetscMalloc(), PetscMalloc2(), PetscFree(), PetscMalloc3(), PetscMalloc4(), PetscMalloc5(), PetscMalloc6(),
1123 PetscMalloc7()
1124
1125M*/
1126#define PetscFree7(m1,m2,m3,m4,m5,m6,m7)PetscFreeA(7,1126,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,&(m1),&(m2),&(m3),&(m4),&(m5),&(m6),
&(m7))
PetscFreeA(7,__LINE__1126,PETSC_FUNCTION_NAME__func__,__FILE__"/sandbox/petsc/petsc.next/include/petscsys.h",&(m1),&(m2),&(m3),&(m4),&(m5),&(m6),&(m7))
1127
1128PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscMallocA(int,PetscBool,int,const char *,const char *,size_t,void *,...);
1129PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscFreeA(int,int,const char *,const char *,void *,...);
1130PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode (*PetscTrMalloc)(size_t,PetscBool,int,const char[],const char[],void**);
1131PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode (*PetscTrFree)(void*,int,const char[],const char[]);
1132PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode (*PetscTrRealloc)(size_t,int,const char[],const char[],void**);
1133PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscMallocSetCoalesce(PetscBool);
1134PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscMallocSet(PetscErrorCode (*)(size_t,PetscBool,int,const char[],const char[],void**),PetscErrorCode (*)(void*,int,const char[],const char[]));
1135PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscMallocClear(void);
1136
1137/*
1138 Unlike PetscMallocSet and PetscMallocClear which overwrite the existing settings, these two functions save the previous choice of allocator, and should be used in pair.
1139*/
1140PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscMallocSetDRAM(void);
1141PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscMallocResetDRAM(void);
1142
1143#define MPIU_PETSCLOGDOUBLE((MPI_Datatype)0x4c00080b) MPI_DOUBLE((MPI_Datatype)0x4c00080b)
1144#define MPIU_2PETSCLOGDOUBLE((MPI_Datatype)1275072547) MPI_2DOUBLE_PRECISION((MPI_Datatype)1275072547)
1145
1146/*
1147 Routines for tracing memory corruption/bleeding with default PETSc memory allocation
1148*/
1149PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscMallocDump(FILE *);
1150PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscMallocDumpLog(FILE *);
1151PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscMallocGetCurrentUsage(PetscLogDouble *);
1152PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscMallocGetMaximumUsage(PetscLogDouble *);
1153PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscMallocPushMaximumUsage(int);
1154PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscMallocPopMaximumUsage(int,PetscLogDouble*);
1155PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscMallocDebug(PetscBool);
1156PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscMallocGetDebug(PetscBool*);
1157PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscMallocValidate(int,const char[],const char[]);
1158PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscMallocSetDumpLog(void);
1159PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscMallocSetDumpLogThreshold(PetscLogDouble);
1160PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscMallocGetDumpLog(PetscBool*);
1161
1162PETSC_EXTERNextern __attribute__((visibility ("default"))) const char *const PetscDataTypes[];
1163PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscDataTypeToMPIDataType(PetscDataType,MPI_Datatype*);
1164PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscMPIDataTypeToPetscDataType(MPI_Datatype,PetscDataType*);
1165PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscDataTypeGetSize(PetscDataType,size_t*);
1166PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscDataTypeFromString(const char*,PetscDataType*,PetscBool*);
1167
1168/*
1169 Basic memory and string operations. These are usually simple wrappers
1170 around the basic Unix system calls, but a few of them have additional
1171 functionality and/or error checking.
1172*/
1173PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscBitMemcpy(void*,PetscInt,const void*,PetscInt,PetscInt,PetscDataType);
1174PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscMemcmp(const void*,const void*,size_t,PetscBool *);
1175PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscStrlen(const char[],size_t*);
1176PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscStrToArray(const char[],char,int*,char ***);
1177PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscStrToArrayDestroy(int,char **);
1178PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscStrcmp(const char[],const char[],PetscBool *);
1179PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscStrgrt(const char[],const char[],PetscBool *);
1180PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscStrcasecmp(const char[],const char[],PetscBool *);
1181PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscStrncmp(const char[],const char[],size_t,PetscBool *);
1182PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscStrcpy(char[],const char[]);
1183PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscStrcat(char[],const char[]);
1184PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscStrlcat(char[],const char[],size_t);
1185PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscStrncpy(char[],const char[],size_t);
1186PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscStrchr(const char[],char,char *[]);
1187PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscStrtolower(char[]);
1188PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscStrtoupper(char[]);
1189PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscStrrchr(const char[],char,char *[]);
1190PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscStrstr(const char[],const char[],char *[]);
1191PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscStrrstr(const char[],const char[],char *[]);
1192PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscStrendswith(const char[],const char[],PetscBool*);
1193PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscStrbeginswith(const char[],const char[],PetscBool*);
1194PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscStrendswithwhich(const char[],const char *const*,PetscInt*);
1195PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscStrallocpy(const char[],char *[]);
1196PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscStrArrayallocpy(const char *const*,char***);
1197PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscStrArrayDestroy(char***);
1198PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscStrNArrayallocpy(PetscInt,const char *const*,char***);
1199PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscStrNArrayDestroy(PetscInt,char***);
1200PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscStrreplace(MPI_Comm,const char[],char[],size_t);
1201
1202PETSC_EXTERNextern __attribute__((visibility ("default"))) void PetscStrcmpNoError(const char[],const char[],PetscBool *);
1203
1204PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscTokenCreate(const char[],const char,PetscToken*);
1205PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscTokenFind(PetscToken,char *[]);
1206PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscTokenDestroy(PetscToken*);
1207PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscStrInList(const char[],const char[],char,PetscBool*);
1208
1209PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscEListFind(PetscInt,const char *const*,const char*,PetscInt*,PetscBool*);
1210PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscEnumFind(const char *const*,const char*,PetscEnum*,PetscBool*);
1211
1212/*
1213 These are MPI operations for MPI_Allreduce() etc
1214*/
1215PETSC_EXTERNextern __attribute__((visibility ("default"))) MPI_Op MPIU_MAXSUM_OP;
1216#if (defined(PETSC_HAVE_COMPLEX1) && !defined(PETSC_HAVE_MPI_C_DOUBLE_COMPLEX1)) || defined(PETSC_USE_REAL___FLOAT128) || defined(PETSC_USE_REAL___FP16)
1217PETSC_EXTERNextern __attribute__((visibility ("default"))) MPI_Op MPIU_SUM(MPI_Op)(0x58000003);
1218#else
1219#define MPIU_SUM(MPI_Op)(0x58000003) MPI_SUM(MPI_Op)(0x58000003)
1220#endif
1221#if defined(PETSC_USE_REAL___FLOAT128) || defined(PETSC_USE_REAL___FP16)
1222PETSC_EXTERNextern __attribute__((visibility ("default"))) MPI_Op MPIU_MAX(MPI_Op)(0x58000001);
1223PETSC_EXTERNextern __attribute__((visibility ("default"))) MPI_Op MPIU_MIN(MPI_Op)(0x58000002);
1224#else
1225#define MPIU_MAX(MPI_Op)(0x58000001) MPI_MAX(MPI_Op)(0x58000001)
1226#define MPIU_MIN(MPI_Op)(0x58000002) MPI_MIN(MPI_Op)(0x58000002)
1227#endif
1228PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscMaxSum(MPI_Comm,const PetscInt[],PetscInt*,PetscInt*);
1229
1230PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode MPIULong_Send(void*,PetscInt,MPI_Datatype,PetscMPIInt,PetscMPIInt,MPI_Comm);
1231PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode MPIULong_Recv(void*,PetscInt,MPI_Datatype,PetscMPIInt,PetscMPIInt,MPI_Comm);
1232
1233PETSC_EXTERNextern __attribute__((visibility ("default"))) const char *const PetscFileModes[];
1234
1235/*
1236 Defines PETSc error handling.
1237*/
1238#include <petscerror.h>
1239
1240#define PETSC_SMALLEST_CLASSID1211211 1211211
1241PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscClassId PETSC_LARGEST_CLASSID;
1242PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscClassId PETSC_OBJECT_CLASSID;
1243PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscClassIdRegister(const char[],PetscClassId *);
1244
1245/*
1246 Routines that get memory usage information from the OS
1247*/
1248PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscMemoryGetCurrentUsage(PetscLogDouble *);
1249PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscMemoryGetMaximumUsage(PetscLogDouble *);
1250PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscMemorySetGetMaximumUsage(void);
1251PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscMemoryTrace(const char[]);
1252
1253PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscInfoAllow(PetscBool ,const char []);
1254PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSleep(PetscReal);
1255
1256/*
1257 Initialization of PETSc
1258*/
1259PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscInitialize(int*,char***,const char[],const char[]);
1260PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscInitializeNoPointers(int,char**,const char[],const char[]);
1261PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscInitializeNoArguments(void);
1262PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscInitialized(PetscBool *);
1263PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscFinalized(PetscBool *);
1264PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscFinalize(void);
1265PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscInitializeFortran(void);
1266PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscGetArgs(int*,char ***);
1267PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscGetArguments(char ***);
1268PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscFreeArguments(char **);
1269
1270PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscEnd(void);
1271PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSysInitializePackage(void);
1272
1273PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscPythonInitialize(const char[],const char[]);
1274PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscPythonFinalize(void);
1275PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscPythonPrintError(void);
1276PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscPythonMonitorSet(PetscObject,const char[]);
1277
1278/*
1279 These are so that in extern C code we can caste function pointers to non-extern C
1280 function pointers. Since the regular C++ code expects its function pointers to be C++
1281*/
1282PETSC_EXTERN_TYPEDEF typedef void (**PetscVoidStarFunction)(void);
1283PETSC_EXTERN_TYPEDEF typedef void (*PetscVoidFunction)(void);
1284PETSC_EXTERN_TYPEDEF typedef PetscErrorCode (*PetscErrorCodeFunction)(void);
1285
1286/*
1287 Functions that can act on any PETSc object.
1288*/
1289PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectDestroy(PetscObject*);
1290PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectGetComm(PetscObject,MPI_Comm *);
1291PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectGetClassId(PetscObject,PetscClassId *);
1292PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectGetClassName(PetscObject,const char *[]);
1293PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectSetType(PetscObject,const char []);
1294PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectGetType(PetscObject,const char *[]);
1295PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectSetName(PetscObject,const char[]);
1296PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectGetName(PetscObject,const char*[]);
1297PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectSetTabLevel(PetscObject,PetscInt);
1298PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectGetTabLevel(PetscObject,PetscInt*);
1299PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectIncrementTabLevel(PetscObject,PetscObject,PetscInt);
1300PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectReference(PetscObject);
1301PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectGetReference(PetscObject,PetscInt*);
1302PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectDereference(PetscObject);
1303PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectGetNewTag(PetscObject,PetscMPIInt *);
1304PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectCompose(PetscObject,const char[],PetscObject);
1305PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectRemoveReference(PetscObject,const char[]);
1306PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectQuery(PetscObject,const char[],PetscObject *);
1307PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectComposeFunction_Private(PetscObject,const char[],void (*)(void));
1308#define PetscObjectComposeFunction(a,b,d)PetscObjectComposeFunction_Private(a,b,(PetscVoidFunction)(d)
)
PetscObjectComposeFunction_Private(a,b,(PetscVoidFunction)(d))
1309PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectSetFromOptions(PetscObject);
1310PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectSetUp(PetscObject);
1311PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectSetPrintedOptions(PetscObject);
1312PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectInheritPrintedOptions(PetscObject,PetscObject);
1313PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscCommGetNewTag(MPI_Comm,PetscMPIInt *);
1314
1315#include <petscviewertypes.h>
1316#include <petscoptions.h>
1317
1318PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectsListGetGlobalNumbering(MPI_Comm,PetscInt,PetscObject*,PetscInt*,PetscInt*);
1319
1320PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscMemoryShowUsage(PetscViewer,const char[]);
1321PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscMemoryView(PetscViewer,const char[]);
1322PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectPrintClassNamePrefixType(PetscObject,PetscViewer);
1323PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectView(PetscObject,PetscViewer);
1324#define PetscObjectQueryFunction(obj,name,fptr)PetscObjectQueryFunction_Private((obj),(name),(PetscVoidFunction
*)(fptr))
PetscObjectQueryFunction_Private((obj),(name),(PetscVoidFunction*)(fptr))
1325PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectQueryFunction_Private(PetscObject,const char[],void (**)(void));
1326PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectSetOptionsPrefix(PetscObject,const char[]);
1327PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectAppendOptionsPrefix(PetscObject,const char[]);
1328PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectPrependOptionsPrefix(PetscObject,const char[]);
1329PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectGetOptionsPrefix(PetscObject,const char*[]);
1330PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectChangeTypeName(PetscObject,const char[]);
1331PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectRegisterDestroy(PetscObject);
1332PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectRegisterDestroyAll(void);
1333PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectViewFromOptions(PetscObject,PetscObject,const char[]);
1334PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectName(PetscObject);
1335PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectTypeCompare(PetscObject,const char[],PetscBool *);
1336PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectBaseTypeCompare(PetscObject,const char[],PetscBool *);
1337PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectTypeCompareAny(PetscObject,PetscBool*,const char[],...);
1338PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectBaseTypeCompareAny(PetscObject,PetscBool*,const char[],...);
1339PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscRegisterFinalize(PetscErrorCode (*)(void));
1340PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscRegisterFinalizeAll(void);
1341
1342#if defined(PETSC_HAVE_SAWS)
1343PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSAWsBlock(void);
1344PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectSAWsViewOff(PetscObject)0;
1345PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectSAWsSetBlock(PetscObject,PetscBool)0;
1346PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectSAWsBlock(PetscObject)0;
1347PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectSAWsGrantAccess(PetscObject)0;
1348PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectSAWsTakeAccess(PetscObject)0;
1349PETSC_EXTERNextern __attribute__((visibility ("default"))) void PetscStackSAWsGrantAccess(void);
1350PETSC_EXTERNextern __attribute__((visibility ("default"))) void PetscStackSAWsTakeAccess(void);
1351PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscStackViewSAWs(void);
1352PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscStackSAWsViewOff(void);
1353
1354#else
1355#define PetscSAWsBlock()0 0
1356#define PetscObjectSAWsViewOff(obj)0 0
1357#define PetscObjectSAWsSetBlock(obj,flg)0 0
1358#define PetscObjectSAWsBlock(obj)0 0
1359#define PetscObjectSAWsGrantAccess(obj)0 0
1360#define PetscObjectSAWsTakeAccess(obj)0 0
1361#define PetscStackViewSAWs()0 0
1362#define PetscStackSAWsViewOff()0 0
1363#define PetscStackSAWsTakeAccess()
1364#define PetscStackSAWsGrantAccess()
1365
1366#endif
1367
1368PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscDLOpen(const char[],PetscDLMode,PetscDLHandle *);
1369PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscDLClose(PetscDLHandle *);
1370PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscDLSym(PetscDLHandle,const char[],void **);
1371
1372#if defined(PETSC_USE_DEBUG1)
1373PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscMallocGetStack(void*,PetscStack**);
1374#endif
1375PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectsDump(FILE*,PetscBool);
1376
1377PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectListDestroy(PetscObjectList*);
1378PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectListFind(PetscObjectList,const char[],PetscObject*);
1379PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectListReverseFind(PetscObjectList,PetscObject,char**,PetscBool*);
1380PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectListAdd(PetscObjectList *,const char[],PetscObject);
1381PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectListRemoveReference(PetscObjectList *,const char[]);
1382PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscObjectListDuplicate(PetscObjectList,PetscObjectList *);
1383
1384/*
1385 Dynamic library lists. Lists of names of routines in objects or in dynamic
1386 link libraries that will be loaded as needed.
1387*/
1388
1389#define PetscFunctionListAdd(list,name,fptr)PetscFunctionListAdd_Private((list),(name),(PetscVoidFunction
)(fptr))
PetscFunctionListAdd_Private((list),(name),(PetscVoidFunction)(fptr))
1390PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscFunctionListAdd_Private(PetscFunctionList*,const char[],void (*)(void));
1391PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscFunctionListDestroy(PetscFunctionList*);
1392#define PetscFunctionListFind(list,name,fptr)PetscFunctionListFind_Private((list),(name),(PetscVoidFunction
*)(fptr))
PetscFunctionListFind_Private((list),(name),(PetscVoidFunction*)(fptr))
1393PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscFunctionListFind_Private(PetscFunctionList,const char[],void (**)(void));
1394PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscFunctionListPrintTypes(MPI_Comm,FILE*,const char[],const char[],const char[],const char[],PetscFunctionList,const char[],const char[]);
1395PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscFunctionListDuplicate(PetscFunctionList,PetscFunctionList *);
1396PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscFunctionListView(PetscFunctionList,PetscViewer);
1397PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscFunctionListGet(PetscFunctionList,const char ***,int*);
1398
1399PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscDLLibrary PetscDLLibrariesLoaded;
1400PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscDLLibraryAppend(MPI_Comm,PetscDLLibrary *,const char[]);
1401PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscDLLibraryPrepend(MPI_Comm,PetscDLLibrary *,const char[]);
1402PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscDLLibrarySym(MPI_Comm,PetscDLLibrary *,const char[],const char[],void **);
1403PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscDLLibraryPrintPath(PetscDLLibrary);
1404PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscDLLibraryRetrieve(MPI_Comm,const char[],char *,size_t,PetscBool *);
1405PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscDLLibraryOpen(MPI_Comm,const char[],PetscDLLibrary *);
1406PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscDLLibraryClose(PetscDLLibrary);
1407
1408/*
1409 Useful utility routines
1410*/
1411PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSplitOwnership(MPI_Comm,PetscInt*,PetscInt*);
1412PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSplitOwnershipBlock(MPI_Comm,PetscInt,PetscInt*,PetscInt*);
1413PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSequentialPhaseBegin(MPI_Comm,PetscMPIInt);
1414PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSequentialPhaseEnd(MPI_Comm,PetscMPIInt);
1415PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscBarrier(PetscObject);
1416PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscMPIDump(FILE*);
1417PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscGlobalMinMaxInt(MPI_Comm,PetscInt[],PetscInt[]);
1418PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscGlobalMinMaxReal(MPI_Comm,PetscReal[],PetscReal[]);
1419
1420/*MC
1421 PetscNot - negates a logical type value and returns result as a PetscBool
1422
1423 Notes:
1424 This is useful in cases like
1425$ int *a;
1426$ PetscBool flag = PetscNot(a)
1427 where !a would not return a PetscBool because we cannot provide a cast from int to PetscBool in C.
1428
1429 Level: beginner
1430
1431 .seealso : PetscBool, PETSC_TRUE, PETSC_FALSE
1432M*/
1433#define PetscNot(a)((a) ? PETSC_FALSE : PETSC_TRUE) ((a) ? PETSC_FALSE : PETSC_TRUE)
1434
1435/*MC
1436 PetscHelpPrintf - Prints help messages.
1437
1438 Synopsis:
1439 #include <petscsys.h>
1440 PetscErrorCode (*PetscHelpPrintf)(const char format[],...);
1441
1442 Not Collective
1443
1444 Input Parameters:
1445. format - the usual printf() format string
1446
1447 Level: developer
1448
1449 Fortran Note:
1450 This routine is not supported in Fortran.
1451
1452
1453.seealso: PetscFPrintf(), PetscSynchronizedPrintf(), PetscErrorPrintf()
1454M*/
1455PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode (*PetscHelpPrintf)(MPI_Comm,const char[],...);
1456
1457/*
1458 Defines PETSc profiling.
1459*/
1460#include <petsclog.h>
1461
1462/*
1463 Simple PETSc parallel IO for ASCII printing
1464*/
1465PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscFixFilename(const char[],char[]);
1466PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscFOpen(MPI_Comm,const char[],const char[],FILE**);
1467PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscFClose(MPI_Comm,FILE*);
1468PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscFPrintf(MPI_Comm,FILE*,const char[],...);
1469PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscPrintf(MPI_Comm,const char[],...);
1470PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSNPrintf(char*,size_t,const char [],...);
1471PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSNPrintfCount(char*,size_t,const char [],size_t*,...);
1472PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscFormatRealArray(char[],size_t,const char*,PetscInt,const PetscReal[]);
1473
1474PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscErrorPrintfDefault(const char [],...);
1475PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscErrorPrintfNone(const char [],...);
1476PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscHelpPrintfDefault(MPI_Comm,const char [],...);
1477
1478PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscFormatConvertGetSize(const char*,size_t*);
1479PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscFormatConvert(const char*,char *);
1480
1481#if defined(PETSC_HAVE_POPEN1)
1482PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscPOpen(MPI_Comm,const char[],const char[],const char[],FILE **);
1483PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscPClose(MPI_Comm,FILE*);
1484PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscPOpenSetMachine(const char[]);
1485#endif
1486
1487PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSynchronizedPrintf(MPI_Comm,const char[],...);
1488PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSynchronizedFPrintf(MPI_Comm,FILE*,const char[],...);
1489PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSynchronizedFlush(MPI_Comm,FILE*);
1490PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSynchronizedFGets(MPI_Comm,FILE*,size_t,char[]);
1491PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscStartMatlab(MPI_Comm,const char[],const char[],FILE**);
1492PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscStartJava(MPI_Comm,const char[],const char[],FILE**);
1493PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscGetPetscDir(const char*[]);
1494
1495PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscClassId PETSC_CONTAINER_CLASSID;
1496PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscContainerGetPointer(PetscContainer,void **);
1497PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscContainerSetPointer(PetscContainer,void *);
1498PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscContainerDestroy(PetscContainer*);
1499PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscContainerCreate(MPI_Comm,PetscContainer *);
1500PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscContainerSetUserDestroy(PetscContainer, PetscErrorCode (*)(void*));
1501PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscContainerUserDestroyDefault(void*);
1502
1503/*
1504 For use in debuggers
1505*/
1506PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscMPIInt PetscGlobalRank;
1507PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscMPIInt PetscGlobalSize;
1508PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscIntView(PetscInt,const PetscInt[],PetscViewer);
1509PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscRealView(PetscInt,const PetscReal[],PetscViewer);
1510PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscScalarView(PetscInt,const PetscScalar[],PetscViewer);
1511
1512#include <stddef.h>
1513#include <string.h> /* for memcpy, memset */
1514#include <stdlib.h>
1515
1516#if defined(PETSC_HAVE_XMMINTRIN_H1) && !defined(__CUDACC__)
1517#include <xmmintrin.h>
1518#endif
1519
1520/*@C
1521 PetscMemmove - Copies n bytes, beginning at location b, to the space
1522 beginning at location a. Copying between regions that overlap will
1523 take place correctly. Use PetscMemcpy() if the locations do not overlap
1524
1525 Not Collective
1526
1527 Input Parameters:
1528+ b - pointer to initial memory space
1529. a - pointer to copy space
1530- n - length (in bytes) of space to copy
1531
1532 Level: intermediate
1533
1534 Note:
1535 PetscArraymove() is preferred
1536 This routine is analogous to memmove().
1537
1538 Developers Note: This is inlined for performance
1539
1540.seealso: PetscMemcpy(), PetscMemcmp(), PetscArrayzero(), PetscMemzero(), PetscArraycmp(), PetscArraycpy(), PetscStrallocpy(),
1541 PetscArraymove()
1542@*/
1543PETSC_STATIC_INLINEstatic inline PetscErrorCode PetscMemmove(void *a,const void *b,size_t n)
1544{
1545 PetscFunctionBegindo { do { ; if (petscstack && (petscstack->currentsize
< 64)) { petscstack->function[petscstack->currentsize
] = __func__; petscstack->file[petscstack->currentsize]
= "/sandbox/petsc/petsc.next/include/petscsys.h"; petscstack
->line[petscstack->currentsize] = 1545; petscstack->
petscroutine[petscstack->currentsize] = PETSC_TRUE; petscstack
->currentsize++; } if (petscstack) { petscstack->hotdepth
+= (PETSC_FALSE || petscstack->hotdepth); } ; } while (0)
; ; } while (0)
;
1546 if (n > 0 && !a) SETERRQ(PETSC_COMM_SELF,PETSC_ERR_ARG_NULL,"Trying to copy to null pointer")return PetscError(((MPI_Comm)0x44000001),1546,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,85,PETSC_ERROR_INITIAL,"Trying to copy to null pointer")
;
1547 if (n > 0 && !b) SETERRQ(PETSC_COMM_SELF,PETSC_ERR_ARG_NULL,"Trying to copy from a null pointer")return PetscError(((MPI_Comm)0x44000001),1547,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,85,PETSC_ERROR_INITIAL,"Trying to copy from a null pointer")
;
1548#if !defined(PETSC_HAVE_MEMMOVE1)
1549 if (a < b) {
1550 if (a <= b - n) memcpy(a,b,n);
1551 else {
1552 memcpy(a,b,(int)(b - a));
1553 PetscMemmove(b,b + (int)(b - a),n - (int)(b - a));
1554 }
1555 } else {
1556 if (b <= a - n) memcpy(a,b,n);
1557 else {
1558 memcpy(b + n,b + (n - (int)(a - b)),(int)(a - b));
1559 PetscMemmove(a,b,n - (int)(a - b));
1560 }
1561 }
1562#else
1563 memmove((char*)(a),(char*)(b),n);
1564#endif
1565 PetscFunctionReturn(0)do { do { ; if (petscstack && petscstack->currentsize
> 0) { petscstack->currentsize--; petscstack->function
[petscstack->currentsize] = 0; petscstack->file[petscstack
->currentsize] = 0; petscstack->line[petscstack->currentsize
] = 0; petscstack->petscroutine[petscstack->currentsize
] = PETSC_FALSE; } if (petscstack) { petscstack->hotdepth =
(((petscstack->hotdepth-1)<(0)) ? (0) : (petscstack->
hotdepth-1)); } ; } while (0); return(0);} while (0)
;
1566}
1567
1568/*@C
1569 PetscMemcpy - Copies n bytes, beginning at location b, to the space
1570 beginning at location a. The two memory regions CANNOT overlap, use
1571 PetscMemmove() in that case.
1572
1573 Not Collective
1574
1575 Input Parameters:
1576+ b - pointer to initial memory space
1577- n - length (in bytes) of space to copy
1578
1579 Output Parameter:
1580. a - pointer to copy space
1581
1582 Level: intermediate
1583
1584 Compile Option:
1585 PETSC_PREFER_DCOPY_FOR_MEMCPY will cause the BLAS dcopy() routine to be used
1586 for memory copies on double precision values.
1587 PETSC_PREFER_COPY_FOR_MEMCPY will cause C code to be used
1588 for memory copies on double precision values.
1589 PETSC_PREFER_FORTRAN_FORMEMCPY will cause Fortran code to be used
1590 for memory copies on double precision values.
1591
1592 Note:
1593 Prefer PetscArraycpy()
1594 This routine is analogous to memcpy().
1595 Not available from Fortran
1596
1597 Developer Note: this is inlined for fastest performance
1598
1599.seealso: PetscMemzero(), PetscMemcmp(), PetscArrayzero(), PetscArraycmp(), PetscArraycpy(), PetscMemmove(), PetscStrallocpy()
1600
1601@*/
1602PETSC_STATIC_INLINEstatic inline PetscErrorCode PetscMemcpy(void *a,const void *b,size_t n)
1603{
1604#if defined(PETSC_USE_DEBUG1)
1605 size_t al = (size_t) a,bl = (size_t) b;
1606 size_t nl = (size_t) n;
1607 PetscFunctionBegindo { do { ; if (petscstack && (petscstack->currentsize
< 64)) { petscstack->function[petscstack->currentsize
] = __func__; petscstack->file[petscstack->currentsize]
= "/sandbox/petsc/petsc.next/include/petscsys.h"; petscstack
->line[petscstack->currentsize] = 1607; petscstack->
petscroutine[petscstack->currentsize] = PETSC_TRUE; petscstack
->currentsize++; } if (petscstack) { petscstack->hotdepth
+= (PETSC_FALSE || petscstack->hotdepth); } ; } while (0)
; ; } while (0)
;
1608 if (n > 0 && !b) SETERRQ(PETSC_COMM_SELF,PETSC_ERR_ARG_NULL,"Trying to copy from a null pointer")return PetscError(((MPI_Comm)0x44000001),1608,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,85,PETSC_ERROR_INITIAL,"Trying to copy from a null pointer")
;
1609 if (n > 0 && !a) SETERRQ(PETSC_COMM_SELF,PETSC_ERR_ARG_NULL,"Trying to copy to a null pointer")return PetscError(((MPI_Comm)0x44000001),1609,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,85,PETSC_ERROR_INITIAL,"Trying to copy to a null pointer")
;
1610#else
1611 PetscFunctionBegindo { do { ; if (petscstack && (petscstack->currentsize
< 64)) { petscstack->function[petscstack->currentsize
] = __func__; petscstack->file[petscstack->currentsize]
= "/sandbox/petsc/petsc.next/include/petscsys.h"; petscstack
->line[petscstack->currentsize] = 1611; petscstack->
petscroutine[petscstack->currentsize] = PETSC_TRUE; petscstack
->currentsize++; } if (petscstack) { petscstack->hotdepth
+= (PETSC_FALSE || petscstack->hotdepth); } ; } while (0)
; ; } while (0)
;
1612#endif
1613 if (a != b && n > 0) {
1614#if defined(PETSC_USE_DEBUG1)
1615 if ((al > bl && (al - bl) < nl) || (bl - al) < nl) SETERRQ3(PETSC_COMM_SELF,PETSC_ERR_ARG_INCOMP,"Memory regions overlap: either use PetscMemmov()\n\return PetscError(((MPI_Comm)0x44000001),1617,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,75,PETSC_ERROR_INITIAL,"Memory regions overlap: either use PetscMemmov()\n or make sure your copy regions and lengths are correct. \n Length (bytes) %ld first address %ld second address %ld"
,nl,al,bl)
1616 or make sure your copy regions and lengths are correct. \n\return PetscError(((MPI_Comm)0x44000001),1617,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,75,PETSC_ERROR_INITIAL,"Memory regions overlap: either use PetscMemmov()\n or make sure your copy regions and lengths are correct. \n Length (bytes) %ld first address %ld second address %ld"
,nl,al,bl)
1617 Length (bytes) %ld first address %ld second address %ld",nl,al,bl)return PetscError(((MPI_Comm)0x44000001),1617,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,75,PETSC_ERROR_INITIAL,"Memory regions overlap: either use PetscMemmov()\n or make sure your copy regions and lengths are correct. \n Length (bytes) %ld first address %ld second address %ld"
,nl,al,bl)
;
1618#endif
1619#if (defined(PETSC_PREFER_DCOPY_FOR_MEMCPY) || defined(PETSC_PREFER_COPY_FOR_MEMCPY) || defined(PETSC_PREFER_FORTRAN_FORMEMCPY))
1620 if (!(a % sizeof(PetscScalar)) && !(n % sizeof(PetscScalar))) {
1621 size_t len = n/sizeof(PetscScalar);
1622#if defined(PETSC_PREFER_DCOPY_FOR_MEMCPY)
1623 PetscBLASInt one = 1,blen;
1624 PetscErrorCode ierr;
1625 ierr = PetscBLASIntCast(len,&blen);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),1625,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
1626 PetscStackCallBLAS("BLAScopy",BLAScopy_(&blen,(PetscScalar *)b,&one,(PetscScalar *)a,&one));
1627#elif defined(PETSC_PREFER_FORTRAN_FORMEMCPY)
1628 fortrancopy_(&len,(PetscScalar*)b,(PetscScalar*)a);
1629#else
1630 size_t i;
1631 PetscScalar *x = (PetscScalar*)b, *y = (PetscScalar*)a;
1632 for (i=0; i<len; i++) y[i] = x[i];
1633#endif
1634 } else {
1635 memcpy((char*)(a),(char*)(b),n);
1636 }
1637#else
1638 memcpy((char*)(a),(char*)(b),n);
1639#endif
1640 }
1641 PetscFunctionReturn(0)do { do { ; if (petscstack && petscstack->currentsize
> 0) { petscstack->currentsize--; petscstack->function
[petscstack->currentsize] = 0; petscstack->file[petscstack
->currentsize] = 0; petscstack->line[petscstack->currentsize
] = 0; petscstack->petscroutine[petscstack->currentsize
] = PETSC_FALSE; } if (petscstack) { petscstack->hotdepth =
(((petscstack->hotdepth-1)<(0)) ? (0) : (petscstack->
hotdepth-1)); } ; } while (0); return(0);} while (0)
;
1642}
1643
1644/*@C
1645 PetscMemzero - Zeros the specified memory.
1646
1647 Not Collective
1648
1649 Input Parameters:
1650+ a - pointer to beginning memory location
1651- n - length (in bytes) of memory to initialize
1652
1653 Level: intermediate
1654
1655 Compile Option:
1656 PETSC_PREFER_BZERO - on certain machines (the IBM RS6000) the bzero() routine happens
1657 to be faster than the memset() routine. This flag causes the bzero() routine to be used.
1658
1659 Not available from Fortran
1660 Prefer PetscArrayzero()
1661
1662 Developer Note: this is inlined for fastest performance
1663
1664.seealso: PetscMemcpy(), PetscMemcmp(), PetscArrayzero(), PetscArraycmp(), PetscArraycpy(), PetscMemmove(), PetscStrallocpy()
1665@*/
1666PETSC_STATIC_INLINEstatic inline PetscErrorCode PetscMemzero(void *a,size_t n)
1667{
1668 if (n > 0) {
1669#if defined(PETSC_USE_DEBUG1)
1670 if (!a) SETERRQ(PETSC_COMM_SELF,PETSC_ERR_ARG_NULL,"Trying to zero at a null pointer")return PetscError(((MPI_Comm)0x44000001),1670,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,85,PETSC_ERROR_INITIAL,"Trying to zero at a null pointer")
;
1671#endif
1672#if defined(PETSC_PREFER_ZERO_FOR_MEMZERO)
1673 if (!(((long) a) % sizeof(PetscScalar)) && !(n % sizeof(PetscScalar))) {
1674 size_t i,len = n/sizeof(PetscScalar);
1675 PetscScalar *x = (PetscScalar*)a;
1676 for (i=0; i<len; i++) x[i] = 0.0;
1677 } else {
1678#elif defined(PETSC_PREFER_FORTRAN_FOR_MEMZERO)
1679 if (!(((long) a) % sizeof(PetscScalar)) && !(n % sizeof(PetscScalar))) {
1680 PetscInt len = n/sizeof(PetscScalar);
1681 fortranzero_(&len,(PetscScalar*)a);
1682 } else {
1683#endif
1684#if defined(PETSC_PREFER_BZERO)
1685 bzero((char *)a,n);
1686#else
1687 memset((char*)a,0,n);
1688#endif
1689#if defined(PETSC_PREFER_ZERO_FOR_MEMZERO) || defined(PETSC_PREFER_FORTRAN_FOR_MEMZERO)
1690 }
1691#endif
1692 }
1693 return 0;
1694}
1695
1696/*MC
1697 PetscArraycmp - Compares two arrays in memory.
1698
1699 Synopsis:
1700 #include <petscsys.h>
1701 PetscErrorCode PetscArraycmp(const anytype *str1,const anytype *str2,size_t cnt,PetscBool *e)
1702
1703 Not Collective
1704
1705 Input Parameters:
1706+ str1 - First array
1707. str2 - Second array
1708- cnt - Count of the array, not in bytes, but number of entries in the arrays
1709
1710 Output Parameters:
1711. e - PETSC_TRUE if equal else PETSC_FALSE.
1712
1713 Level: intermediate
1714
1715 Note:
1716 This routine is a preferred replacement to PetscMemcmp()
1717 The arrays must be of the same type
1718
1719.seealso: PetscMemcpy(), PetscMemcmp(), PetscArrayzero(), PetscMemzero(), PetscArraycpy(), PetscMemmove(), PetscStrallocpy(),
1720 PetscArraymove()
1721M*/
1722#define PetscArraycmp(str1,str2,cnt,e)((sizeof(*(str1)) != sizeof(*(str2))) || PetscMemcmp(str1,str2
,(cnt)*sizeof(*(str1)),e))
((sizeof(*(str1)) != sizeof(*(str2))) || PetscMemcmp(str1,str2,(cnt)*sizeof(*(str1)),e))
1723
1724/*MC
1725 PetscArraymove - Copies from one array in memory to another, the arrays may overlap. Use PetscArraycpy() when the arrays
1726 do not overlap
1727
1728 Synopsis:
1729 #include <petscsys.h>
1730 PetscErrorCode PetscArraymove(anytype *str1,const anytype *str2,size_t cnt)
1731
1732 Not Collective
1733
1734 Input Parameters:
1735+ str1 - First array
1736. str2 - Second array
1737- cnt - Count of the array, not in bytes, but number of entries in the arrays
1738
1739 Level: intermediate
1740
1741 Note:
1742 This routine is a preferred replacement to PetscMemmove()
1743 The arrays must be of the same type
1744
1745.seealso: PetscMemcpy(), PetscMemcmp(), PetscArrayzero(), PetscMemzero(), PetscArraycpy(), PetscMemmove(), PetscArraycmp(), PetscStrallocpy()
1746M*/
1747#define PetscArraymove(str1,str2,cnt)((sizeof(*(str1)) != sizeof(*(str2))) || PetscMemmove(str1,str2
,(cnt)*sizeof(*(str1))))
((sizeof(*(str1)) != sizeof(*(str2))) || PetscMemmove(str1,str2,(cnt)*sizeof(*(str1))))
1748
1749/*MC
1750 PetscArraycpy - Copies from one array in memory to another
1751
1752 Synopsis:
1753 #include <petscsys.h>
1754 PetscErrorCode PetscArraycpy(anytype *str1,const anytype *str2,size_t cnt)
1755
1756 Not Collective
1757
1758 Input Parameters:
1759+ str1 - First array (destination)
1760. str2 - Second array (source)
1761- cnt - Count of the array, not in bytes, but number of entries in the arrays
1762
1763 Level: intermediate
1764
1765 Note:
1766 This routine is a preferred replacement to PetscMemcpy()
1767 The arrays must be of the same type
1768
1769.seealso: PetscMemcpy(), PetscMemcmp(), PetscArrayzero(), PetscMemzero(), PetscArraymove(), PetscMemmove(), PetscArraycmp(), PetscStrallocpy()
1770M*/
1771#define PetscArraycpy(str1,str2,cnt)((sizeof(*(str1)) != sizeof(*(str2))) || PetscMemcpy(str1,str2
,(cnt)*sizeof(*(str1))))
((sizeof(*(str1)) != sizeof(*(str2))) || PetscMemcpy(str1,str2,(cnt)*sizeof(*(str1))))
1772
1773/*MC
1774 PetscArrayzero - Zeros an array in memory.
1775
1776 Synopsis:
1777 #include <petscsys.h>
1778 PetscErrorCode PetscArrayzero(anytype *str1,size_t cnt)
1779
1780 Not Collective
1781
1782 Input Parameters:
1783+ str1 - array
1784- cnt - Count of the array, not in bytes, but number of entries in the array
1785
1786 Level: intermediate
1787
1788 Note:
1789 This routine is a preferred replacement to PetscMemzero()
1790
1791.seealso: PetscMemcpy(), PetscMemcmp(), PetscMemzero(), PetscArraycmp(), PetscArraycpy(), PetscMemmove(), PetscStrallocpy(), PetscArraymove()
1792M*/
1793#define PetscArrayzero(str1,cnt)PetscMemzero(str1,(cnt)*sizeof(*(str1))) PetscMemzero(str1,(cnt)*sizeof(*(str1)))
1794
1795/*MC
1796 PetscPrefetchBlock - Prefetches a block of memory
1797
1798 Synopsis:
1799 #include <petscsys.h>
1800 void PetscPrefetchBlock(const anytype *a,size_t n,int rw,int t)
1801
1802 Not Collective
1803
1804 Input Parameters:
1805+ a - pointer to first element to fetch (any type but usually PetscInt or PetscScalar)
1806. n - number of elements to fetch
1807. rw - 1 if the memory will be written to, otherwise 0 (ignored by many processors)
1808- t - temporal locality (PETSC_PREFETCH_HINT_{NTA,T0,T1,T2}), see note
1809
1810 Level: developer
1811
1812 Notes:
1813 The last two arguments (rw and t) must be compile-time constants.
1814
1815 Adopting Intel's x86/x86-64 conventions, there are four levels of temporal locality. Not all architectures offer
1816 equivalent locality hints, but the following macros are always defined to their closest analogue.
1817+ PETSC_PREFETCH_HINT_NTA - Non-temporal. Prefetches directly to L1, evicts to memory (skips higher level cache unless it was already there when prefetched).
1818. PETSC_PREFETCH_HINT_T0 - Fetch to all levels of cache and evict to the closest level. Use this when the memory will be reused regularly despite necessary eviction from L1.
1819. PETSC_PREFETCH_HINT_T1 - Fetch to level 2 and higher (not L1).
1820- PETSC_PREFETCH_HINT_T2 - Fetch to high-level cache only. (On many systems, T0 and T1 are equivalent.)
1821
1822 This function does nothing on architectures that do not support prefetch and never errors (even if passed an invalid
1823 address).
1824
1825M*/
1826#define PetscPrefetchBlock(a,n,rw,t)do { const char *_p = (const char*)(a),*_end = (const char*)(
(a)+(n)); for ( ; _p < _end; _p += 64) (__builtin_prefetch
((void *)((const char*)(_p)), ((((t))) >> 2) & 1, (
((t))) & 0x3)); } while (0)
do { \
1827 const char *_p = (const char*)(a),*_end = (const char*)((a)+(n)); \
1828 for ( ; _p < _end; _p += PETSC_LEVEL1_DCACHE_LINESIZE64) PETSC_Prefetch(_p,(rw),(t))(__builtin_prefetch((void *)((const char*)(_p)), ((((t))) >>
2) & 1, (((t))) & 0x3))
; \
1829 } while (0)
1830
1831/*
1832 Determine if some of the kernel computation routines use
1833 Fortran (rather than C) for the numerical calculations. On some machines
1834 and compilers (like complex numbers) the Fortran version of the routines
1835 is faster than the C/C++ versions. The flag --with-fortran-kernels
1836 should be used with ./configure to turn these on.
1837*/
1838#if defined(PETSC_USE_FORTRAN_KERNELS)
1839
1840#if !defined(PETSC_USE_FORTRAN_KERNEL_MULTCRL)
1841#define PETSC_USE_FORTRAN_KERNEL_MULTCRL
1842#endif
1843
1844#if !defined(PETSC_USE_FORTRAN_KERNEL_MULTAIJPERM)
1845#define PETSC_USE_FORTRAN_KERNEL_MULTAIJPERM
1846#endif
1847
1848#if !defined(PETSC_USE_FORTRAN_KERNEL_MULTAIJ)
1849#define PETSC_USE_FORTRAN_KERNEL_MULTAIJ
1850#endif
1851
1852#if !defined(PETSC_USE_FORTRAN_KERNEL_MULTTRANSPOSEAIJ)
1853#define PETSC_USE_FORTRAN_KERNEL_MULTTRANSPOSEAIJ
1854#endif
1855
1856#if !defined(PETSC_USE_FORTRAN_KERNEL_NORM)
1857#define PETSC_USE_FORTRAN_KERNEL_NORM
1858#endif
1859
1860#if !defined(PETSC_USE_FORTRAN_KERNEL_MAXPY)
1861#define PETSC_USE_FORTRAN_KERNEL_MAXPY
1862#endif
1863
1864#if !defined(PETSC_USE_FORTRAN_KERNEL_SOLVEAIJ)
1865#define PETSC_USE_FORTRAN_KERNEL_SOLVEAIJ
1866#endif
1867
1868#if !defined(PETSC_USE_FORTRAN_KERNEL_RELAXAIJ)
1869#define PETSC_USE_FORTRAN_KERNEL_RELAXAIJ
1870#endif
1871
1872#if !defined(PETSC_USE_FORTRAN_KERNEL_SOLVEBAIJ)
1873#define PETSC_USE_FORTRAN_KERNEL_SOLVEBAIJ
1874#endif
1875
1876#if !defined(PETSC_USE_FORTRAN_KERNEL_MULTADDAIJ)
1877#define PETSC_USE_FORTRAN_KERNEL_MULTADDAIJ
1878#endif
1879
1880#if !defined(PETSC_USE_FORTRAN_KERNEL_MDOT)
1881#define PETSC_USE_FORTRAN_KERNEL_MDOT
1882#endif
1883
1884#if !defined(PETSC_USE_FORTRAN_KERNEL_XTIMESY)
1885#define PETSC_USE_FORTRAN_KERNEL_XTIMESY
1886#endif
1887
1888#if !defined(PETSC_USE_FORTRAN_KERNEL_AYPX)
1889#define PETSC_USE_FORTRAN_KERNEL_AYPX
1890#endif
1891
1892#if !defined(PETSC_USE_FORTRAN_KERNEL_WAXPY)
1893#define PETSC_USE_FORTRAN_KERNEL_WAXPY
1894#endif
1895
1896#endif
1897
1898/*
1899 Macros for indicating code that should be compiled with a C interface,
1900 rather than a C++ interface. Any routines that are dynamically loaded
1901 (such as the PCCreate_XXX() routines) must be wrapped so that the name
1902 mangler does not change the functions symbol name. This just hides the
1903 ugly extern "C" {} wrappers.
1904*/
1905#if defined(__cplusplus)
1906# define EXTERN_C_BEGIN extern "C" {
1907# define EXTERN_C_END }
1908#else
1909# define EXTERN_C_BEGIN
1910# define EXTERN_C_END
1911#endif
1912
1913/* --------------------------------------------------------------------*/
1914
1915/*MC
1916 MPI_Comm - the basic object used by MPI to determine which processes are involved in a
1917 communication
1918
1919 Level: beginner
1920
1921 Note: This manual page is a place-holder because MPICH does not have a manual page for MPI_Comm
1922
1923.seealso: PETSC_COMM_WORLD, PETSC_COMM_SELF
1924M*/
1925
1926#if defined(PETSC_HAVE_MPIIO1)
1927PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode MPIU_File_write_all(MPI_File,void*,PetscMPIInt,MPI_Datatype,MPI_Status*);
1928PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode MPIU_File_read_all(MPI_File,void*,PetscMPIInt,MPI_Datatype,MPI_Status*);
1929#endif
1930
1931/* the following petsc_static_inline require petscerror.h */
1932
1933/* Limit MPI to 32-bits */
1934#define PETSC_MPI_INT_MAX2147483647 2147483647
1935#define PETSC_MPI_INT_MIN-2147483647 -2147483647
1936/* Limit BLAS to 32-bits */
1937#define PETSC_BLAS_INT_MAX2147483647 2147483647
1938#define PETSC_BLAS_INT_MIN-2147483647 -2147483647
1939
1940/*@C
1941 PetscIntCast - casts a PetscInt64 (which is 64 bits in size) to a PetscInt (which may be 32 bits in size), generates an
1942 error if the PetscInt is not large enough to hold the number.
1943
1944 Not Collective
1945
1946 Input Parameter:
1947. a - the PetscInt64 value
1948
1949 Output Parameter:
1950. b - the resulting PetscInt value
1951
1952 Level: advanced
1953
1954 Notes: If integers needed for the applications are too large to fit in 32 bit ints you can ./configure using --with-64-bit-indices to make PetscInt use 64 bit ints
1955
1956 Not available from Fortran
1957
1958.seealso: PetscBLASInt, PetscMPIInt, PetscInt, PetscMPIIntCast(), PetscBLASIntCast()
1959@*/
1960PETSC_STATIC_INLINEstatic inline PetscErrorCode PetscIntCast(PetscInt64 a,PetscInt *b)
1961{
1962 PetscFunctionBegindo { do { ; if (petscstack && (petscstack->currentsize
< 64)) { petscstack->function[petscstack->currentsize
] = __func__; petscstack->file[petscstack->currentsize]
= "/sandbox/petsc/petsc.next/include/petscsys.h"; petscstack
->line[petscstack->currentsize] = 1962; petscstack->
petscroutine[petscstack->currentsize] = PETSC_TRUE; petscstack
->currentsize++; } if (petscstack) { petscstack->hotdepth
+= (PETSC_FALSE || petscstack->hotdepth); } ; } while (0)
; ; } while (0)
;
1963#if !defined(PETSC_USE_64BIT_INDICES)
1964 if ((a) > PETSC_MAX_INT2147483647) SETERRQ(PETSC_COMM_SELF,PETSC_ERR_ARG_OUTOFRANGE,"Integer too long for PetscInt, you may need to ./configure using --with-64-bit-indices")return PetscError(((MPI_Comm)0x44000001),1964,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,63,PETSC_ERROR_INITIAL,"Integer too long for PetscInt, you may need to ./configure using --with-64-bit-indices"
)
;
1965#endif
1966 *b = (PetscInt)(a);
1967 PetscFunctionReturn(0)do { do { ; if (petscstack && petscstack->currentsize
> 0) { petscstack->currentsize--; petscstack->function
[petscstack->currentsize] = 0; petscstack->file[petscstack
->currentsize] = 0; petscstack->line[petscstack->currentsize
] = 0; petscstack->petscroutine[petscstack->currentsize
] = PETSC_FALSE; } if (petscstack) { petscstack->hotdepth =
(((petscstack->hotdepth-1)<(0)) ? (0) : (petscstack->
hotdepth-1)); } ; } while (0); return(0);} while (0)
;
1968}
1969
1970/*@C
1971 PetscBLASIntCast - casts a PetscInt (which may be 64 bits in size) to a PetscBLASInt (which may be 32 bits in size), generates an
1972 error if the PetscBLASInt is not large enough to hold the number.
1973
1974 Not Collective
1975
1976 Input Parameter:
1977. a - the PetscInt value
1978
1979 Output Parameter:
1980. b - the resulting PetscBLASInt value
1981
1982 Level: advanced
1983
1984 Not available from Fortran
1985
1986.seealso: PetscBLASInt, PetscMPIInt, PetscInt, PetscMPIIntCast(), PetscIntCast()
1987@*/
1988PETSC_STATIC_INLINEstatic inline PetscErrorCode PetscBLASIntCast(PetscInt a,PetscBLASInt *b)
1989{
1990 PetscFunctionBegindo { do { ; if (petscstack && (petscstack->currentsize
< 64)) { petscstack->function[petscstack->currentsize
] = __func__; petscstack->file[petscstack->currentsize]
= "/sandbox/petsc/petsc.next/include/petscsys.h"; petscstack
->line[petscstack->currentsize] = 1990; petscstack->
petscroutine[petscstack->currentsize] = PETSC_TRUE; petscstack
->currentsize++; } if (petscstack) { petscstack->hotdepth
+= (PETSC_FALSE || petscstack->hotdepth); } ; } while (0)
; ; } while (0)
;
1991#if defined(PETSC_USE_64BIT_INDICES) && !defined(PETSC_HAVE_64BIT_BLAS_INDICES)
1992 *b = 0;
1993 if ((a) > PETSC_BLAS_INT_MAX2147483647) SETERRQ(PETSC_COMM_SELF,PETSC_ERR_ARG_OUTOFRANGE,"Array too long for BLAS/LAPACK")return PetscError(((MPI_Comm)0x44000001),1993,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,63,PETSC_ERROR_INITIAL,"Array too long for BLAS/LAPACK")
;
1994#endif
1995 *b = (PetscBLASInt)(a);
1996 PetscFunctionReturn(0)do { do { ; if (petscstack && petscstack->currentsize
> 0) { petscstack->currentsize--; petscstack->function
[petscstack->currentsize] = 0; petscstack->file[petscstack
->currentsize] = 0; petscstack->line[petscstack->currentsize
] = 0; petscstack->petscroutine[petscstack->currentsize
] = PETSC_FALSE; } if (petscstack) { petscstack->hotdepth =
(((petscstack->hotdepth-1)<(0)) ? (0) : (petscstack->
hotdepth-1)); } ; } while (0); return(0);} while (0)
;
1997}
1998
1999/*@C
2000 PetscMPIIntCast - casts a PetscInt (which may be 64 bits in size) to a PetscMPIInt (which may be 32 bits in size), generates an
2001 error if the PetscMPIInt is not large enough to hold the number.
2002
2003 Not Collective
2004
2005 Input Parameter:
2006. a - the PetscInt value
2007
2008 Output Parameter:
2009. b - the resulting PetscMPIInt value
2010
2011 Level: advanced
2012
2013 Not available from Fortran
2014
2015.seealso: PetscBLASInt, PetscMPIInt, PetscInt, PetscBLASIntCast(), PetscIntCast()
2016@*/
2017PETSC_STATIC_INLINEstatic inline PetscErrorCode PetscMPIIntCast(PetscInt a,PetscMPIInt *b)
2018{
2019 PetscFunctionBegindo { do { ; if (petscstack && (petscstack->currentsize
< 64)) { petscstack->function[petscstack->currentsize
] = __func__; petscstack->file[petscstack->currentsize]
= "/sandbox/petsc/petsc.next/include/petscsys.h"; petscstack
->line[petscstack->currentsize] = 2019; petscstack->
petscroutine[petscstack->currentsize] = PETSC_TRUE; petscstack
->currentsize++; } if (petscstack) { petscstack->hotdepth
+= (PETSC_FALSE || petscstack->hotdepth); } ; } while (0)
; ; } while (0)
;
2020#if defined(PETSC_USE_64BIT_INDICES)
2021 *b = 0;
2022 if ((a) > PETSC_MPI_INT_MAX2147483647) SETERRQ(PETSC_COMM_SELF,PETSC_ERR_ARG_OUTOFRANGE,"Array too long for MPI")return PetscError(((MPI_Comm)0x44000001),2022,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,63,PETSC_ERROR_INITIAL,"Array too long for MPI")
;
2023#endif
2024 *b = (PetscMPIInt)(a);
2025 PetscFunctionReturn(0)do { do { ; if (petscstack && petscstack->currentsize
> 0) { petscstack->currentsize--; petscstack->function
[petscstack->currentsize] = 0; petscstack->file[petscstack
->currentsize] = 0; petscstack->line[petscstack->currentsize
] = 0; petscstack->petscroutine[petscstack->currentsize
] = PETSC_FALSE; } if (petscstack) { petscstack->hotdepth =
(((petscstack->hotdepth-1)<(0)) ? (0) : (petscstack->
hotdepth-1)); } ; } while (0); return(0);} while (0)
;
2026}
2027
2028#define PetscInt64Mult(a,b)((PetscInt64)(a))*((PetscInt64)(b)) ((PetscInt64)(a))*((PetscInt64)(b))
2029
2030/*@C
2031
2032 PetscRealIntMultTruncate - Computes the product of a positive PetscReal and a positive PetscInt and truncates the value to slightly less than the maximal possible value
2033
2034 Not Collective
2035
2036 Input Parameter:
2037+ a - the PetscReal value
2038- b - the second value
2039
2040 Returns:
2041 the result as a PetscInt value
2042
2043 Use PetscInt64Mult() to compute the product of two PetscInt as a PetscInt64
2044 Use PetscIntMultTruncate() to compute the product of two positive PetscInt and truncate to fit a PetscInt
2045 Use PetscIntMultError() to compute the product of two PetscInt if you wish to generate an error if the result will not fit in a PetscInt
2046
2047 Developers Note:
2048 We currently assume that PetscInt addition can never overflow, this is obviously wrong but requires many more checks.
2049
2050 This is used where we compute approximate sizes for workspace and need to insure the workspace is index-able.
2051
2052 Not available from Fortran
2053
2054 Level: advanced
2055
2056.seealso: PetscBLASInt, PetscMPIInt, PetscInt, PetscBLASIntCast(), PetscInt64Mult()
2057@*/
2058PETSC_STATIC_INLINEstatic inline PetscInt PetscRealIntMultTruncate(PetscReal a,PetscInt b)
2059{
2060 PetscInt64 r;
2061
2062 r = (PetscInt64) (a*(PetscReal)b);
2063 if (r > PETSC_MAX_INT2147483647 - 100) r = PETSC_MAX_INT2147483647 - 100;
2064 return (PetscInt) r;
2065}
2066
2067/*@C
2068
2069 PetscIntMultTruncate - Computes the product of two positive PetscInt and truncates the value to slightly less than the maximal possible value
2070
2071 Not Collective
2072
2073 Input Parameter:
2074+ a - the PetscInt value
2075- b - the second value
2076
2077 Returns:
2078 the result as a PetscInt value
2079
2080 Use PetscInt64Mult() to compute the product of two PetscInt as a PetscInt64
2081 Use PetscRealIntMultTruncate() to compute the product of a PetscReal and a PetscInt and truncate to fit a PetscInt
2082 Use PetscIntMultError() to compute the product of two PetscInt if you wish to generate an error if the result will not fit in a PetscInt
2083
2084 Not available from Fortran
2085
2086 Developers Note: We currently assume that PetscInt addition can never overflow, this is obviously wrong but requires many more checks.
2087
2088 This is used where we compute approximate sizes for workspace and need to insure the workspace is index-able.
2089
2090 Level: advanced
2091
2092.seealso: PetscBLASInt, PetscMPIInt, PetscInt, PetscBLASIntCast(), PetscInt64Mult()
2093@*/
2094PETSC_STATIC_INLINEstatic inline PetscInt PetscIntMultTruncate(PetscInt a,PetscInt b)
2095{
2096 PetscInt64 r;
2097
2098 r = PetscInt64Mult(a,b)((PetscInt64)(a))*((PetscInt64)(b));
2099 if (r > PETSC_MAX_INT2147483647 - 100) r = PETSC_MAX_INT2147483647 - 100;
2100 return (PetscInt) r;
2101}
2102
2103/*@C
2104
2105 PetscIntSumTruncate - Computes the sum of two positive PetscInt and truncates the value to slightly less than the maximal possible value
2106
2107 Not Collective
2108
2109 Input Parameter:
2110+ a - the PetscInt value
2111- b - the second value
2112
2113 Returns:
2114 the result as a PetscInt value
2115
2116 Use PetscInt64Mult() to compute the product of two PetscInt as a PetscInt64
2117 Use PetscRealIntMultTruncate() to compute the product of a PetscReal and a PetscInt and truncate to fit a PetscInt
2118 Use PetscIntMultError() to compute the product of two PetscInt if you wish to generate an error if the result will not fit in a PetscInt
2119
2120 This is used where we compute approximate sizes for workspace and need to insure the workspace is index-able.
2121
2122 Not available from Fortran
2123
2124 Level: advanced
2125
2126.seealso: PetscBLASInt, PetscMPIInt, PetscInt, PetscBLASIntCast(), PetscInt64Mult()
2127@*/
2128PETSC_STATIC_INLINEstatic inline PetscInt PetscIntSumTruncate(PetscInt a,PetscInt b)
2129{
2130 PetscInt64 r;
2131
2132 r = ((PetscInt64)a) + ((PetscInt64)b);
2133 if (r > PETSC_MAX_INT2147483647 - 100) r = PETSC_MAX_INT2147483647 - 100;
2134 return (PetscInt) r;
2135}
2136
2137/*@C
2138
2139 PetscIntMultError - Computes the product of two positive PetscInt and generates an error with overflow.
2140
2141 Not Collective
2142
2143 Input Parameter:
2144+ a - the PetscInt value
2145- b - the second value
2146
2147 Output Parameter:ma
2148. result - the result as a PetscInt value, or NULL if you do not want the result, you just want to check if it overflows
2149
2150 Use PetscInt64Mult() to compute the product of two 32 bit PetscInt and store in a PetscInt64
2151 Use PetscIntMultTruncate() to compute the product of two PetscInt and truncate it to fit in a PetscInt
2152
2153 Not available from Fortran
2154
2155 Developers Note: We currently assume that PetscInt addition does not overflow, this is obviously wrong but requires many more checks.
2156
2157 Level: advanced
2158
2159.seealso: PetscBLASInt, PetscMPIInt, PetscInt, PetscBLASIntCast(), PetscIntMult64(), PetscIntSumError()
2160@*/
2161PETSC_STATIC_INLINEstatic inline PetscErrorCode PetscIntMultError(PetscInt a,PetscInt b,PetscInt *result)
2162{
2163 PetscInt64 r;
2164
2165 PetscFunctionBegindo { do { ; if (petscstack && (petscstack->currentsize
< 64)) { petscstack->function[petscstack->currentsize
] = __func__; petscstack->file[petscstack->currentsize]
= "/sandbox/petsc/petsc.next/include/petscsys.h"; petscstack
->line[petscstack->currentsize] = 2165; petscstack->
petscroutine[petscstack->currentsize] = PETSC_TRUE; petscstack
->currentsize++; } if (petscstack) { petscstack->hotdepth
+= (PETSC_FALSE || petscstack->hotdepth); } ; } while (0)
; ; } while (0)
;
2166 r = PetscInt64Mult(a,b)((PetscInt64)(a))*((PetscInt64)(b));
2167#if !defined(PETSC_USE_64BIT_INDICES)
2168 if (r > PETSC_MAX_INT2147483647) SETERRQ2(PETSC_COMM_SELF,PETSC_ERR_SUP,"Product of two integer %d %d overflow, you must ./configure PETSc with --with-64-bit-indices for the case you are running",a,b)return PetscError(((MPI_Comm)0x44000001),2168,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,56,PETSC_ERROR_INITIAL,"Product of two integer %d %d overflow, you must ./configure PETSc with --with-64-bit-indices for the case you are running"
,a,b)
;
2169#endif
2170 if (result) *result = (PetscInt) r;
2171 PetscFunctionReturn(0)do { do { ; if (petscstack && petscstack->currentsize
> 0) { petscstack->currentsize--; petscstack->function
[petscstack->currentsize] = 0; petscstack->file[petscstack
->currentsize] = 0; petscstack->line[petscstack->currentsize
] = 0; petscstack->petscroutine[petscstack->currentsize
] = PETSC_FALSE; } if (petscstack) { petscstack->hotdepth =
(((petscstack->hotdepth-1)<(0)) ? (0) : (petscstack->
hotdepth-1)); } ; } while (0); return(0);} while (0)
;
2172}
2173
2174/*@C
2175
2176 PetscIntSumError - Computes the sum of two positive PetscInt and generates an error with overflow.
2177
2178 Not Collective
2179
2180 Input Parameter:
2181+ a - the PetscInt value
2182- b - the second value
2183
2184 Output Parameter:ma
2185. c - the result as a PetscInt value, or NULL if you do not want the result, you just want to check if it overflows
2186
2187 Use PetscInt64Mult() to compute the product of two 32 bit PetscInt and store in a PetscInt64
2188 Use PetscIntMultTruncate() to compute the product of two PetscInt and truncate it to fit in a PetscInt
2189
2190 Not available from Fortran
2191
2192 Level: advanced
2193
2194.seealso: PetscBLASInt, PetscMPIInt, PetscInt, PetscBLASIntCast(), PetscInt64Mult()
2195@*/
2196PETSC_STATIC_INLINEstatic inline PetscErrorCode PetscIntSumError(PetscInt a,PetscInt b,PetscInt *result)
2197{
2198 PetscInt64 r;
2199
2200 PetscFunctionBegindo { do { ; if (petscstack && (petscstack->currentsize
< 64)) { petscstack->function[petscstack->currentsize
] = __func__; petscstack->file[petscstack->currentsize]
= "/sandbox/petsc/petsc.next/include/petscsys.h"; petscstack
->line[petscstack->currentsize] = 2200; petscstack->
petscroutine[petscstack->currentsize] = PETSC_TRUE; petscstack
->currentsize++; } if (petscstack) { petscstack->hotdepth
+= (PETSC_FALSE || petscstack->hotdepth); } ; } while (0)
; ; } while (0)
;
2201 r = ((PetscInt64)a) + ((PetscInt64)b);
2202#if !defined(PETSC_USE_64BIT_INDICES)
2203 if (r > PETSC_MAX_INT2147483647) SETERRQ2(PETSC_COMM_SELF,PETSC_ERR_SUP,"Sum of two integer %d %d overflow, you must ./configure PETSc with --with-64-bit-indices for the case you are running",a,b)return PetscError(((MPI_Comm)0x44000001),2203,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,56,PETSC_ERROR_INITIAL,"Sum of two integer %d %d overflow, you must ./configure PETSc with --with-64-bit-indices for the case you are running"
,a,b)
;
2204#endif
2205 if (result) *result = (PetscInt) r;
2206 PetscFunctionReturn(0)do { do { ; if (petscstack && petscstack->currentsize
> 0) { petscstack->currentsize--; petscstack->function
[petscstack->currentsize] = 0; petscstack->file[petscstack
->currentsize] = 0; petscstack->line[petscstack->currentsize
] = 0; petscstack->petscroutine[petscstack->currentsize
] = PETSC_FALSE; } if (petscstack) { petscstack->hotdepth =
(((petscstack->hotdepth-1)<(0)) ? (0) : (petscstack->
hotdepth-1)); } ; } while (0); return(0);} while (0)
;
2207}
2208
2209/*
2210 The IBM include files define hz, here we hide it so that it may be used as a regular user variable.
2211*/
2212#if defined(hz)
2213# undef hz
2214#endif
2215
2216#include <limits.h>
2217
2218/* The number of bits in a byte */
2219
2220#define PETSC_BITS_PER_BYTE8 CHAR_BIT8
2221
2222/* For arrays that contain filenames or paths */
2223
2224#if defined(PETSC_HAVE_SYS_PARAM_H1)
2225# include <sys/param.h>
2226#endif
2227#if defined(PETSC_HAVE_SYS_TYPES_H1)
2228# include <sys/types.h>
2229#endif
2230#if defined(MAXPATHLEN4096)
2231# define PETSC_MAX_PATH_LEN4096 MAXPATHLEN4096
2232#elif defined(MAX_PATH)
2233# define PETSC_MAX_PATH_LEN4096 MAX_PATH
2234#elif defined(_MAX_PATH)
2235# define PETSC_MAX_PATH_LEN4096 _MAX_PATH
2236#else
2237# define PETSC_MAX_PATH_LEN4096 4096
2238#endif
2239
2240/*MC
2241
2242 PETSC_VERSION - This manual page provides information about how PETSc documents and uses its version information. This information is available to both C/C++
2243 and Fortran compilers when petscsys.h is included.
2244
2245
2246 The current PETSc version and the API for accessing it are defined in petscversion.h
2247
2248 The complete version number is given as the triple PETSC_VERSION_MAJOR.PETSC_VERSION_MINOR.PETSC_VERSION_SUBMINOR (in short hand x.y.z)
2249
2250 A change in the minor version number (y) indicates possible/likely changes in the PETSc API. Note this is different than with the semantic versioning convention
2251 where only a change in the major version number (x) indicates a change in the API.
2252
2253 A subminor greater than zero indicates a patch release. Version x.y.z maintains source and binary compatibility with version x.y.w for all z and w
2254
2255 Use the macros PETSC_VERSION_EQ(x,y,z), PETSC_VERSION_LT(x,y,z), PETSC_VERSION_LE(x,y,z), PETSC_VERSION_GT(x,y,z),
2256 PETSC_VERSION_GE(x,y,z) to determine if the current version is equal to, less than, less than or equal to, greater than or greater than or equal to a given
2257 version number (x.y.z).
2258
2259 PETSC_RELEASE_DATE is the date the x.y version was released (i.e. the version before any patch releases)
2260
2261 PETSC_VERSION_DATE is the date the x.y.z version was released
2262
2263 PETSC_VERSION_GIT is the last git commit to the repository given in the form vx.y.z-wwwww
2264
2265 PETSC_VERSION_DATE_GIT is the date of the last git commit to the repository
2266
2267 Level: intermediate
2268
2269 PETSC_VERSION_() and PETSC_VERSION_PATCH are deprecated and will eventually be removed. For several releases PETSC_VERSION_PATCH is always 0
2270
2271M*/
2272
2273/*MC
2274
2275 UsingFortran - To use PETSc with Fortran you must use both PETSc include files and modules. At the beginning
2276 of every function and module definition you need something like
2277
2278$
2279$#include "petsc/finclude/petscXXX.h"
2280$ use petscXXX
2281
2282 You can declare PETSc variables using either of the following.
2283
2284$ XXX variablename
2285$ type(tXXX) variablename
2286
2287 For example,
2288
2289$#include "petsc/finclude/petscvec.h"
2290$ use petscvec
2291$
2292$ Vec b
2293$ type(tVec) x
2294
2295 Level: beginner
2296
2297M*/
2298
2299PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscGetArchType(char[],size_t);
2300PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscGetHostName(char[],size_t);
2301PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscGetUserName(char[],size_t);
2302PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscGetProgramName(char[],size_t);
2303PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSetProgramName(const char[]);
2304PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscGetDate(char[],size_t);
2305PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscGetVersion(char[], size_t);
2306PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscGetVersionNumber(PetscInt*,PetscInt*,PetscInt*,PetscInt*);
2307
2308PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSortInt(PetscInt,PetscInt[]);
2309PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSortedRemoveDupsInt(PetscInt*,PetscInt[]);
2310PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSortRemoveDupsInt(PetscInt*,PetscInt[]);
2311PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscFindInt(PetscInt, PetscInt, const PetscInt[], PetscInt*);
2312PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscFindMPIInt(PetscMPIInt, PetscInt, const PetscMPIInt[], PetscInt*);
2313PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSortIntWithPermutation(PetscInt,const PetscInt[],PetscInt[]);
2314PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSortStrWithPermutation(PetscInt,const char*[],PetscInt[]);
2315PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSortIntWithArray(PetscInt,PetscInt[],PetscInt[]);
2316PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSortIntWithArrayPair(PetscInt,PetscInt[],PetscInt[],PetscInt[]);
2317PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSortMPIInt(PetscInt,PetscMPIInt[]);
2318PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSortRemoveDupsMPIInt(PetscInt*,PetscMPIInt[]);
2319PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSortMPIIntWithArray(PetscMPIInt,PetscMPIInt[],PetscMPIInt[]);
2320PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSortMPIIntWithIntArray(PetscMPIInt,PetscMPIInt[],PetscInt[]);
2321PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSortIntWithScalarArray(PetscInt,PetscInt[],PetscScalar[]);
2322PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSortIntWithDataArray(PetscInt,PetscInt[],void*,size_t,void*);
2323PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSortReal(PetscInt,PetscReal[]);
2324PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSortRealWithArrayInt(PetscInt,PetscReal[],PetscInt[]);
2325PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSortRealWithPermutation(PetscInt,const PetscReal[],PetscInt[]);
2326PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSortRemoveDupsReal(PetscInt*,PetscReal[]);
2327PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscFindReal(PetscReal,PetscInt,const PetscReal[],PetscReal,PetscInt*);
2328PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSortSplit(PetscInt,PetscInt,PetscScalar[],PetscInt[]);
2329PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSortSplitReal(PetscInt,PetscInt,PetscReal[],PetscInt[]);
2330PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscProcessTree(PetscInt,const PetscBool [],const PetscInt[],PetscInt*,PetscInt**,PetscInt**,PetscInt**,PetscInt**);
2331PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscMergeIntArrayPair(PetscInt,const PetscInt[],const PetscInt[],PetscInt,const PetscInt[],const PetscInt[],PetscInt*,PetscInt**,PetscInt**);
2332PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscMergeIntArray(PetscInt,const PetscInt[],PetscInt,const PetscInt[],PetscInt*,PetscInt**);
2333PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscMergeMPIIntArray(PetscInt,const PetscMPIInt[],PetscInt,const PetscMPIInt[],PetscInt*,PetscMPIInt**);
2334
2335PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSetDisplay(void);
2336PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscGetDisplay(char[],size_t);
2337
2338/*J
2339 PetscRandomType - String with the name of a PETSc randomizer
2340
2341 Level: beginner
2342
2343 Notes:
2344 To use SPRNG or RANDOM123 you must have ./configure PETSc
2345 with the option --download-sprng or --download-random123
2346
2347.seealso: PetscRandomSetType(), PetscRandom, PetscRandomCreate()
2348J*/
2349typedef const char* PetscRandomType;
2350#define PETSCRAND"rand" "rand"
2351#define PETSCRAND48"rand48" "rand48"
2352#define PETSCSPRNG"sprng" "sprng"
2353#define PETSCRANDER48"rander48" "rander48"
2354#define PETSCRANDOM123"random123" "random123"
2355
2356/* Logging support */
2357PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscClassId PETSC_RANDOM_CLASSID;
2358
2359PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscRandomInitializePackage(void);
2360
2361/* Dynamic creation and loading functions */
2362PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscFunctionList PetscRandomList;
2363
2364PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscRandomRegister(const char[],PetscErrorCode (*)(PetscRandom));
2365PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscRandomSetType(PetscRandom, PetscRandomType);
2366PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscRandomSetFromOptions(PetscRandom);
2367PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscRandomGetType(PetscRandom, PetscRandomType*);
2368PETSC_STATIC_INLINEstatic inline PetscErrorCode PetscRandomViewFromOptions(PetscRandom A,PetscObject obj,const char name[]) {return PetscObjectViewFromOptions((PetscObject)A,obj,name);}
2369PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscRandomView(PetscRandom,PetscViewer);
2370
2371PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscRandomCreate(MPI_Comm,PetscRandom*);
2372PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscRandomGetValue(PetscRandom,PetscScalar*);
2373PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscRandomGetValueReal(PetscRandom,PetscReal*);
2374PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscRandomGetInterval(PetscRandom,PetscScalar*,PetscScalar*);
2375PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscRandomSetInterval(PetscRandom,PetscScalar,PetscScalar);
2376PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscRandomSetSeed(PetscRandom,unsigned long);
2377PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscRandomGetSeed(PetscRandom,unsigned long *);
2378PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscRandomSeed(PetscRandom);
2379PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscRandomDestroy(PetscRandom*);
2380
2381PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscGetFullPath(const char[],char[],size_t);
2382PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscGetRelativePath(const char[],char[],size_t);
2383PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscGetWorkingDirectory(char[],size_t);
2384PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscGetRealPath(const char[],char[]);
2385PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscGetHomeDirectory(char[],size_t);
2386PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscTestFile(const char[],char,PetscBool *);
2387PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscTestDirectory(const char[],char,PetscBool *);
2388PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscMkdir(const char[]);
2389PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscMkdtemp(char[]);
2390PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscRMTree(const char[]);
2391
2392PETSC_STATIC_INLINEstatic inline PetscBool PetscBinaryBigEndian(void) {long _petsc_v = 1; return ((char*)&_petsc_v)[0] ? PETSC_FALSE : PETSC_TRUE;}
2393
2394PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscBinaryRead(int,void*,PetscInt,PetscInt*,PetscDataType);
2395PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscBinarySynchronizedRead(MPI_Comm,int,void*,PetscInt,PetscInt*,PetscDataType);
2396PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscBinaryWrite(int,void*,PetscInt,PetscDataType,PetscBool);
2397PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscBinarySynchronizedWrite(MPI_Comm,int,void*,PetscInt,PetscDataType,PetscBool);
2398PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscBinaryOpen(const char[],PetscFileMode,int *);
2399PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscBinaryClose(int);
2400PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSharedTmp(MPI_Comm,PetscBool *);
2401PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSharedWorkingDirectory(MPI_Comm,PetscBool *);
2402PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscGetTmp(MPI_Comm,char[],size_t);
2403PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscFileRetrieve(MPI_Comm,const char[],char[],size_t,PetscBool *);
2404PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscLs(MPI_Comm,const char[],char[],size_t,PetscBool *);
2405#if defined(PETSC_USE_SOCKET_VIEWER1)
2406PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscOpenSocket(const char[],int,int*);
2407#endif
2408
2409PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscBinarySeek(int,off_t,PetscBinarySeekType,off_t*);
2410PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscBinarySynchronizedSeek(MPI_Comm,int,off_t,PetscBinarySeekType,off_t*);
2411PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscByteSwap(void *,PetscDataType,PetscInt);
2412
2413PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSetDebugTerminal(const char[]);
2414PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSetDebugger(const char[],PetscBool );
2415PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSetDefaultDebugger(void);
2416PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSetDebuggerFromString(const char*);
2417PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscAttachDebugger(void);
2418PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscStopForDebugger(void);
2419
2420PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscGatherNumberOfMessages(MPI_Comm,const PetscMPIInt[],const PetscMPIInt[],PetscMPIInt*);
2421PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscGatherMessageLengths(MPI_Comm,PetscMPIInt,PetscMPIInt,const PetscMPIInt[],PetscMPIInt**,PetscMPIInt**);
2422PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscGatherMessageLengths2(MPI_Comm,PetscMPIInt,PetscMPIInt,const PetscMPIInt[],const PetscMPIInt[],PetscMPIInt**,PetscMPIInt**,PetscMPIInt**);
2423PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscPostIrecvInt(MPI_Comm,PetscMPIInt,PetscMPIInt,const PetscMPIInt[],const PetscMPIInt[],PetscInt***,MPI_Request**);
2424PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscPostIrecvScalar(MPI_Comm,PetscMPIInt,PetscMPIInt,const PetscMPIInt[],const PetscMPIInt[],PetscScalar***,MPI_Request**);
2425PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscCommBuildTwoSided(MPI_Comm,PetscMPIInt,MPI_Datatype,PetscMPIInt,const PetscMPIInt*,const void*,PetscMPIInt*,PetscMPIInt**,void*)
2426 PetscAttrMPIPointerWithType(6,3);
2427PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscCommBuildTwoSidedF(MPI_Comm,PetscMPIInt,MPI_Datatype,PetscMPIInt,const PetscMPIInt[],const void*,PetscMPIInt*,PetscMPIInt**,void*,PetscMPIInt,
2428 PetscErrorCode (*send)(MPI_Comm,const PetscMPIInt[],PetscMPIInt,PetscMPIInt,void*,MPI_Request[],void*),
2429 PetscErrorCode (*recv)(MPI_Comm,const PetscMPIInt[],PetscMPIInt,void*,MPI_Request[],void*),void *ctx)
2430 PetscAttrMPIPointerWithType(6,3);
2431PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscCommBuildTwoSidedFReq(MPI_Comm,PetscMPIInt,MPI_Datatype,PetscMPIInt,const PetscMPIInt[],const void*,PetscMPIInt*,PetscMPIInt**,void*,PetscMPIInt,
2432 MPI_Request**,MPI_Request**,
2433 PetscErrorCode (*send)(MPI_Comm,const PetscMPIInt[],PetscMPIInt,PetscMPIInt,void*,MPI_Request[],void*),
2434 PetscErrorCode (*recv)(MPI_Comm,const PetscMPIInt[],PetscMPIInt,void*,MPI_Request[],void*),void *ctx)
2435 PetscAttrMPIPointerWithType(6,3);
2436
2437PETSC_EXTERNextern __attribute__((visibility ("default"))) const char *const PetscBuildTwoSidedTypes[];
2438PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscCommBuildTwoSidedSetType(MPI_Comm,PetscBuildTwoSidedType);
2439PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscCommBuildTwoSidedGetType(MPI_Comm,PetscBuildTwoSidedType*);
2440
2441PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSSEIsEnabled(MPI_Comm,PetscBool*,PetscBool*);
2442
2443PETSC_EXTERNextern __attribute__((visibility ("default"))) MPI_Comm PetscObjectComm(PetscObject);
2444
2445PETSC_EXTERNextern __attribute__((visibility ("default"))) const char *const PetscSubcommTypes[];
2446
2447struct _n_PetscSubcomm {
2448 MPI_Comm parent; /* parent communicator */
2449 MPI_Comm dupparent; /* duplicate parent communicator, under which the processors of this subcomm have contiguous rank */
2450 MPI_Comm child; /* the sub-communicator */
2451 PetscMPIInt n; /* num of subcommunicators under the parent communicator */
2452 PetscMPIInt color; /* color of processors belong to this communicator */
2453 PetscMPIInt *subsize; /* size of subcommunicator[color] */
2454 PetscSubcommType type;
2455 char *subcommprefix;
2456};
2457
2458PETSC_STATIC_INLINEstatic inline MPI_Comm PetscSubcommParent(PetscSubcomm scomm) {return scomm->parent;}
2459PETSC_STATIC_INLINEstatic inline MPI_Comm PetscSubcommChild(PetscSubcomm scomm) {return scomm->child;}
9
Access to field 'child' results in a dereference of a null pointer (loaded from variable 'scomm')
2460PETSC_STATIC_INLINEstatic inline MPI_Comm PetscSubcommContiguousParent(PetscSubcomm scomm) {return scomm->dupparent;}
2461PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSubcommCreate(MPI_Comm,PetscSubcomm*);
2462PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSubcommDestroy(PetscSubcomm*);
2463PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSubcommSetNumber(PetscSubcomm,PetscInt);
2464PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSubcommSetType(PetscSubcomm,PetscSubcommType);
2465PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSubcommSetTypeGeneral(PetscSubcomm,PetscMPIInt,PetscMPIInt);
2466PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSubcommView(PetscSubcomm,PetscViewer);
2467PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSubcommSetFromOptions(PetscSubcomm);
2468PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSubcommSetOptionsPrefix(PetscSubcomm,const char[]);
2469
2470PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscHeapCreate(PetscInt,PetscHeap*);
2471PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscHeapAdd(PetscHeap,PetscInt,PetscInt);
2472PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscHeapPop(PetscHeap,PetscInt*,PetscInt*);
2473PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscHeapPeek(PetscHeap,PetscInt*,PetscInt*);
2474PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscHeapStash(PetscHeap,PetscInt,PetscInt);
2475PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscHeapUnstash(PetscHeap);
2476PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscHeapDestroy(PetscHeap*);
2477PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscHeapView(PetscHeap,PetscViewer);
2478
2479PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscProcessPlacementView(PetscViewer);
2480PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscShmCommGet(MPI_Comm,PetscShmComm*);
2481PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscShmCommGlobalToLocal(PetscShmComm,PetscMPIInt,PetscMPIInt*);
2482PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscShmCommLocalToGlobal(PetscShmComm,PetscMPIInt,PetscMPIInt*);
2483PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscShmCommGetMpiShmComm(PetscShmComm,MPI_Comm*);
2484
2485/* routines to better support OpenMP multithreading needs of some PETSc third party libraries */
2486PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscOmpCtrlCreate(MPI_Comm,PetscInt,PetscOmpCtrl*);
2487PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscOmpCtrlGetOmpComms(PetscOmpCtrl,MPI_Comm*,MPI_Comm*,PetscBool*);
2488PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscOmpCtrlDestroy(PetscOmpCtrl*);
2489PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscOmpCtrlBarrier(PetscOmpCtrl);
2490PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscOmpCtrlOmpRegionOnMasterBegin(PetscOmpCtrl);
2491PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscOmpCtrlOmpRegionOnMasterEnd(PetscOmpCtrl);
2492
2493PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSegBufferCreate(size_t,size_t,PetscSegBuffer*);
2494PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSegBufferDestroy(PetscSegBuffer*);
2495PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSegBufferGet(PetscSegBuffer,size_t,void*);
2496PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSegBufferExtractAlloc(PetscSegBuffer,void*);
2497PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSegBufferExtractTo(PetscSegBuffer,void*);
2498PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSegBufferExtractInPlace(PetscSegBuffer,void*);
2499PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSegBufferGetSize(PetscSegBuffer,size_t*);
2500PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscSegBufferUnuse(PetscSegBuffer,size_t);
2501
2502
2503/* Type-safe wrapper to encourage use of PETSC_RESTRICT. Does not use PetscFunctionBegin because the error handling
2504 * prevents the compiler from completely erasing the stub. This is called in inner loops so it has to be as fast as
2505 * possible. */
2506PETSC_STATIC_INLINEstatic inline PetscErrorCode PetscSegBufferGetInts(PetscSegBuffer seg,size_t count,PetscInt *PETSC_RESTRICT__restrict *slot) {return PetscSegBufferGet(seg,count,(void**)slot);}
2507
2508extern PetscOptionsHelpPrinted PetscOptionsHelpPrintedSingleton;
2509PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscOptionsHelpPrintedDestroy(PetscOptionsHelpPrinted*);
2510PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscOptionsHelpPrintedCreate(PetscOptionsHelpPrinted*);
2511PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscOptionsHelpPrintedCheck(PetscOptionsHelpPrinted,const char*,const char*,PetscBool*);
2512
2513#include <stdarg.h>
2514PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscVSNPrintf(char*,size_t,const char[],size_t*,va_list);
2515PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode (*PetscVFPrintf)(FILE*,const char[],va_list);
2516
2517PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscSegBuffer PetscCitationsList;
2518
2519/*@C
2520 PetscCitationsRegister - Register a bibtex item to obtain credit for an implemented algorithm used in the code.
2521
2522 Not Collective - only what is registered on rank 0 of PETSC_COMM_WORLD will be printed
2523
2524 Input Parameters:
2525+ cite - the bibtex item, formated to displayed on multiple lines nicely
2526- set - a boolean variable initially set to PETSC_FALSE; this is used to insure only a single registration of the citation
2527
2528 Level: intermediate
2529
2530 Not available from Fortran
2531
2532 Options Database:
2533. -citations [filename] - print out the bibtex entries for the given computation
2534@*/
2535PETSC_STATIC_INLINEstatic inline PetscErrorCode PetscCitationsRegister(const char cit[],PetscBool *set)
2536{
2537 size_t len;
2538 char *vstring;
2539 PetscErrorCode ierr;
2540
2541 PetscFunctionBegindo { do { ; if (petscstack && (petscstack->currentsize
< 64)) { petscstack->function[petscstack->currentsize
] = __func__; petscstack->file[petscstack->currentsize]
= "/sandbox/petsc/petsc.next/include/petscsys.h"; petscstack
->line[petscstack->currentsize] = 2541; petscstack->
petscroutine[petscstack->currentsize] = PETSC_TRUE; petscstack
->currentsize++; } if (petscstack) { petscstack->hotdepth
+= (PETSC_FALSE || petscstack->hotdepth); } ; } while (0)
; ; } while (0)
;
2542 if (set && *set) PetscFunctionReturn(0)do { do { ; if (petscstack && petscstack->currentsize
> 0) { petscstack->currentsize--; petscstack->function
[petscstack->currentsize] = 0; petscstack->file[petscstack
->currentsize] = 0; petscstack->line[petscstack->currentsize
] = 0; petscstack->petscroutine[petscstack->currentsize
] = PETSC_FALSE; } if (petscstack) { petscstack->hotdepth =
(((petscstack->hotdepth-1)<(0)) ? (0) : (petscstack->
hotdepth-1)); } ; } while (0); return(0);} while (0)
;
2543 ierr = PetscStrlen(cit,&len);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),2543,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
2544 ierr = PetscSegBufferGet(PetscCitationsList,len,&vstring);CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),2544,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
2545 ierr = PetscArraycpy(vstring,cit,len)((sizeof(*(vstring)) != sizeof(*(cit))) || PetscMemcpy(vstring
,cit,(len)*sizeof(*(vstring))))
;CHKERRQ(ierr)do {if (__builtin_expect(!!(ierr),0)) return PetscError(((MPI_Comm
)0x44000001),2545,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
,ierr,PETSC_ERROR_REPEAT," ");} while (0)
;
2546 if (set) *set = PETSC_TRUE;
2547 PetscFunctionReturn(0)do { do { ; if (petscstack && petscstack->currentsize
> 0) { petscstack->currentsize--; petscstack->function
[petscstack->currentsize] = 0; petscstack->file[petscstack
->currentsize] = 0; petscstack->line[petscstack->currentsize
] = 0; petscstack->petscroutine[petscstack->currentsize
] = PETSC_FALSE; } if (petscstack) { petscstack->hotdepth =
(((petscstack->hotdepth-1)<(0)) ? (0) : (petscstack->
hotdepth-1)); } ; } while (0); return(0);} while (0)
;
2548}
2549
2550PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscURLShorten(const char[],char[],size_t);
2551PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscGoogleDriveAuthorize(MPI_Comm,char[],char[],size_t);
2552PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscGoogleDriveRefresh(MPI_Comm,const char[],char[],size_t);
2553PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscGoogleDriveUpload(MPI_Comm,const char[],const char []);
2554
2555PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscBoxAuthorize(MPI_Comm,char[],char[],size_t);
2556PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscBoxRefresh(MPI_Comm,const char[],char[],char[],size_t);
2557
2558PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscGlobusGetTransfers(MPI_Comm,const char[],char[],size_t);
2559
2560PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscTextBelt(MPI_Comm,const char[],const char[],PetscBool*);
2561PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscTellMyCell(MPI_Comm,const char[],const char[],PetscBool*);
2562
2563PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscPullJSONValue(const char[],const char[],char[],size_t,PetscBool*);
2564PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscPushJSONValue(char[],const char[],const char[],size_t);
2565
2566
2567#if defined(PETSC_USE_DEBUG1)
2568/*
2569 Verify that all processes in the communicator have called this from the same line of code
2570 */
2571PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode PetscAllreduceBarrierCheck(MPI_Comm,PetscMPIInt,int,const char*,const char *);
2572
2573/*MC
2574 MPIU_Allreduce - a PETSc replacement for MPI_Allreduce() that tries to determine if the call from all the MPI processes occur from the
2575 same place in the PETSc code. This helps to detect bugs where different MPI processes follow different code paths
2576 resulting in inconsistent and incorrect calls to MPI_Allreduce().
2577
2578 Collective
2579
2580 Synopsis:
2581 #include <petscsys.h>
2582 PetscErrorCode MPIU_Allreduce(void *indata,void *outdata,PetscMPIInt count,MPI_Datatype datatype, MPI_Op op, MPI_Comm comm);
2583
2584 Input Parameters:
2585+ indata - pointer to the input data to be reduced
2586. count - the number of MPI data items in indata and outdata
2587. datatype - the MPI datatype, for example MPI_INT
2588. op - the MPI operation, for example MPI_SUM
2589- comm - the MPI communicator on which the operation occurs
2590
2591 Output Parameter:
2592. outdata - the reduced values
2593
2594 Notes:
2595 In optimized mode this directly calls MPI_Allreduce()
2596
2597 Level: developer
2598
2599.seealso: MPI_Allreduce()
2600M*/
2601#define MPIU_Allreduce(a,b,c,d,e,fcomm)(PetscAllreduceBarrierCheck(fcomm,c,2601,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
) || ((petsc_allreduce_ct += PetscMPIParallelComm((fcomm)),0)
|| MPI_Allreduce((a),(b),(c),(d),(e),(fcomm))))
(PetscAllreduceBarrierCheck(fcomm,c,__LINE__2601,PETSC_FUNCTION_NAME__func__,__FILE__"/sandbox/petsc/petsc.next/include/petscsys.h") || MPI_Allreduce(a,b,c,d,e,fcomm)((petsc_allreduce_ct += PetscMPIParallelComm((fcomm)),0) || MPI_Allreduce
((a),(b),(c),(d),(e),(fcomm)))
)
2602#else
2603#define MPIU_Allreduce(a,b,c,d,e,fcomm)(PetscAllreduceBarrierCheck(fcomm,c,2603,__func__,"/sandbox/petsc/petsc.next/include/petscsys.h"
) || ((petsc_allreduce_ct += PetscMPIParallelComm((fcomm)),0)
|| MPI_Allreduce((a),(b),(c),(d),(e),(fcomm))))
MPI_Allreduce(a,b,c,d,e,fcomm)((petsc_allreduce_ct += PetscMPIParallelComm((fcomm)),0) || MPI_Allreduce
((a),(b),(c),(d),(e),(fcomm)))
2604#endif
2605
2606#if defined(PETSC_HAVE_MPI_WIN_CREATE_FEATURE)
2607PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode MPIU_Win_allocate_shared(MPI_Aint,PetscMPIInt,MPI_Info,MPI_Comm,void*,MPI_Win*);
2608PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscErrorCode MPIU_Win_shared_query(MPI_Win,PetscMPIInt,MPI_Aint*,PetscMPIInt*,void*);
2609#endif
2610
2611/*
2612 Returned from PETSc functions that are called from MPI, such as related to attributes
2613*/
2614PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscMPIInt PETSC_MPI_ERROR_CLASS;
2615PETSC_EXTERNextern __attribute__((visibility ("default"))) PetscMPIInt PETSC_MPI_ERROR_CODE;
2616
2617#endif