Skip to content
GitLab
Menu
Projects
Groups
Snippets
Help
Help
Support
Community forum
Keyboard shortcuts
?
Submit feedback
Contribute to GitLab
Sign in
Toggle navigation
Menu
Open sidebar
Aland, Sebastian
amdis
Commits
515af748
Commit
515af748
authored
Nov 29, 2012
by
Thomas Witkowski
Browse files
Fixed mpi tests.
parent
1908148a
Changes
10
Hide whitespace changes
Inline
Side-by-side
test/mpi/init/test0001.dat.2d
View file @
515af748
...
...
@@ -8,7 +8,7 @@ ellipt->dim: 2
ellipt->components: 1
ellipt->polynomial degree[0]: 1
ellipt->solver: c
g
ellipt->solver:
pets
c
ellipt->solver->max iteration: 10
ellipt->solver->tolerance: 1.e-8
ellipt->solver->info: 10
...
...
test/mpi/init/test0002.dat.2d
View file @
515af748
...
...
@@ -8,11 +8,11 @@ ellipt->dim: 2
ellipt->components: 1
ellipt->polynomial degree[0]: 1
ellipt->solver:
cg
ellipt->solver:
petsc-feti
ellipt->solver->max iteration: 10
ellipt->solver->tolerance: 1.e-8
ellipt->solver->info: 10
ellipt->solver->left precon:
diag
ellipt->solver->left precon:
no
ellipt->solver->right precon: no
ellipt->estimator[0]: 0
...
...
@@ -22,4 +22,4 @@ parallel->log main rank: 0
parallel->pre refine: 0
parallel->partitioner: checker
parallel->multi level test: 0
parallel->solver: petsc-feti
test/mpi/init/test0003.dat.2d
View file @
515af748
...
...
@@ -8,11 +8,11 @@ ellipt->dim: 2
ellipt->components: 1
ellipt->polynomial degree[0]: 1
ellipt->solver:
cg
ellipt->solver:
petsc-feti
ellipt->solver->max iteration: 10
ellipt->solver->tolerance: 1.e-8
ellipt->solver->info: 10
ellipt->solver->left precon:
diag
ellipt->solver->left precon:
no
ellipt->solver->right precon: no
ellipt->estimator[0]: 0
...
...
@@ -23,4 +23,3 @@ parallel->pre refine: 0
parallel->partitioner: checker
parallel->partitioner->mode: multilevel
parallel->multi level test: 1
parallel->solver: petsc-feti
test/mpi/init/test0004.dat.2d
View file @
515af748
...
...
@@ -8,11 +8,11 @@ ellipt->dim: 2
ellipt->components: 1
ellipt->polynomial degree[0]: 1
ellipt->solver:
cg
ellipt->solver:
petsc-feti
ellipt->solver->max iteration: 10
ellipt->solver->tolerance: 1.e-8
ellipt->solver->info: 10
ellipt->solver->left precon:
diag
ellipt->solver->left precon:
no
ellipt->solver->right precon: no
ellipt->estimator[0]: 0
...
...
@@ -23,4 +23,4 @@ parallel->pre refine: 0
parallel->partitioner: checker
parallel->partitioner->mode: multilevel
parallel->multi level test: 1
parallel->solver: petsc-feti
test/mpi/init/test0005.dat.2d
View file @
515af748
...
...
@@ -8,11 +8,11 @@ ellipt->dim: 2
ellipt->components: 1
ellipt->polynomial degree[0]: 1
ellipt->solver:
cg
ellipt->solver:
petsc-feti
ellipt->solver->max iteration: 10
ellipt->solver->tolerance: 1.e-8
ellipt->solver->info: 10
ellipt->solver->left precon:
diag
ellipt->solver->left precon:
no
ellipt->solver->right precon: no
ellipt->estimator[0]: 0
...
...
@@ -23,4 +23,4 @@ parallel->pre refine: 0
parallel->partitioner: checker
parallel->partitioner->mode: multilevel
parallel->multi level test: 1
parallel->solver: petsc-feti
test/mpi/init/test0006.dat.2d
View file @
515af748
...
...
@@ -8,11 +8,11 @@ ellipt->dim: 2
ellipt->components: 1
ellipt->polynomial degree[0]: 1
ellipt->solver:
cg
ellipt->solver:
petsc-feti
ellipt->solver->max iteration: 10
ellipt->solver->tolerance: 1.e-8
ellipt->solver->info: 10
ellipt->solver->left precon:
diag
ellipt->solver->left precon:
no
ellipt->solver->right precon: no
ellipt->estimator[0]: 0
...
...
@@ -23,4 +23,4 @@ parallel->pre refine: 0
parallel->partitioner: checker
parallel->partitioner->mode: multilevel
parallel->multi level test: 1
parallel->solver: petsc-feti
test/mpi/init/test0007.dat.2d
View file @
515af748
...
...
@@ -8,7 +8,7 @@ ellipt->dim: 2
ellipt->components: 1
ellipt->polynomial degree[0]: 1
ellipt->solver: c
g
ellipt->solver:
pets
c
ellipt->solver->max iteration: 10
ellipt->solver->tolerance: 1.e-8
ellipt->solver->info: 10
...
...
test/mpi/src/test0002.cc
View file @
515af748
...
...
@@ -22,7 +22,6 @@ BOOST_AUTO_TEST_CASE(amdis_mpi_feti)
MeshDistributor
*
meshDist
=
MeshDistributor
::
globalMeshDistributor
;
meshDist
->
initParallelization
();
ParallelDofMapping
&
dofMap
=
meshDist
->
getDofMap
();
Mesh
*
mesh
=
ellipt
.
getMesh
();
BOOST_REQUIRE
(
mesh
->
getNumberOfLeaves
()
==
8
);
...
...
@@ -32,13 +31,13 @@ BOOST_AUTO_TEST_CASE(amdis_mpi_feti)
feSpaces
.
push_back
(
ellipt
.
getFeSpace
(
0
));
PetscSolverFeti
&
feti
=
*
(
dynamic_cast
<
PetscSolverFeti
*>
(
ellipt
.
getPetscSolver
()));
feti
.
setMeshDistributor
(
meshDist
);
feti
.
initialize
(
feSpaces
);
feti
.
initialize
();
feti
.
createFetiData
();
BOOST_REQUIRE
(
feti
.
getNumberOfPrimals
()
==
21
);
BOOST_REQUIRE
(
feti
.
getNumberOfDuals
()
==
48
);
ParallelDofMapping
&
dofMap
=
ellipt
.
getPetscSolver
()
->
getDofMap
();
Spreadsheet
sheet
;
sheet
.
read
(
"data/data0002a"
);
...
...
test/mpi/src/test0003.cc
View file @
515af748
...
...
@@ -32,8 +32,7 @@ BOOST_AUTO_TEST_CASE(amdis_mpi_feti_multilevel)
feSpaces
.
push_back
(
ellipt
.
getFeSpace
(
0
));
PetscSolverFeti
&
feti
=
*
(
dynamic_cast
<
PetscSolverFeti
*>
(
ellipt
.
getPetscSolver
()));
feti
.
setMeshDistributor
(
meshDist
);
feti
.
initialize
(
feSpaces
);
feti
.
initialize
();
feti
.
createFetiData
();
if
(
feti
.
getNumberOfPrimals
()
!=
5
)
{
...
...
@@ -48,8 +47,10 @@ BOOST_AUTO_TEST_CASE(amdis_mpi_feti_multilevel)
BOOST_ERROR
(
"Number of lagrange constraints is "
+
lexical_cast
<
string
>
(
feti
.
getNumberOfLagrange
())
+
" and not 24!
\n
"
);
}
if
(
meshDist
->
getDofMapSd
().
getOverallDofs
()
!=
25
)
{
BOOST_ERROR
(
"Number of subdomain DOFs is "
+
lexical_cast
<
string
>
(
meshDist
->
getDofMapSd
().
getOverallDofs
())
+
" and not 25!
\n
"
);
ParallelDofMapping
&
dofMapSd
=
ellipt
.
getPetscSolver
()
->
getDofMapSd
();
if
(
dofMapSd
.
getOverallDofs
()
!=
25
)
{
BOOST_ERROR
(
"Number of subdomain DOFs is "
+
lexical_cast
<
string
>
(
dofMapSd
.
getOverallDofs
())
+
" and not 25!
\n
"
);
}
DofComm
&
commSd
=
meshDist
->
getDofCommSd
();
...
...
test/mpi/src/test0004.cc
View file @
515af748
...
...
@@ -37,8 +37,8 @@ BOOST_AUTO_TEST_CASE(amdis_mpi_feti_multilevel)
DofComm
::
LevelDataType
&
sendDofs
=
meshDist
->
getDofComm
().
getSendDofs
();
DofComm
::
LevelDataType
&
recvDofs
=
meshDist
->
getDofComm
().
getRecvDofs
();
ParallelDofMapping
dofMap
;
dofMap
.
init
(
levelData
,
feSpaces
,
feSpaces
,
true
,
true
);
ParallelDofMapping
dofMap
(
COMPONENT_WISE
)
;
dofMap
.
init
(
levelData
,
feSpaces
,
feSpaces
);
dofMap
.
setDofComm
(
meshDist
->
getDofComm
());
dofMap
.
clear
();
...
...
Write
Preview
Supports
Markdown
0%
Try again
or
attach a new file
.
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment