// -*- mode: C++; tab-width: 4; indent-tabs-mode: nil; c-basic-offset: 4 -*-
// vi: set et ts=4 sw=4 sts=4:
/*
This file is part of the Open Porous Media project (OPM).
OPM is free software: you can redistribute it and/or modify
it under the terms of the GNU General Public License as published by
the Free Software Foundation, either version 2 of the License, or
(at your option) any later version.
OPM is distributed in the hope that it will be useful,
but WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
GNU General Public License for more details.
You should have received a copy of the GNU General Public License
along with OPM. If not, see .
Consult the COPYING file in the top-level source directory of this
module for the precise wording of the license and the list of
copyright holders.
*/
/*!
* \file
* \copydoc Opm::Linear::ParallelAmgBackend
*/
#ifndef EWOMS_PARALLEL_AMG_BACKEND_HH
#define EWOMS_PARALLEL_AMG_BACKEND_HH
#include "linalgproperties.hh"
#include "parallelbasebackend.hh"
#include "bicgstabsolver.hh"
#include "combinedcriterion.hh"
#include "istlsparsematrixadapter.hh"
#include
#include
#include
#include
#include
#include
namespace Opm::Linear {
template
class ParallelAmgBackend;
} // namespace Opm::Linear
namespace Opm::Properties {
// Create new type tags
namespace TTag {
struct ParallelAmgLinearSolver { using InheritsFrom = std::tuple; };
} // end namespace TTag
//! The target number of DOFs per processor for the parallel algebraic
//! multi-grid solver
template
struct AmgCoarsenTarget { static constexpr int value = 5000; };
template
struct LinearSolverMaxError
{
using type = GetPropType;
static constexpr type value = 1e7;
};
template
struct LinearSolverBackend
{ using type = Opm::Linear::ParallelAmgBackend; };
} // namespace Opm::Properties
namespace Opm {
namespace Linear {
/*!
* \ingroup Linear
*
* \brief Provides a linear solver backend using the parallel
* algebraic multi-grid (AMG) linear solver from DUNE-ISTL.
*/
template
class ParallelAmgBackend : public ParallelBaseBackend
{
using ParentType = ParallelBaseBackend;
using Scalar = GetPropType;
using LinearSolverScalar = GetPropType;
using Simulator = GetPropType;
using GridView = GetPropType;
using Overlap = GetPropType;
using SparseMatrixAdapter = GetPropType;
using ParallelOperator = typename ParentType::ParallelOperator;
using OverlappingVector = typename ParentType::OverlappingVector;
using ParallelPreconditioner = typename ParentType::ParallelPreconditioner;
using ParallelScalarProduct = typename ParentType::ParallelScalarProduct;
static constexpr int numEq = getPropValue();
using VectorBlock = Dune::FieldVector;
using MatrixBlock = typename SparseMatrixAdapter::MatrixBlock;
using IstlMatrix = typename SparseMatrixAdapter::IstlMatrix;
using Vector = Dune::BlockVector;
// define the smoother used for the AMG and specify its
// arguments
using SequentialSmoother = Dune::SeqSOR;
// using SequentialSmoother = Dune::SeqSSOR;
// using SequentialSmoother = Dune::SeqJac;
#if DUNE_VERSION_NEWER(DUNE_ISTL, 2,7)
// using SequentialSmoother = Dune::SeqILU;
#else
// using SequentialSmoother = Dune::SeqILU0;
// using SequentialSmoother = Dune::SeqILUn;
#endif
#if HAVE_MPI
using OwnerOverlapCopyCommunication = Dune::OwnerOverlapCopyCommunication;
using FineOperator = Dune::OverlappingSchwarzOperator;
using FineScalarProduct = Dune::OverlappingSchwarzScalarProduct;
using ParallelSmoother = Dune::BlockPreconditioner;
using AMG = Dune::Amg::AMG;
#else
using FineOperator = Dune::MatrixAdapter;
using FineScalarProduct = Dune::SeqScalarProduct;
using ParallelSmoother = SequentialSmoother;
using AMG = Dune::Amg::AMG;
#endif
using RawLinearSolver = BiCGStabSolver ;
static_assert(std::is_same >::value,
"The ParallelAmgBackend linear solver backend requires the IstlSparseMatrixAdapter");
public:
ParallelAmgBackend(const Simulator& simulator)
: ParentType(simulator)
{ }
static void registerParameters()
{
ParentType::registerParameters();
EWOMS_REGISTER_PARAM(TypeTag, Scalar, LinearSolverMaxError,
"The maximum residual error which the linear solver tolerates"
" without giving up");
EWOMS_REGISTER_PARAM(TypeTag, int, AmgCoarsenTarget,
"The coarsening target for the agglomerations of "
"the AMG preconditioner");
}
protected:
friend ParentType;
std::shared_ptr preparePreconditioner_()
{
#if HAVE_MPI
// create and initialize DUNE's OwnerOverlapCopyCommunication
// using the domestic overlap
istlComm_ = std::make_shared(MPI_COMM_WORLD);
setupAmgIndexSet_(this->overlappingMatrix_->overlap(), istlComm_->indexSet());
istlComm_->remoteIndices().template rebuild();
#endif
// create the parallel scalar product and the parallel operator
#if HAVE_MPI
fineOperator_ = std::make_shared(*this->overlappingMatrix_, *istlComm_);
#else
fineOperator_ = std::make_shared(*this->overlappingMatrix_);
#endif
setupAmg_();
return amg_;
}
void cleanupPreconditioner_()
{ /* nothing to do */ }
std::shared_ptr prepareSolver_(ParallelOperator& parOperator,
ParallelScalarProduct& parScalarProduct,
AMG& parPreCond)
{
const auto& gridView = this->simulator_.gridView();
using CCC = CombinedCriterion;
Scalar linearSolverTolerance = EWOMS_GET_PARAM(TypeTag, Scalar, LinearSolverTolerance);
Scalar linearSolverAbsTolerance = EWOMS_GET_PARAM(TypeTag, Scalar, LinearSolverAbsTolerance);
if(linearSolverAbsTolerance < 0.0)
linearSolverAbsTolerance = this->simulator_.model().newtonMethod().tolerance()/100.0;
convCrit_.reset(new CCC(gridView.comm(),
/*residualReductionTolerance=*/linearSolverTolerance,
/*absoluteResidualTolerance=*/linearSolverAbsTolerance,
EWOMS_GET_PARAM(TypeTag, Scalar, LinearSolverMaxError)));
auto bicgstabSolver =
std::make_shared(parPreCond, *convCrit_, parScalarProduct);
int verbosity = 0;
if (parOperator.overlap().myRank() == 0)
verbosity = EWOMS_GET_PARAM(TypeTag, int, LinearSolverVerbosity);
bicgstabSolver->setVerbosity(verbosity);
bicgstabSolver->setMaxIterations(EWOMS_GET_PARAM(TypeTag, int, LinearSolverMaxIterations));
bicgstabSolver->setLinearOperator(&parOperator);
bicgstabSolver->setRhs(this->overlappingb_);
return bicgstabSolver;
}
std::pair runSolver_(std::shared_ptr solver)
{
bool converged = solver->apply(*this->overlappingx_);
return std::make_pair(converged, int(solver->report().iterations()));
}
void cleanupSolver_()
{ /* nothing to do */ }
#if HAVE_MPI
template
void setupAmgIndexSet_(const Overlap& overlap, ParallelIndexSet& istlIndices)
{
using GridAttributes = Dune::OwnerOverlapCopyAttributeSet;
using GridAttributeSet = Dune::OwnerOverlapCopyAttributeSet::AttributeSet;
// create DUNE's ParallelIndexSet from a domestic overlap
istlIndices.beginResize();
for (Index curIdx = 0; static_cast(curIdx) < overlap.numDomestic(); ++curIdx) {
GridAttributeSet gridFlag =
overlap.iAmMasterOf(curIdx)
? GridAttributes::owner
: GridAttributes::copy;
// an index is used by other processes if it is in the
// domestic or in the foreign overlap.
bool isShared = overlap.isInOverlap(curIdx);
assert(curIdx == overlap.globalToDomestic(overlap.domesticToGlobal(curIdx)));
istlIndices.add(/*globalIdx=*/overlap.domesticToGlobal(curIdx),
Dune::ParallelLocalIndex(static_cast(curIdx),
gridFlag,
isShared));
}
istlIndices.endResize();
}
#endif
void setupAmg_()
{
if (amg_)
amg_.reset();
int verbosity = 0;
if (this->simulator_.vanguard().gridView().comm().rank() == 0)
verbosity = EWOMS_GET_PARAM(TypeTag, int, LinearSolverVerbosity);
using SmootherArgs = typename Dune::Amg::SmootherTraits::Arguments;
SmootherArgs smootherArgs;
smootherArgs.iterations = 1;
smootherArgs.relaxationFactor = 1.0;
// specify the coarsen criterion:
//
// using CoarsenCriterion =
// Dune::Amg::CoarsenCriterion>
using CoarsenCriterion = Dune::Amg::
CoarsenCriterion >;
int coarsenTarget = EWOMS_GET_PARAM(TypeTag, int, AmgCoarsenTarget);
CoarsenCriterion coarsenCriterion(/*maxLevel=*/15, coarsenTarget);
coarsenCriterion.setDefaultValuesAnisotropic(GridView::dimension,
/*aggregateSizePerDim=*/3);
if (verbosity > 0)
coarsenCriterion.setDebugLevel(1);
else
coarsenCriterion.setDebugLevel(0); // make the AMG shut up
// reduce the minium coarsen rate (default is 1.2)
coarsenCriterion.setMinCoarsenRate(1.05);
// coarsenCriterion.setAccumulate(Dune::Amg::noAccu);
coarsenCriterion.setAccumulate(Dune::Amg::atOnceAccu);
coarsenCriterion.setSkipIsolated(false);
// instantiate the AMG preconditioner
#if HAVE_MPI
amg_ = std::make_shared(*fineOperator_, coarsenCriterion, smootherArgs, *istlComm_);
#else
amg_ = std::make_shared(*fineOperator_, coarsenCriterion, smootherArgs);
#endif
}
std::unique_ptr > convCrit_;
std::shared_ptr fineOperator_;
std::shared_ptr amg_;
#if HAVE_MPI
std::shared_ptr istlComm_;
#endif
};
} // namespace Linear
} // namespace Opm
#endif