tpfalinearizer.hh
Go to the documentation of this file.
1// -*- mode: C++; tab-width: 4; indent-tabs-mode: nil; c-basic-offset: 4 -*-
2// vi: set et ts=4 sw=4 sts=4:
3/*
4 This file is part of the Open Porous Media project (OPM).
5
6 OPM is free software: you can redistribute it and/or modify
7 it under the terms of the GNU General Public License as published by
8 the Free Software Foundation, either version 2 of the License, or
9 (at your option) any later version.
10
11 OPM is distributed in the hope that it will be useful,
12 but WITHOUT ANY WARRANTY; without even the implied warranty of
13 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
14 GNU General Public License for more details.
15
16 You should have received a copy of the GNU General Public License
17 along with OPM. If not, see <http://www.gnu.org/licenses/>.
18
19 Consult the COPYING file in the top-level source directory of this
20 module for the precise wording of the license and the list of
21 copyright holders.
22*/
28#ifndef TPFA_LINEARIZER_HH
29#define TPFA_LINEARIZER_HH
30
31#include <dune/common/version.hh>
32#include <dune/common/fvector.hh>
33#include <dune/common/fmatrix.hh>
34
35#include <opm/common/Exceptions.hpp>
36#include <opm/common/TimingMacros.hpp>
37
38#include <opm/grid/utility/SparseTable.hpp>
39
40#include <opm/material/common/ConditionalStorage.hpp>
41
42#include <opm/input/eclipse/EclipseState/Grid/FaceDir.hpp>
43#include <opm/input/eclipse/Schedule/BCProp.hpp>
44
51
52#include <cassert>
53#include <cstddef>
54#include <exception> // current_exception, rethrow_exception
55#include <iostream>
56#include <map>
57#include <memory>
58#include <numeric>
59#include <set>
60#include <stdexcept>
61#include <unordered_map>
62#include <vector>
63
64#include <fmt/format.h>
65
66#include <opm/common/utility/gpuDecorators.hpp>
67#if HAVE_CUDA
68#if USE_HIP
69#include <opm/simulators/linalg/gpuistl_hip/GpuBuffer.hpp>
70#include <opm/simulators/linalg/gpuistl_hip/GpuView.hpp>
71#include <opm/simulators/linalg/gpuistl_hip/MiniMatrix.hpp>
72#include <opm/simulators/linalg/gpuistl_hip/MiniVector.hpp>
73#else
78#endif
79#endif
80
81namespace Opm::Parameters {
82
83struct SeparateSparseSourceTerms { static constexpr bool value = false; };
84
85} // namespace Opm::Parameters
86
87namespace Opm {
88
89// forward declarations
90template<class TypeTag>
92
93// Moved these structs out of the class to make them visible in the GPU code.
94template<class Storage = std::vector<int>>
96{
97 Storage cells;
98};
99
100#if HAVE_CUDA && OPM_IS_COMPILING_WITH_GPU_COMPILER
102 {
103 if (CPUDomain.cells.size() == 0) {
104 OPM_THROW(std::runtime_error, "Cannot copy empty full domain to GPU.");
105 }
106 return FullDomain<gpuistl::GpuBuffer<int>>{
107 gpuistl::GpuBuffer<int>(CPUDomain.cells)
108 };
109 };
110
111 FullDomain<gpuistl::GpuView<int>> make_view(FullDomain<gpuistl::GpuBuffer<int>>& buffer)
112 {
113 if (buffer.cells.size() == 0) {
114 OPM_THROW(std::runtime_error, "Cannot make view of empty full domain buffer.");
115 }
116 return FullDomain<gpuistl::GpuView<int>>{
117 gpuistl::make_view(buffer.cells)
118 };
119 };
120#endif
121
122template <class ResidualNBInfoType,class BlockType>
124{
125 unsigned int neighbor;
126 ResidualNBInfoType res_nbinfo;
127 BlockType* matBlockAddress;
128
129 template <class OtherBlockType>
131 : neighbor(other.neighbor)
132 , res_nbinfo(other.res_nbinfo)
133 , matBlockAddress(nullptr)
134 {
135 if (other.matBlockAddress) {
136 matBlockAddress = reinterpret_cast<BlockType*>(other.matBlockAddress);
137 }
138 }
139
140 template <class PtrType>
141 NeighborInfoStruct(unsigned int n, const ResidualNBInfoType& r, PtrType ptr)
142 : neighbor(n)
143 , res_nbinfo(r)
144 , matBlockAddress(static_cast<BlockType*>(ptr))
145 {
146 }
147
148 // Add a default constructor
150 : neighbor(0)
151 , res_nbinfo()
152 , matBlockAddress(nullptr)
153 {
154 }
155};
156
157#if HAVE_CUDA && OPM_IS_COMPILING_WITH_GPU_COMPILER
158namespace gpuistl {
159 template<class MiniMatrixType, class MatrixBlockType, class ResidualNBInfoType>
160 auto copy_to_gpu(const SparseTable<NeighborInfoStruct<ResidualNBInfoType, MatrixBlockType>>& cpuNeighborInfoTable)
161 {
162 // Convert the DUNE FieldMatrix/MatrixBlock to MiniMatrix types
163 using StructWithMinimatrix = NeighborInfoStruct<ResidualNBInfoType, MiniMatrixType>;
164 std::vector<StructWithMinimatrix> minimatrices(cpuNeighborInfoTable.dataSize());
165 size_t idx = 0;
166 for (auto e : cpuNeighborInfoTable.dataStorage()) {
167 minimatrices[idx++] = StructWithMinimatrix(e);
168 }
169
170 return SparseTable<StructWithMinimatrix, gpuistl::GpuBuffer>(
171 gpuistl::GpuBuffer<StructWithMinimatrix>(minimatrices),
172 gpuistl::GpuBuffer<int>(cpuNeighborInfoTable.rowStarts())
173 );
174 }
175}
176#endif
177
187template<class TypeTag>
189{
197
207
208 using Element = typename GridView::template Codim<0>::Entity;
209 using ElementIterator = typename GridView::template Codim<0>::Iterator;
210
211 using Vector = GlobalEqVector;
212
213 enum { numEq = getPropValue<TypeTag, Properties::NumEq>() };
214 enum { historySize = getPropValue<TypeTag, Properties::TimeDiscHistorySize>() };
215 enum { dimWorld = GridView::dimensionworld };
216
217 using MatrixBlock = typename SparseMatrixAdapter::MatrixBlock;
218 using VectorBlock = Dune::FieldVector<Scalar, numEq>;
220
221#if HAVE_CUDA && OPM_IS_COMPILING_WITH_GPU_COMPILER
222 using MatrixBlockGPU = gpuistl::MiniMatrix<Scalar, numEq * numEq>;
223 using VectorBlockGPU = gpuistl::MiniVector<Scalar, numEq>;
224#endif
225
226 static constexpr bool linearizeNonLocalElements =
227 getPropValue<TypeTag, Properties::LinearizeNonLocalElements>();
228 static constexpr bool enableFullyImplicitThermal = (getPropValue<TypeTag, Properties::EnergyModuleType>() == EnergyModules::FullyImplicitThermal);
229 static constexpr bool enableDiffusion = getPropValue<TypeTag, Properties::EnableDiffusion>();
230 static constexpr bool enableDispersion = getPropValue<TypeTag, Properties::EnableDispersion>();
231 static const bool enableBioeffects = getPropValue<TypeTag, Properties::EnableBioeffects>();
232
233 // copying the linearizer is not a good idea
234 TpfaLinearizer(const TpfaLinearizer&) = delete;
236
237public:
239 {
240 simulatorPtr_ = nullptr;
241 separateSparseSourceTerms_ = Parameters::Get<Parameters::SeparateSparseSourceTerms>();
242 exportIndex_=-1;
243 exportCount_=-1;
244 }
245
249 static void registerParameters()
250 {
251 Parameters::Register<Parameters::SeparateSparseSourceTerms>
252 ("Treat well source terms all in one go, instead of on a cell by cell basis.");
253 }
254
264 void init(Simulator& simulator)
265 {
266 simulatorPtr_ = &simulator;
267 eraseMatrix();
268 }
269
278 {
279 jacobian_.reset();
280 }
281
292 {
295 }
296
308 {
309 int succeeded;
310 try {
311 linearizeDomain(fullDomain_);
312 succeeded = 1;
313 }
314 catch (const std::exception& e) {
315 std::cout << "rank " << simulator_().gridView().comm().rank()
316 << " caught an exception while linearizing:" << e.what()
317 << "\n" << std::flush;
318 succeeded = 0;
319 }
320 catch (...) {
321 std::cout << "rank " << simulator_().gridView().comm().rank()
322 << " caught an exception while linearizing"
323 << "\n" << std::flush;
324 succeeded = 0;
325 }
326 succeeded = simulator_().gridView().comm().min(succeeded);
327
328 if (!succeeded) {
329 throw NumericalProblem("A process did not succeed in linearizing the system");
330 }
331 }
332
346 template <class SubDomainType>
347 void linearizeDomain(const SubDomainType& domain, const bool isNlddLocalSolve = false)
348 {
349 OPM_TIMEBLOCK(linearizeDomain);
350 // we defer the initialization of the Jacobian matrix until here because the
351 // auxiliary modules usually assume the problem, model and grid to be fully
352 // initialized...
353 if (!jacobian_) {
354 initFirstIteration_();
355 }
356
357 // Called here because it is no longer called from linearize_().
358 if (isNlddLocalSolve) {
359 resetSystem_(domain);
360 }
361 else {
362 resetSystem_();
363 }
364
365 linearize_(domain, isNlddLocalSolve);
366 }
367
368 void finalize()
369 { jacobian_->finalize(); }
370
376 {
377 OPM_TIMEBLOCK(linearizeAuxilaryEquations);
378 // flush possible local caches into matrix structure
379 jacobian_->commit();
380
381 auto& model = model_();
382 const auto& comm = simulator_().gridView().comm();
383 for (unsigned auxModIdx = 0; auxModIdx < model.numAuxiliaryModules(); ++auxModIdx) {
384 bool succeeded = true;
385 try {
386 model.auxiliaryModule(auxModIdx)->linearize(*jacobian_, residual_);
387 }
388 catch (const std::exception& e) {
389 succeeded = false;
390
391 std::cout << "rank " << simulator_().gridView().comm().rank()
392 << " caught an exception while linearizing:" << e.what()
393 << "\n" << std::flush;
394 }
395
396 succeeded = comm.min(succeeded);
397
398 if (!succeeded) {
399 throw NumericalProblem("linearization of an auxiliary equation failed");
400 }
401 }
402 }
403
407 const SparseMatrixAdapter& jacobian() const
408 { return *jacobian_; }
409
410 SparseMatrixAdapter& jacobian()
411 { return *jacobian_; }
412
416 const GlobalEqVector& residual() const
417 { return residual_; }
418
419 GlobalEqVector& residual()
420 { return residual_; }
421
425 void exportSystem(const int idx, std::string& tag, const char *path="export")
426 {
427 const bool export_sparsity = exportIndex_ == -1;
428
429 // increment indices and generate tag
430 exportCount_ = exportIndex_ == idx ? ++exportCount_ : 0;
431 exportIndex_ = idx;
432 tag = fmt::format("_{:03d}_{:02d}", exportIndex_, exportCount_);
433
434 fmt::print("index = {:d}\n", exportIndex_);
435 fmt::print("count = {:d}\n", exportCount_);
436
437 Opm::exportSystem(jacobian_->istlMatrix(), residual_, export_sparsity, tag.c_str(), path);
438 }
439
441 { linearizationType_ = linearizationType; }
442
444 { return linearizationType_; }
445
451 const auto& getFlowsInfo() const
452 { return flowsInfo_; }
453
459 const auto& getFloresInfo() const
460 { return floresInfo_; }
461
467 const auto& getVelocityInfo() const
468 { return velocityInfo_; }
469
470 const auto& getNeighborInfo() const {
471 return neighborInfo_;
472 }
473
474
476 {
477 updateStoredTransmissibilities();
478 }
479
481 {
482 for (auto& bdyInfo : boundaryInfo_) {
483 const auto [type, massrateAD] = problem_().boundaryCondition(bdyInfo.cell, bdyInfo.dir);
484
485 // Strip the unnecessary (and zero anyway) derivatives off massrate.
486 VectorBlock massrate(0.0);
487 for (std::size_t ii = 0; ii < massrate.size(); ++ii) {
488 massrate[ii] = massrateAD[ii].value();
489 }
490 if (type != BCType::NONE) {
491 const auto& exFluidState = problem_().boundaryFluidState(bdyInfo.cell, bdyInfo.dir);
492 bdyInfo.bcdata.type = type;
493 bdyInfo.bcdata.massRate = massrate;
494 bdyInfo.bcdata.exFluidState = exFluidState;
495 }
496 }
497 }
498
504 std::map<unsigned, Constraints> constraintsMap() const
505 { return {}; }
506
507 template <class SubDomainType>
508 void resetSystem_(const SubDomainType& domain)
509 {
510 if (!jacobian_) {
511 initFirstIteration_();
512 }
513 for (int globI : domain.cells) {
514 residual_[globI] = 0.0;
515 jacobian_->clearRow(globI, 0.0);
516 }
517 }
518
519private:
520 Simulator& simulator_()
521 { return *simulatorPtr_; }
522
523 const Simulator& simulator_() const
524 { return *simulatorPtr_; }
525
526 Problem& problem_()
527 { return simulator_().problem(); }
528
529 const Problem& problem_() const
530 { return simulator_().problem(); }
531
532 Model& model_()
533 { return simulator_().model(); }
534
535 const Model& model_() const
536 { return simulator_().model(); }
537
538 const GridView& gridView_() const
539 { return problem_().gridView(); }
540
541 void initFirstIteration_()
542 {
543 // initialize the BCRS matrix for the Jacobian of the residual function
544 createMatrix_();
545
546 // initialize the Jacobian matrix and the vector for the residual function
547 residual_.resize(model_().numTotalDof());
548 resetSystem_();
549
550 // initialize the sparse tables for Flows and Flores
551 createFlows_();
552 }
553
554 // Construct the BCRS matrix for the Jacobian of the residual function
555 void createMatrix_()
556 {
557 OPM_TIMEBLOCK(createMatrix);
558 if (!neighborInfo_.empty()) {
559 // It is ok to call this function multiple times, but it
560 // should not do anything if already called.
561 return;
562 }
563 const auto& model = model_();
564 Stencil stencil(gridView_(), model_().dofMapper());
565
566 // for the main model, find out the global indices of the neighboring degrees of
567 // freedom of each primary degree of freedom
568 using NeighborSet = std::set<unsigned>;
569 std::vector<NeighborSet> sparsityPattern(model.numTotalDof());
570 const Scalar gravity = problem_().gravity()[dimWorld - 1];
571 unsigned numCells = model.numTotalDof();
572 neighborInfo_.reserve(numCells, 6 * numCells);
573 std::vector<NeighborInfoCPU> loc_nbinfo;
574 for (const auto& elem : elements(gridView_())) {
575 stencil.update(elem);
576
577 for (unsigned primaryDofIdx = 0; primaryDofIdx < stencil.numPrimaryDof(); ++primaryDofIdx) {
578 const unsigned myIdx = stencil.globalSpaceIndex(primaryDofIdx);
579 loc_nbinfo.resize(stencil.numDof() - 1); // Do not include the primary dof in neighborInfo_
580
581 for (unsigned dofIdx = 0; dofIdx < stencil.numDof(); ++dofIdx) {
582 const unsigned neighborIdx = stencil.globalSpaceIndex(dofIdx);
583 sparsityPattern[myIdx].insert(neighborIdx);
584 if (dofIdx > 0) {
585 const Scalar trans = problem_().transmissibility(myIdx, neighborIdx);
586 const auto scvfIdx = dofIdx - 1;
587 const auto& scvf = stencil.interiorFace(scvfIdx);
588 const Scalar area = scvf.area();
589 const Scalar Vin = problem_().model().dofTotalVolume(myIdx);
590 const Scalar Vex = problem_().model().dofTotalVolume(neighborIdx);
591 const Scalar zIn = problem_().dofCenterDepth(myIdx);
592 const Scalar zEx = problem_().dofCenterDepth(neighborIdx);
593 const Scalar dZg = (zIn - zEx)*gravity;
594 const Scalar thpres = problem_().thresholdPressure(myIdx, neighborIdx);
595 const auto dirId = scvf.dirId();
596 auto faceDir = dirId < 0 ? FaceDir::DirEnum::Unknown
597 : FaceDir::FromIntersectionIndex(dirId);
598 ResidualNBInfo nbinfo{trans, area, thpres, dZg, faceDir, Vin, Vex, {}, {}, {}, {}};
599 if constexpr (enableFullyImplicitThermal) {
600 nbinfo.inAlpha = problem_().thermalHalfTransmissibility(myIdx, neighborIdx);
601 nbinfo.outAlpha = problem_().thermalHalfTransmissibility(neighborIdx, myIdx);
602 }
603 if constexpr (enableDiffusion) {
604 nbinfo.diffusivity = problem_().diffusivity(myIdx, neighborIdx);
605 }
606 if constexpr (enableDispersion) {
607 nbinfo.dispersivity = problem_().dispersivity(myIdx, neighborIdx);
608 }
609 loc_nbinfo[dofIdx - 1] = NeighborInfoCPU{neighborIdx, nbinfo, nullptr};
610 }
611 }
612 neighborInfo_.appendRow(loc_nbinfo.begin(), loc_nbinfo.end());
613 if (problem_().nonTrivialBoundaryConditions()) {
614 for (unsigned bfIndex = 0; bfIndex < stencil.numBoundaryFaces(); ++bfIndex) {
615 const auto& bf = stencil.boundaryFace(bfIndex);
616 const int dir_id = bf.dirId();
617 // not for NNCs
618 if (dir_id < 0) {
619 continue;
620 }
621 const auto [type, massrateAD] = problem_().boundaryCondition(myIdx, dir_id);
622 // Strip the unnecessary (and zero anyway) derivatives off massrate.
623 VectorBlock massrate(0.0);
624 for (std::size_t ii = 0; ii < massrate.size(); ++ii) {
625 massrate[ii] = massrateAD[ii].value();
626 }
627 const auto& exFluidState = problem_().boundaryFluidState(myIdx, dir_id);
628 BoundaryConditionData bcdata{type,
629 massrate,
630 exFluidState.pvtRegionIndex(),
631 bfIndex,
632 bf.area(),
633 bf.integrationPos()[dimWorld - 1],
634 exFluidState};
635 boundaryInfo_.push_back({myIdx, dir_id, bfIndex, bcdata});
636 }
637 }
638 }
639 }
640
641 // add the additional neighbors and degrees of freedom caused by the auxiliary
642 // equations
643 const std::size_t numAuxMod = model.numAuxiliaryModules();
644 for (unsigned auxModIdx = 0; auxModIdx < numAuxMod; ++auxModIdx) {
645 model.auxiliaryModule(auxModIdx)->addNeighbors(sparsityPattern);
646 }
647
648 // allocate raw matrix
649 jacobian_ = std::make_unique<SparseMatrixAdapter>(simulator_());
650 diagMatAddress_.resize(numCells);
651 // create matrix structure based on sparsity pattern
652 jacobian_->reserve(sparsityPattern);
653 for (unsigned globI = 0; globI < numCells; globI++) {
654 const auto& nbInfos = neighborInfo_[globI];
655 diagMatAddress_[globI] = jacobian_->blockAddress(globI, globI);
656 for (auto& nbInfo : nbInfos) {
657 nbInfo.matBlockAddress = jacobian_->blockAddress(nbInfo.neighbor, globI);
658 }
659 }
660
661 // Create dummy full domain.
662 fullDomain_.cells.resize(numCells);
663 std::iota(fullDomain_.cells.begin(), fullDomain_.cells.end(), 0);
664 }
665
666 // reset the global linear system of equations.
667 void resetSystem_()
668 {
669 residual_ = 0.0;
670 // zero all matrix entries
671 jacobian_->clear();
672 }
673
674 // Initialize the flows, flores, and velocity sparse tables
675 void createFlows_()
676 {
677 OPM_TIMEBLOCK(createFlows);
678 // If FLOWS/FLORES is set in any RPTRST in the schedule, then we initializate the sparse tables
679 // For now, do the same also if any block flows are requested (TODO: only save requested cells...)
680 // If DISPERC is in the deck, we initialize the sparse table here as well.
681 const bool anyFlows = simulator_().problem().eclWriter().outputModule().getFlows().anyFlows() ||
682 simulator_().problem().eclWriter().outputModule().getFlows().hasBlockFlows();
683 const bool anyFlores = simulator_().problem().eclWriter().outputModule().getFlows().anyFlores();
684 const bool dispersionActive = simulator_().vanguard().eclState().getSimulationConfig().rock_config().dispersion();
685 if (((!anyFlows || !flowsInfo_.empty()) && (!anyFlores || !floresInfo_.empty())) && (!dispersionActive && !enableBioeffects)) {
686 return;
687 }
688 const auto& model = model_();
689 const auto& nncOutput = simulator_().problem().eclWriter().getOutputNnc();
690 Stencil stencil(gridView_(), model_().dofMapper());
691 const unsigned numCells = model.numTotalDof();
692 std::unordered_multimap<int, std::pair<int, int>> nncIndices;
693 std::vector<FlowInfo> loc_flinfo;
694 std::vector<VelocityInfo> loc_vlinfo;
695 unsigned int nncId = 0;
696 VectorBlock flow(0.0);
697
698 // Create a nnc structure to use fast lookup
699 for (unsigned nncIdx = 0; nncIdx < nncOutput.size(); ++nncIdx) {
700 const int ci1 = nncOutput[nncIdx].cell1;
701 const int ci2 = nncOutput[nncIdx].cell2;
702 nncIndices.emplace(ci1, std::make_pair(ci2, nncIdx));
703 }
704
705 if (anyFlows) {
706 flowsInfo_.reserve(numCells, 6 * numCells);
707 }
708 if (anyFlores) {
709 floresInfo_.reserve(numCells, 6 * numCells);
710 }
711 if (dispersionActive || enableBioeffects) {
712 velocityInfo_.reserve(numCells, 6 * numCells);
713 }
714
715 for (const auto& elem : elements(gridView_())) {
716 stencil.update(elem);
717 for (unsigned primaryDofIdx = 0; primaryDofIdx < stencil.numPrimaryDof(); ++primaryDofIdx) {
718 const unsigned myIdx = stencil.globalSpaceIndex(primaryDofIdx);
719 const int numFaces = stencil.numBoundaryFaces() + stencil.numInteriorFaces();
720 loc_flinfo.resize(numFaces);
721 loc_vlinfo.resize(stencil.numDof() - 1);
722
723 for (unsigned dofIdx = 0; dofIdx < stencil.numDof(); ++dofIdx) {
724 const unsigned neighborIdx = stencil.globalSpaceIndex(dofIdx);
725 if (dofIdx > 0) {
726 const auto scvfIdx = dofIdx - 1;
727 const auto& scvf = stencil.interiorFace(scvfIdx);
728 int faceId = scvf.dirId();
729 const int cartMyIdx = simulator_().vanguard().cartesianIndex(myIdx);
730 const int cartNeighborIdx = simulator_().vanguard().cartesianIndex(neighborIdx);
731 const auto& range = nncIndices.equal_range(cartMyIdx);
732 for (auto it = range.first; it != range.second; ++it) {
733 if (it->second.first == cartNeighborIdx){
734 // -1 gives problem since is used for the nncInput from the deck
735 faceId = -2;
736 // the index is stored to be used for writting the outputs
737 nncId = it->second.second;
738 }
739 }
740 loc_flinfo[dofIdx - 1] = FlowInfo{faceId, flow, nncId};
741 loc_vlinfo[dofIdx - 1] = VelocityInfo{flow};
742 }
743 }
744
745 for (unsigned bdfIdx = 0; bdfIdx < stencil.numBoundaryFaces(); ++bdfIdx) {
746 const auto& scvf = stencil.boundaryFace(bdfIdx);
747 const int faceId = scvf.dirId();
748 loc_flinfo[stencil.numInteriorFaces() + bdfIdx] = FlowInfo{faceId, flow, nncId};
749 }
750
751 if (anyFlows) {
752 flowsInfo_.appendRow(loc_flinfo.begin(), loc_flinfo.end());
753 }
754 if (anyFlores) {
755 floresInfo_.appendRow(loc_flinfo.begin(), loc_flinfo.end());
756 }
757 if (dispersionActive || enableBioeffects) {
758 velocityInfo_.appendRow(loc_vlinfo.begin(), loc_vlinfo.end());
759 }
760 }
761 }
762 }
763
764public:
765 void setResAndJacobi(VectorBlock& res, MatrixBlock& bMat, const ADVectorBlock& resid) const
766 {
767 for (unsigned eqIdx = 0; eqIdx < numEq; ++eqIdx) {
768 res[eqIdx] = resid[eqIdx].value();
769 }
770
771 for (unsigned eqIdx = 0; eqIdx < numEq; ++eqIdx) {
772 for (unsigned pvIdx = 0; pvIdx < numEq; ++pvIdx) {
773 // A[dofIdx][focusDofIdx][eqIdx][pvIdx] is the partial derivative of
774 // the residual function 'eqIdx' for the degree of freedom 'dofIdx'
775 // with regard to the focus variable 'pvIdx' of the degree of freedom
776 // 'focusDofIdx'
777 bMat[eqIdx][pvIdx] = resid[eqIdx].derivative(pvIdx);
778 }
779 }
780 }
781
783 {
784 OPM_TIMEBLOCK(updateFlows);
785 const bool enableFlows = simulator_().problem().eclWriter().outputModule().getFlows().hasFlows() ||
786 simulator_().problem().eclWriter().outputModule().getFlows().hasBlockFlows();
787 const bool enableFlores = simulator_().problem().eclWriter().outputModule().getFlows().hasFlores();
788 if (!enableFlows && !enableFlores) {
789 return;
790 }
791 const unsigned int numCells = model_().numTotalDof();
792#ifdef _OPENMP
793#pragma omp parallel for
794#endif
795 for (unsigned globI = 0; globI < numCells; ++globI) {
796 OPM_TIMEBLOCK_LOCAL(linearizationForEachCell, Subsystem::Assembly);
797 const auto& nbInfos = neighborInfo_[globI];
798 ADVectorBlock adres(0.0);
799 ADVectorBlock darcyFlux(0.0);
800 const IntensiveQuantities& intQuantsIn = model_().intensiveQuantities(globI, /*timeIdx*/ 0);
801 // Flux term.
802 {
803 OPM_TIMEBLOCK_LOCAL(fluxCalculationForEachCell, Subsystem::Assembly);
804 short loc = 0;
805 for (const auto& nbInfo : nbInfos) {
806 OPM_TIMEBLOCK_LOCAL(fluxCalculationForEachFace, Subsystem::Assembly);
807 const unsigned globJ = nbInfo.neighbor;
808 assert(globJ != globI);
809 adres = 0.0;
810 darcyFlux = 0.0;
811 const IntensiveQuantities& intQuantsEx = model_().intensiveQuantities(globJ, /*timeIdx*/ 0);
812 LocalResidual::computeFlux(adres, darcyFlux, globI, globJ, intQuantsIn,
813 intQuantsEx, nbInfo.res_nbinfo, problem_().moduleParams());
814 adres *= nbInfo.res_nbinfo.faceArea;
815 if (enableFlows) {
816 for (unsigned eqIdx = 0; eqIdx < numEq; ++eqIdx) {
817 flowsInfo_[globI][loc].flow[eqIdx] = adres[eqIdx].value();
818 }
819 }
820 if (enableFlores) {
821 for (unsigned eqIdx = 0; eqIdx < numEq; ++eqIdx) {
822 floresInfo_[globI][loc].flow[eqIdx] = darcyFlux[eqIdx].value();
823 }
824 }
825 ++loc;
826 }
827 }
828 }
829
830 // Boundary terms. Only looping over cells with nontrivial bcs.
831 for (const auto& bdyInfo : boundaryInfo_) {
832 if (bdyInfo.bcdata.type == BCType::NONE) {
833 continue;
834 }
835
836 ADVectorBlock adres(0.0);
837 const unsigned globI = bdyInfo.cell;
838 const auto& nbInfos = neighborInfo_[globI];
839 const IntensiveQuantities& insideIntQuants = model_().intensiveQuantities(globI, /*timeIdx*/ 0);
840 LocalResidual::computeBoundaryFlux(adres, problem_(), bdyInfo.bcdata, insideIntQuants, globI);
841 adres *= bdyInfo.bcdata.faceArea;
842 const unsigned bfIndex = bdyInfo.bfIndex;
843 if (enableFlows) {
844 for (unsigned eqIdx = 0; eqIdx < numEq; ++eqIdx) {
845 flowsInfo_[globI][nbInfos.size() + bfIndex].flow[eqIdx] = adres[eqIdx].value();
846 }
847 }
848 // TODO also store Flores?
849 }
850 }
851
852private:
853 template <class SubDomainType>
854 void linearize_(const SubDomainType& domain, bool isNlddLocalSolve)
855 {
856 // This check should be removed once this is addressed by
857 // for example storing the previous timesteps' values for
858 // rsmax (for DRSDT) and similar.
859 if (!problem_().recycleFirstIterationStorage()) {
860 if (!model_().storeIntensiveQuantities() && !model_().enableStorageCache()) {
861 OPM_THROW(std::runtime_error, "Must have cached either IQs or storage when we cannot recycle.");
862 }
863 }
864
865 OPM_TIMEBLOCK(linearize);
866
867 // We do not call resetSystem_() here, since that will set
868 // the full system to zero, not just our part.
869 // Instead, that must be called before starting the linearization.
870 const bool dispersionActive = simulator_().vanguard().eclState().getSimulationConfig().rock_config().dispersion();
871 const unsigned int numCells = domain.cells.size();
872
873 // Fetch timestepsize used later in accumulation term.
874 const double dt = simulator_().timeStepSize();
875
876#ifdef _OPENMP
877#pragma omp parallel for
878#endif
879 for (unsigned ii = 0; ii < numCells; ++ii) {
880 OPM_TIMEBLOCK_LOCAL(linearizationForEachCell, Subsystem::Assembly);
881 const unsigned globI = domain.cells[ii];
882 const auto& nbInfos = neighborInfo_[globI];
883 VectorBlock res(0.0);
884 MatrixBlock bMat(0.0);
885 ADVectorBlock adres(0.0);
886 ADVectorBlock darcyFlux(0.0);
887 const IntensiveQuantities& intQuantsIn = model_().intensiveQuantities(globI, /*timeIdx*/ 0);
888
889 // Flux term.
890 {
891 OPM_TIMEBLOCK_LOCAL(fluxCalculationForEachCell, Subsystem::Assembly);
892 short loc = 0;
893 for (const auto& nbInfo : nbInfos) {
894 OPM_TIMEBLOCK_LOCAL(fluxCalculationForEachFace, Subsystem::Assembly);
895 const unsigned globJ = nbInfo.neighbor;
896 assert(globJ != globI);
897 res = 0.0;
898 bMat = 0.0;
899 adres = 0.0;
900 darcyFlux = 0.0;
901 const IntensiveQuantities& intQuantsEx = model_().intensiveQuantities(globJ, /*timeIdx*/ 0);
902 LocalResidual::computeFlux(adres, darcyFlux, globI, globJ, intQuantsIn, intQuantsEx,
903 nbInfo.res_nbinfo, problem_().moduleParams());
904 adres *= nbInfo.res_nbinfo.faceArea;
905 if (dispersionActive || enableBioeffects) {
906 for (unsigned phaseIdx = 0; phaseIdx < numEq; ++phaseIdx) {
907 velocityInfo_[globI][loc].velocity[phaseIdx] =
908 darcyFlux[phaseIdx].value() / nbInfo.res_nbinfo.faceArea;
909 }
910 }
911 setResAndJacobi(res, bMat, adres);
912 residual_[globI] += res;
913 //SparseAdapter syntax: jacobian_->addToBlock(globI, globI, bMat);
914 *diagMatAddress_[globI] += bMat;
915 bMat *= -1.0;
916 //SparseAdapter syntax: jacobian_->addToBlock(globJ, globI, bMat);
917 *nbInfo.matBlockAddress += bMat;
918 ++loc;
919 }
920 }
921
922 // Accumulation term.
923 const double volume = model_().dofTotalVolume(globI);
924 const Scalar storefac = volume / dt;
925 adres = 0.0;
926 {
927 OPM_TIMEBLOCK_LOCAL(computeStorage, Subsystem::Assembly);
928 LocalResidual::computeStorage(adres, intQuantsIn);
929 }
930 setResAndJacobi(res, bMat, adres);
931 // Either use cached storage term, or compute it on the fly.
932 if (model_().enableStorageCache()) {
933 // The cached storage for timeIdx 0 (current time) is not
934 // used, but after storage cache is shifted at the end of the
935 // timestep, it will become cached storage for timeIdx 1.
936 model_().updateCachedStorage(globI, /*timeIdx=*/0, res);
937 // We should not update the storage cache here for NLDD local solves.
938 // This will reset the start-of-step storage to incorrect numbers when
939 // we do local solves, where the iteration number will start from 0,
940 // but the starting state may not be identical to the start-of-step state.
941 // Note that a full assembly must be done before local solves
942 // otherwise this will be left un-updated.
943 if (model_().newtonMethod().numIterations() == 0 && !isNlddLocalSolve) {
944 // Need to update the storage cache.
945 if (problem_().recycleFirstIterationStorage()) {
946 // Assumes nothing have changed in the system which
947 // affects masses calculated from primary variables.
948 model_().updateCachedStorage(globI, /*timeIdx=*/1, res);
949 }
950 else {
951 Dune::FieldVector<Scalar, numEq> tmp;
952 const IntensiveQuantities intQuantOld = model_().intensiveQuantities(globI, 1);
953 LocalResidual::computeStorage(tmp, intQuantOld);
954 model_().updateCachedStorage(globI, /*timeIdx=*/1, tmp);
955 }
956 }
957 res -= model_().cachedStorage(globI, 1);
958 }
959 else {
960 OPM_TIMEBLOCK_LOCAL(computeStorage0, Subsystem::Assembly);
961 Dune::FieldVector<Scalar, numEq> tmp;
962 const IntensiveQuantities intQuantOld = model_().intensiveQuantities(globI, 1);
963 LocalResidual::computeStorage(tmp, intQuantOld);
964 // assume volume do not change
965 res -= tmp;
966 }
967 res *= storefac;
968 bMat *= storefac;
969 residual_[globI] += res;
970 //SparseAdapter syntax: jacobian_->addToBlock(globI, globI, bMat);
971 *diagMatAddress_[globI] += bMat;
972
973 // Cell-wise source terms.
974 // This will include well sources if SeparateSparseSourceTerms is false.
975 res = 0.0;
976 bMat = 0.0;
977 adres = 0.0;
978 if (separateSparseSourceTerms_) {
979 LocalResidual::computeSourceDense(adres, problem_(), intQuantsIn, globI, 0);
980 }
981 else {
982 LocalResidual::computeSource(adres, problem_(), intQuantsIn, globI, 0);
983 }
984 adres *= -volume;
985 setResAndJacobi(res, bMat, adres);
986 residual_[globI] += res;
987 //SparseAdapter syntax: jacobian_->addToBlock(globI, globI, bMat);
988 *diagMatAddress_[globI] += bMat;
989 } // end of loop for cell globI.
990
991 // Add sparse source terms. For now only wells.
992 if (separateSparseSourceTerms_) {
993 problem_().wellModel().addReservoirSourceTerms(residual_, diagMatAddress_);
994 }
995
996 // Boundary terms. Only looping over cells with nontrivial bcs.
997 for (const auto& bdyInfo : boundaryInfo_) {
998 if (bdyInfo.bcdata.type == BCType::NONE) {
999 continue;
1000 }
1001
1002 VectorBlock res(0.0);
1003 MatrixBlock bMat(0.0);
1004 ADVectorBlock adres(0.0);
1005 const unsigned globI = bdyInfo.cell;
1006 const IntensiveQuantities& insideIntQuants = model_().intensiveQuantities(globI, /*timeIdx*/ 0);
1007 LocalResidual::computeBoundaryFlux(adres, problem_(), bdyInfo.bcdata, insideIntQuants, globI);
1008 adres *= bdyInfo.bcdata.faceArea;
1009 setResAndJacobi(res, bMat, adres);
1010 residual_[globI] += res;
1012 *diagMatAddress_[globI] += bMat;
1013 }
1014 }
1015
1016 void updateStoredTransmissibilities()
1017 {
1018 if (neighborInfo_.empty()) {
1019 // This function was called before createMatrix_() was called.
1020 // We call initFirstIteration_(), not createMatrix_(), because
1021 // that will also initialize the residual consistently.
1022 initFirstIteration_();
1023 }
1024
1025 const unsigned numCells = model_().numTotalDof();
1026#ifdef _OPENMP
1027#pragma omp parallel for
1028#endif
1029 for (unsigned globI = 0; globI < numCells; globI++) {
1030 auto nbInfos = neighborInfo_[globI]; // nbInfos will be a SparseTable<...>::mutable_iterator_range.
1031 for (auto& nbInfo : nbInfos) {
1032 const unsigned globJ = nbInfo.neighbor;
1033 nbInfo.res_nbinfo.trans = problem_().transmissibility(globI, globJ);
1034 }
1035 }
1036 }
1037
1038 Simulator* simulatorPtr_{};
1039
1040 // the jacobian matrix
1041 std::unique_ptr<SparseMatrixAdapter> jacobian_{};
1042
1043 // the right-hand side
1044 GlobalEqVector residual_;
1045
1046 LinearizationType linearizationType_{};
1047
1048 using ResidualNBInfo = typename LocalResidual::ResidualNBInfo;
1049 using NeighborInfoCPU = NeighborInfoStruct<ResidualNBInfo, MatrixBlock>;
1050
1051 SparseTable<NeighborInfoCPU> neighborInfo_{};
1052 std::vector<MatrixBlock*> diagMatAddress_{};
1053
1054 struct FlowInfo
1055 {
1056 int faceId;
1057 VectorBlock flow;
1058 unsigned int nncId;
1059 };
1060 SparseTable<FlowInfo> flowsInfo_;
1061 SparseTable<FlowInfo> floresInfo_;
1062
1063 struct VelocityInfo
1064 {
1065 VectorBlock velocity;
1066 };
1067 SparseTable<VelocityInfo> velocityInfo_;
1068
1069 using ScalarFluidState = typename IntensiveQuantities::ScalarFluidState;
1070 struct BoundaryConditionData
1071 {
1072 BCType type;
1073 VectorBlock massRate;
1074 unsigned pvtRegionIdx;
1075 unsigned boundaryFaceIndex;
1076 double faceArea;
1077 double faceZCoord;
1078 ScalarFluidState exFluidState;
1079 };
1080
1081 struct BoundaryInfo
1082 {
1083 unsigned int cell;
1084 int dir;
1085 unsigned int bfIndex;
1086 BoundaryConditionData bcdata;
1087 };
1088 std::vector<BoundaryInfo> boundaryInfo_;
1089
1090 bool separateSparseSourceTerms_ = false;
1091
1092 FullDomain<> fullDomain_;
1093
1094 int exportIndex_;
1095 int exportCount_;
1096};
1097} // namespace Opm
1098
1099#endif // TPFA_LINEARIZER
Declares the properties required by the black oil model.
The base class for the element-centered finite-volume discretization scheme.
Definition: ecfvdiscretization.hh:160
Definition: matrixblock.hh:229
Manages the initializing and running of time dependent problems.
Definition: simulator.hh:84
Scalar timeStepSize() const
Returns the time step length so that we don't miss the beginning of the next episode or cross the en...
Definition: simulator.hh:413
Vanguard & vanguard()
Return a reference to the grid manager of simulation.
Definition: simulator.hh:234
Problem & problem()
Return the object which specifies the pysical setup of the simulation.
Definition: simulator.hh:265
const GridView & gridView() const
Return the grid view for which the simulation is done.
Definition: simulator.hh:246
Model & model()
Return the physical model used in the simulation.
Definition: simulator.hh:252
The common code for the linearizers of non-linear systems of equations.
Definition: tpfalinearizer.hh:189
const auto & getFloresInfo() const
Return constant reference to the floresInfo.
Definition: tpfalinearizer.hh:459
const LinearizationType & getLinearizationType() const
Definition: tpfalinearizer.hh:443
const auto & getNeighborInfo() const
Definition: tpfalinearizer.hh:470
void updateBoundaryConditionData()
Definition: tpfalinearizer.hh:480
void linearize()
Linearize the full system of non-linear equations.
Definition: tpfalinearizer.hh:291
SparseMatrixAdapter & jacobian()
Definition: tpfalinearizer.hh:410
const auto & getFlowsInfo() const
Return constant reference to the flowsInfo.
Definition: tpfalinearizer.hh:451
std::map< unsigned, Constraints > constraintsMap() const
Returns the map of constraint degrees of freedom.
Definition: tpfalinearizer.hh:504
TpfaLinearizer()
Definition: tpfalinearizer.hh:238
void finalize()
Definition: tpfalinearizer.hh:368
void init(Simulator &simulator)
Initialize the linearizer.
Definition: tpfalinearizer.hh:264
void updateFlowsInfo()
Definition: tpfalinearizer.hh:782
void setResAndJacobi(VectorBlock &res, MatrixBlock &bMat, const ADVectorBlock &resid) const
Definition: tpfalinearizer.hh:765
static void registerParameters()
Register all run-time parameters for the Jacobian linearizer.
Definition: tpfalinearizer.hh:249
void exportSystem(const int idx, std::string &tag, const char *path="export")
Export block sparse linear system.
Definition: tpfalinearizer.hh:425
void linearizeDomain()
Linearize the part of the non-linear system of equations that is associated with the spatial domain.
Definition: tpfalinearizer.hh:307
const SparseMatrixAdapter & jacobian() const
Return constant reference to global Jacobian matrix backend.
Definition: tpfalinearizer.hh:407
void setLinearizationType(LinearizationType linearizationType)
Definition: tpfalinearizer.hh:440
void linearizeDomain(const SubDomainType &domain, const bool isNlddLocalSolve=false)
Linearize the part of the non-linear system of equations that is associated with a part of the spatia...
Definition: tpfalinearizer.hh:347
GlobalEqVector & residual()
Definition: tpfalinearizer.hh:419
void eraseMatrix()
Causes the Jacobian matrix to be recreated from scratch before the next iteration.
Definition: tpfalinearizer.hh:277
const auto & getVelocityInfo() const
Return constant reference to the velocityInfo.
Definition: tpfalinearizer.hh:467
void updateDiscretizationParameters()
Definition: tpfalinearizer.hh:475
void resetSystem_(const SubDomainType &domain)
Definition: tpfalinearizer.hh:508
void linearizeAuxiliaryEquations()
Linearize the part of the non-linear system of equations that is associated with the spatial domain.
Definition: tpfalinearizer.hh:375
const GlobalEqVector & residual() const
Return constant reference to global residual vector.
Definition: tpfalinearizer.hh:416
A small fixed-size square matrix class for use in CUDA kernels.
Definition: MiniMatrix.hpp:37
Definition: MiniVector.hpp:52
Declare the properties used by the infrastructure code of the finite volume discretizations.
Defines the common properties required by the porous medium multi-phase models.
@ NONE
Definition: DeferredLogger.hpp:46
Definition: blackoilnewtonmethodparams.hpp:31
HYPRE_IJMatrix createMatrix(HYPRE_Int N, HYPRE_Int dof_offset, const CommType &comm)
Create Hypre matrix.
Definition: HypreSetup.hpp:170
PointerView< T > make_view(const std::shared_ptr< T > &ptr)
Definition: gpu_smart_pointer.hpp:318
Definition: blackoilbioeffectsmodules.hh:43
typename Properties::Detail::GetPropImpl< TypeTag, Property >::type::type GetPropType
get the type alias defined in the property (equivalent to old macro GET_PROP_TYPE(....
Definition: propertysystem.hh:233
void exportSystem(const IstlMatrix &jacobian, const GlobalEqVector &residual, const bool export_sparsity, const char *tag, const char *path="export")
Export blocks-sparse linear system.
Definition: exportSystem.hpp:42
Definition: tpfalinearizer.hh:96
Storage cells
Definition: tpfalinearizer.hh:97
Definition: linearizationtype.hh:34
Definition: tpfalinearizer.hh:124
NeighborInfoStruct(unsigned int n, const ResidualNBInfoType &r, PtrType ptr)
Definition: tpfalinearizer.hh:141
NeighborInfoStruct()
Definition: tpfalinearizer.hh:149
NeighborInfoStruct(const NeighborInfoStruct< ResidualNBInfoType, OtherBlockType > &other)
Definition: tpfalinearizer.hh:130
BlockType * matBlockAddress
Definition: tpfalinearizer.hh:127
ResidualNBInfoType res_nbinfo
Definition: tpfalinearizer.hh:126
unsigned int neighbor
Definition: tpfalinearizer.hh:125
Definition: tpfalinearizer.hh:83
static constexpr bool value
Definition: tpfalinearizer.hh:83