https://mooseframework.inl.gov
NonlinearSystemBase.C
Go to the documentation of this file.
1 //* This file is part of the MOOSE framework
2 //* https://mooseframework.inl.gov
3 //*
4 //* All rights reserved, see COPYRIGHT for full restrictions
5 //* https://github.com/idaholab/moose/blob/master/COPYRIGHT
6 //*
7 //* Licensed under LGPL 2.1, please see LICENSE for details
8 //* https://www.gnu.org/licenses/lgpl-2.1.html
9 
10 #include "NonlinearSystemBase.h"
11 #include "AuxiliarySystem.h"
12 #include "Problem.h"
13 #include "FEProblem.h"
14 #include "MooseVariableFE.h"
15 #include "MooseVariableScalar.h"
16 #include "PetscSupport.h"
17 #include "Factory.h"
18 #include "ParallelUniqueId.h"
19 #include "ThreadedElementLoop.h"
20 #include "MaterialData.h"
21 #include "ComputeResidualThread.h"
23 #include "ComputeFVFluxThread.h"
24 #include "ComputeJacobianThread.h"
28 #include "ComputeDiracThread.h"
35 #include "TimeKernel.h"
36 #include "BoundaryCondition.h"
37 #include "DirichletBCBase.h"
38 #include "NodalBCBase.h"
39 #include "IntegratedBCBase.h"
40 #include "DGKernel.h"
41 #include "InterfaceKernelBase.h"
42 #include "ElementDamper.h"
43 #include "NodalDamper.h"
44 #include "GeneralDamper.h"
45 #include "DisplacedProblem.h"
46 #include "NearestNodeLocator.h"
47 #include "PenetrationLocator.h"
48 #include "NodalConstraint.h"
49 #include "NodeFaceConstraint.h"
50 #include "NodeElemConstraintBase.h"
51 #include "MortarConstraint.h"
52 #include "ElemElemConstraint.h"
53 #include "ScalarKernelBase.h"
54 #include "Parser.h"
55 #include "Split.h"
57 #include "MooseMesh.h"
58 #include "MooseUtils.h"
59 #include "MooseApp.h"
60 #include "NodalKernelBase.h"
61 #include "DiracKernelBase.h"
62 #include "TimeIntegrator.h"
63 #include "Predictor.h"
64 #include "Assembly.h"
65 #include "ElementPairLocator.h"
66 #include "ODETimeKernel.h"
69 #include "ADKernel.h"
70 #include "ADDirichletBCBase.h"
71 #include "Moose.h"
72 #include "ConsoleStream.h"
73 #include "MooseError.h"
74 #include "FVElementalKernel.h"
77 #include "FVFluxKernel.h"
79 #include "UserObject.h"
81 #include "HDGKernel.h"
83 
84 // libMesh
85 #include "libmesh/nonlinear_solver.h"
86 #include "libmesh/quadrature_gauss.h"
87 #include "libmesh/dense_vector.h"
88 #include "libmesh/boundary_info.h"
89 #include "libmesh/petsc_matrix.h"
90 #include "libmesh/petsc_vector.h"
91 #include "libmesh/petsc_nonlinear_solver.h"
92 #include "libmesh/numeric_vector.h"
93 #include "libmesh/mesh.h"
94 #include "libmesh/dense_subvector.h"
95 #include "libmesh/dense_submatrix.h"
96 #include "libmesh/dof_map.h"
97 #include "libmesh/sparse_matrix.h"
98 #include "libmesh/petsc_matrix.h"
99 #include "libmesh/default_coupling.h"
100 #include "libmesh/diagonal_matrix.h"
101 #include "libmesh/fe_interface.h"
102 #include "libmesh/petsc_solver_exception.h"
103 
104 #include <ios>
105 
106 #include "petscsnes.h"
107 #include <PetscDMMoose.h>
108 EXTERN_C_BEGIN
109 extern PetscErrorCode DMCreate_Moose(DM);
110 EXTERN_C_END
111 
112 using namespace libMesh;
113 
115  System & sys,
116  const std::string & name)
117  : SolverSystem(fe_problem, fe_problem, name, Moose::VAR_SOLVER),
118  PerfGraphInterface(fe_problem.getMooseApp().perfGraph(), "NonlinearSystemBase"),
119  _sys(sys),
120  _last_nl_rnorm(0.),
121  _current_nl_its(0),
122  _residual_ghosted(NULL),
123  _Re_time_tag(-1),
124  _Re_time(NULL),
125  _Re_non_time_tag(-1),
126  _Re_non_time(NULL),
127  _scalar_kernels(/*threaded=*/false),
128  _nodal_bcs(/*threaded=*/false),
129  _preset_nodal_bcs(/*threaded=*/false),
130  _ad_preset_nodal_bcs(/*threaded=*/false),
131 #ifdef MOOSE_KOKKOS_ENABLED
132  _kokkos_kernels(/*threaded=*/false),
133  _kokkos_integrated_bcs(/*threaded=*/false),
134  _kokkos_nodal_bcs(/*threaded=*/false),
135  _kokkos_preset_nodal_bcs(/*threaded=*/false),
136  _kokkos_nodal_kernels(/*threaded=*/false),
137 #endif
138  _general_dampers(/*threaded=*/false),
139  _splits(/*threaded=*/false),
140  _increment_vec(NULL),
141  _use_finite_differenced_preconditioner(false),
142  _fdcoloring(nullptr),
143  _fsp(nullptr),
144  _add_implicit_geometric_coupling_entries_to_jacobian(false),
145  _assemble_constraints_separately(false),
146  _need_residual_ghosted(false),
147  _debugging_residuals(false),
148  _doing_dg(false),
149  _n_iters(0),
150  _n_linear_iters(0),
151  _n_residual_evaluations(0),
152  _final_residual(0.),
153  _computing_pre_smo_residual(false),
154  _pre_smo_residual(0),
155  _initial_residual(0),
156  _use_pre_smo_residual(false),
157  _print_all_var_norms(false),
158  _has_save_in(false),
159  _has_diag_save_in(false),
160  _has_nodalbc_save_in(false),
161  _has_nodalbc_diag_save_in(false),
162  _computed_scaling(false),
163  _compute_scaling_once(true),
164  _resid_vs_jac_scaling_param(0),
165  _off_diagonals_in_auto_scaling(false),
166  _auto_scaling_initd(false)
167 {
169  // Don't need to add the matrix - it already exists (for now)
171 
172  // The time matrix tag is not normally used - but must be added to the system
173  // in case it is so that objects can have 'time' in their matrix tags by default
174  _fe_problem.addMatrixTag("TIME");
175 
176  _Re_tag = _fe_problem.addVectorTag("RESIDUAL");
177 
179 
181  {
182  auto & dof_map = _sys.get_dof_map();
183  dof_map.remove_algebraic_ghosting_functor(dof_map.default_algebraic_ghosting());
184  dof_map.set_implicit_neighbor_dofs(false);
185  }
186 }
187 
189 
190 void
192 {
194 
195  if (_fe_problem.hasDampers())
196  setupDampers();
197 
198  if (_residual_copy.get())
199  _residual_copy->init(_sys.n_dofs(), false, SERIAL);
200 
201 #ifdef MOOSE_KOKKOS_ENABLED
204 #endif
205 }
206 
207 void
209 {
210  // reinit is called on meshChanged() in FEProblemBase. We could implement meshChanged() instead.
211  // Subdomains might have changed
212  for (auto & functor : _displaced_mortar_functors)
213  functor.second.setupMortarMaterials();
214  for (auto & functor : _undisplaced_mortar_functors)
215  functor.second.setupMortarMaterials();
216 }
217 
218 void
220 {
222  nonlinearSolver()->jacobian = NULL;
223 }
224 
225 void
227 {
228  TIME_SECTION("nlInitialSetup", 2, "Setting Up Nonlinear System");
229 
231 
232  {
233  TIME_SECTION("kernelsInitialSetup", 2, "Setting Up Kernels/BCs/Constraints");
234 
235  for (THREAD_ID tid = 0; tid < libMesh::n_threads(); tid++)
236  {
237  _kernels.initialSetup(tid);
240  if (_doing_dg)
243 
247 
248  if (_fe_problem.haveFV())
249  {
250  std::vector<FVElementalKernel *> fv_elemental_kernels;
252  .query()
253  .template condition<AttribSystem>("FVElementalKernel")
254  .template condition<AttribThread>(tid)
255  .queryInto(fv_elemental_kernels);
256 
257  for (auto * fv_kernel : fv_elemental_kernels)
258  fv_kernel->initialSetup();
259 
260  std::vector<FVFluxKernel *> fv_flux_kernels;
262  .query()
263  .template condition<AttribSystem>("FVFluxKernel")
264  .template condition<AttribThread>(tid)
265  .queryInto(fv_flux_kernels);
266 
267  for (auto * fv_kernel : fv_flux_kernels)
268  fv_kernel->initialSetup();
269  }
270  }
271 
278 
279 #ifdef MOOSE_KOKKOS_ENABLED
284 #endif
285  }
286 
287  {
288  TIME_SECTION("mortarSetup", 2, "Initializing Mortar Interfaces");
289 
290  auto create_mortar_functors = [this](const bool displaced)
291  {
292  // go over mortar interfaces and construct functors
293  const auto & mortar_interfaces = _fe_problem.getMortarInterfaces(displaced);
294  for (const auto & [primary_secondary_boundary_pair, mortar_generation_ptr] :
295  mortar_interfaces)
296  {
297  if (!_constraints.hasActiveMortarConstraints(primary_secondary_boundary_pair, displaced))
298  continue;
299 
300  auto & mortar_constraints =
301  _constraints.getActiveMortarConstraints(primary_secondary_boundary_pair, displaced);
302 
303  auto & subproblem = displaced
304  ? static_cast<SubProblem &>(*_fe_problem.getDisplacedProblem())
305  : static_cast<SubProblem &>(_fe_problem);
306 
307  auto & mortar_functors =
309 
310  mortar_functors.emplace(primary_secondary_boundary_pair,
311  ComputeMortarFunctor(mortar_constraints,
312  *mortar_generation_ptr,
313  subproblem,
314  _fe_problem,
315  displaced,
316  subproblem.assembly(0, number())));
317  }
318  };
319 
320  create_mortar_functors(false);
321  create_mortar_functors(true);
322  }
323 
324  if (_automatic_scaling)
325  {
327  _scaling_matrix = std::make_unique<OffDiagonalScalingMatrix<Number>>(_communicator);
328  else
329  _scaling_matrix = std::make_unique<DiagonalMatrix<Number>>(_communicator);
330  }
331 
332  if (_preconditioner)
333  _preconditioner->initialSetup();
334 }
335 
336 void
338 {
340 
341  for (THREAD_ID tid = 0; tid < libMesh::n_threads(); tid++)
342  {
343  _kernels.timestepSetup(tid);
346  if (_doing_dg)
352 
353  if (_fe_problem.haveFV())
354  {
355  std::vector<FVFluxBC *> bcs;
357  .query()
358  .template condition<AttribSystem>("FVFluxBC")
359  .template condition<AttribThread>(tid)
360  .queryInto(bcs);
361 
362  std::vector<FVInterfaceKernel *> iks;
364  .query()
365  .template condition<AttribSystem>("FVInterfaceKernel")
366  .template condition<AttribThread>(tid)
367  .queryInto(iks);
368 
369  std::vector<FVFluxKernel *> kernels;
371  .query()
372  .template condition<AttribSystem>("FVFluxKernel")
373  .template condition<AttribThread>(tid)
374  .queryInto(kernels);
375 
376  for (auto * bc : bcs)
377  bc->timestepSetup();
378  for (auto * ik : iks)
379  ik->timestepSetup();
380  for (auto * kernel : kernels)
381  kernel->timestepSetup();
382  }
383  }
390 
391 #ifdef MOOSE_KOKKOS_ENABLED
396 #endif
397 }
398 
399 void
401 {
402  SolverSystem::customSetup(exec_type);
403 
404  for (THREAD_ID tid = 0; tid < libMesh::n_threads(); tid++)
405  {
406  _kernels.customSetup(exec_type, tid);
407  _nodal_kernels.customSetup(exec_type, tid);
408  _dirac_kernels.customSetup(exec_type, tid);
409  if (_doing_dg)
410  _dg_kernels.customSetup(exec_type, tid);
411  _interface_kernels.customSetup(exec_type, tid);
412  _element_dampers.customSetup(exec_type, tid);
413  _nodal_dampers.customSetup(exec_type, tid);
414  _integrated_bcs.customSetup(exec_type, tid);
415 
416  if (_fe_problem.haveFV())
417  {
418  std::vector<FVFluxBC *> bcs;
420  .query()
421  .template condition<AttribSystem>("FVFluxBC")
422  .template condition<AttribThread>(tid)
423  .queryInto(bcs);
424 
425  std::vector<FVInterfaceKernel *> iks;
427  .query()
428  .template condition<AttribSystem>("FVInterfaceKernel")
429  .template condition<AttribThread>(tid)
430  .queryInto(iks);
431 
432  std::vector<FVFluxKernel *> kernels;
434  .query()
435  .template condition<AttribSystem>("FVFluxKernel")
436  .template condition<AttribThread>(tid)
437  .queryInto(kernels);
438 
439  for (auto * bc : bcs)
440  bc->customSetup(exec_type);
441  for (auto * ik : iks)
442  ik->customSetup(exec_type);
443  for (auto * kernel : kernels)
444  kernel->customSetup(exec_type);
445  }
446  }
447  _scalar_kernels.customSetup(exec_type);
448  _constraints.customSetup(exec_type);
449  _general_dampers.customSetup(exec_type);
450  _nodal_bcs.customSetup(exec_type);
451  _preset_nodal_bcs.customSetup(exec_type);
453 
454 #ifdef MOOSE_KOKKOS_ENABLED
455  _kokkos_kernels.customSetup(exec_type);
458  _kokkos_nodal_bcs.customSetup(exec_type);
459 #endif
460 }
461 
462 void
464 {
465  if (_fsp)
466  _fsp->setupDM();
467 }
468 
469 void
470 NonlinearSystemBase::addKernel(const std::string & kernel_name,
471  const std::string & name,
472  InputParameters & parameters)
473 {
474  for (THREAD_ID tid = 0; tid < libMesh::n_threads(); tid++)
475  {
476  // Create the kernel object via the factory and add to warehouse
477  std::shared_ptr<KernelBase> kernel =
478  _factory.create<KernelBase>(kernel_name, name, parameters, tid);
479  _kernels.addObject(kernel, tid);
480  postAddResidualObject(*kernel);
481  // Add to theWarehouse, a centralized storage for all moose objects
482  _fe_problem.theWarehouse().add(kernel);
483  }
484 
485  if (parameters.get<std::vector<AuxVariableName>>("save_in").size() > 0)
486  _has_save_in = true;
487  if (parameters.get<std::vector<AuxVariableName>>("diag_save_in").size() > 0)
488  _has_diag_save_in = true;
489 }
490 
491 void
492 NonlinearSystemBase::addHDGKernel(const std::string & kernel_name,
493  const std::string & name,
494  InputParameters & parameters)
495 {
496  for (THREAD_ID tid = 0; tid < libMesh::n_threads(); tid++)
497  {
498  // Create the kernel object via the factory and add to warehouse
499  auto kernel = _factory.create<HDGKernel>(kernel_name, name, parameters, tid);
500  _kernels.addObject(kernel, tid);
501  _hybridized_kernels.addObject(kernel, tid);
502  // Add to theWarehouse, a centralized storage for all moose objects
503  _fe_problem.theWarehouse().add(kernel);
504  postAddResidualObject(*kernel);
505  }
506 }
507 
508 void
509 NonlinearSystemBase::addNodalKernel(const std::string & kernel_name,
510  const std::string & name,
511  InputParameters & parameters)
512 {
513  for (THREAD_ID tid = 0; tid < libMesh::n_threads(); tid++)
514  {
515  // Create the kernel object via the factory and add to the warehouse
516  std::shared_ptr<NodalKernelBase> kernel =
517  _factory.create<NodalKernelBase>(kernel_name, name, parameters, tid);
518  _nodal_kernels.addObject(kernel, tid);
519  // Add to theWarehouse, a centralized storage for all moose objects
520  _fe_problem.theWarehouse().add(kernel);
521  postAddResidualObject(*kernel);
522  }
523 
524  if (parameters.get<std::vector<AuxVariableName>>("save_in").size() > 0)
525  _has_save_in = true;
526  if (parameters.get<std::vector<AuxVariableName>>("diag_save_in").size() > 0)
527  _has_diag_save_in = true;
528 }
529 
530 void
531 NonlinearSystemBase::addScalarKernel(const std::string & kernel_name,
532  const std::string & name,
533  InputParameters & parameters)
534 {
535  std::shared_ptr<ScalarKernelBase> kernel =
536  _factory.create<ScalarKernelBase>(kernel_name, name, parameters);
537  postAddResidualObject(*kernel);
538  // Add to theWarehouse, a centralized storage for all moose objects
539  _fe_problem.theWarehouse().add(kernel);
540  _scalar_kernels.addObject(kernel);
541 }
542 
543 void
544 NonlinearSystemBase::addBoundaryCondition(const std::string & bc_name,
545  const std::string & name,
546  InputParameters & parameters)
547 {
548  // ThreadID
549  THREAD_ID tid = 0;
550 
551  // Create the object
552  std::shared_ptr<BoundaryCondition> bc =
553  _factory.create<BoundaryCondition>(bc_name, name, parameters, tid);
555 
556  // Active BoundaryIDs for the object
557  const std::set<BoundaryID> & boundary_ids = bc->boundaryIDs();
558  auto bc_var = dynamic_cast<const MooseVariableFieldBase *>(&bc->variable());
559  _vars[tid].addBoundaryVar(boundary_ids, bc_var);
560 
561  // Cast to the various types of BCs
562  std::shared_ptr<NodalBCBase> nbc = std::dynamic_pointer_cast<NodalBCBase>(bc);
563  std::shared_ptr<IntegratedBCBase> ibc = std::dynamic_pointer_cast<IntegratedBCBase>(bc);
564 
565  // NodalBCBase
566  if (nbc)
567  {
568  if (nbc->checkNodalVar() && !nbc->variable().isNodal())
569  mooseError("Trying to use nodal boundary condition '",
570  nbc->name(),
571  "' on a non-nodal variable '",
572  nbc->variable().name(),
573  "'.");
574 
575  _nodal_bcs.addObject(nbc);
576  // Add to theWarehouse, a centralized storage for all moose objects
578  _vars[tid].addBoundaryVars(boundary_ids, nbc->getCoupledVars());
579 
580  if (parameters.get<std::vector<AuxVariableName>>("save_in").size() > 0)
581  _has_nodalbc_save_in = true;
582  if (parameters.get<std::vector<AuxVariableName>>("diag_save_in").size() > 0)
584 
585  // DirichletBCs that are preset
586  std::shared_ptr<DirichletBCBase> dbc = std::dynamic_pointer_cast<DirichletBCBase>(bc);
587  if (dbc && dbc->preset())
589 
590  std::shared_ptr<ADDirichletBCBase> addbc = std::dynamic_pointer_cast<ADDirichletBCBase>(bc);
591  if (addbc && addbc->preset())
593  }
594 
595  // IntegratedBCBase
596  else if (ibc)
597  {
598  _integrated_bcs.addObject(ibc, tid);
599  // Add to theWarehouse, a centralized storage for all moose objects
601  _vars[tid].addBoundaryVars(boundary_ids, ibc->getCoupledVars());
602 
603  if (parameters.get<std::vector<AuxVariableName>>("save_in").size() > 0)
604  _has_save_in = true;
605  if (parameters.get<std::vector<AuxVariableName>>("diag_save_in").size() > 0)
606  _has_diag_save_in = true;
607 
608  for (tid = 1; tid < libMesh::n_threads(); tid++)
609  {
610  // Create the object
611  bc = _factory.create<BoundaryCondition>(bc_name, name, parameters, tid);
612 
613  // Give users opportunity to set some parameters
615 
616  // Active BoundaryIDs for the object
617  const std::set<BoundaryID> & boundary_ids = bc->boundaryIDs();
618  _vars[tid].addBoundaryVar(boundary_ids, bc_var);
619 
620  ibc = std::static_pointer_cast<IntegratedBCBase>(bc);
621 
622  _integrated_bcs.addObject(ibc, tid);
623  _vars[tid].addBoundaryVars(boundary_ids, ibc->getCoupledVars());
624  }
625  }
626 
627  else
628  mooseError("Unknown BoundaryCondition type for object named ", bc->name());
629 }
630 
631 void
632 NonlinearSystemBase::addConstraint(const std::string & c_name,
633  const std::string & name,
634  InputParameters & parameters)
635 {
636  std::shared_ptr<Constraint> constraint = _factory.create<Constraint>(c_name, name, parameters);
637  _constraints.addObject(constraint);
638  postAddResidualObject(*constraint);
639 
640  if (constraint && constraint->addCouplingEntriesToJacobian())
642 }
643 
644 void
645 NonlinearSystemBase::addDiracKernel(const std::string & kernel_name,
646  const std::string & name,
647  InputParameters & parameters)
648 {
649  for (THREAD_ID tid = 0; tid < libMesh::n_threads(); tid++)
650  {
651  std::shared_ptr<DiracKernelBase> kernel =
652  _factory.create<DiracKernelBase>(kernel_name, name, parameters, tid);
653  postAddResidualObject(*kernel);
654  _dirac_kernels.addObject(kernel, tid);
655  // Add to theWarehouse, a centralized storage for all moose objects
656  _fe_problem.theWarehouse().add(kernel);
657  }
658 }
659 
660 void
661 NonlinearSystemBase::addDGKernel(std::string dg_kernel_name,
662  const std::string & name,
663  InputParameters & parameters)
664 {
665  for (THREAD_ID tid = 0; tid < libMesh::n_threads(); ++tid)
666  {
667  auto dg_kernel = _factory.create<DGKernelBase>(dg_kernel_name, name, parameters, tid);
668  _dg_kernels.addObject(dg_kernel, tid);
669  // Add to theWarehouse, a centralized storage for all moose objects
670  _fe_problem.theWarehouse().add(dg_kernel);
671  postAddResidualObject(*dg_kernel);
672  }
673 
674  _doing_dg = true;
675 
676  if (parameters.get<std::vector<AuxVariableName>>("save_in").size() > 0)
677  _has_save_in = true;
678  if (parameters.get<std::vector<AuxVariableName>>("diag_save_in").size() > 0)
679  _has_diag_save_in = true;
680 }
681 
682 void
683 NonlinearSystemBase::addInterfaceKernel(std::string interface_kernel_name,
684  const std::string & name,
685  InputParameters & parameters)
686 {
687  for (THREAD_ID tid = 0; tid < libMesh::n_threads(); ++tid)
688  {
689  std::shared_ptr<InterfaceKernelBase> interface_kernel =
690  _factory.create<InterfaceKernelBase>(interface_kernel_name, name, parameters, tid);
691  postAddResidualObject(*interface_kernel);
692 
693  const std::set<BoundaryID> & boundary_ids = interface_kernel->boundaryIDs();
694  auto ik_var = dynamic_cast<const MooseVariableFieldBase *>(&interface_kernel->variable());
695  _vars[tid].addBoundaryVar(boundary_ids, ik_var);
696 
697  _interface_kernels.addObject(interface_kernel, tid);
698  // Add to theWarehouse, a centralized storage for all moose objects
699  _fe_problem.theWarehouse().add(interface_kernel);
700  _vars[tid].addBoundaryVars(boundary_ids, interface_kernel->getCoupledVars());
701  }
702 }
703 
704 void
705 NonlinearSystemBase::addDamper(const std::string & damper_name,
706  const std::string & name,
707  InputParameters & parameters)
708 {
709  for (THREAD_ID tid = 0; tid < libMesh::n_threads(); ++tid)
710  {
711  std::shared_ptr<Damper> damper = _factory.create<Damper>(damper_name, name, parameters, tid);
712 
713  // Attempt to cast to the damper types
714  std::shared_ptr<ElementDamper> ed = std::dynamic_pointer_cast<ElementDamper>(damper);
715  std::shared_ptr<NodalDamper> nd = std::dynamic_pointer_cast<NodalDamper>(damper);
716  std::shared_ptr<GeneralDamper> gd = std::dynamic_pointer_cast<GeneralDamper>(damper);
717 
718  if (gd)
719  {
721  break; // not threaded
722  }
723  else if (ed)
724  _element_dampers.addObject(ed, tid);
725  else if (nd)
726  _nodal_dampers.addObject(nd, tid);
727  else
728  mooseError("Invalid damper type");
729  }
730 }
731 
732 void
733 NonlinearSystemBase::addSplit(const std::string & split_name,
734  const std::string & name,
735  InputParameters & parameters)
736 {
737  std::shared_ptr<Split> split = _factory.create<Split>(split_name, name, parameters);
739  // Add to theWarehouse, a centralized storage for all moose objects
741 }
742 
743 std::shared_ptr<Split>
744 NonlinearSystemBase::getSplit(const std::string & name)
745 {
746  return _splits.getActiveObject(name);
747 }
748 
749 bool
751 {
753  return false;
754 
755  // The legacy behavior (#10464) _always_ performs the pre-SMO residual evaluation
756  // regardless of whether it is needed.
757  //
758  // This is not ideal and has been fixed by #23472. This legacy option ensures a smooth transition
759  // to the new behavior. Modules and Apps that want to migrate to the new behavior should set this
760  // parameter to false.
761  if (_app.parameters().get<bool>("use_legacy_initial_residual_evaluation_behavior"))
762  return true;
763 
764  return _use_pre_smo_residual;
765 }
766 
767 Real
769 {
771 }
772 
773 Real
775 {
777  mooseError("pre-SMO residual is requested but not evaluated.");
778 
779  return _pre_smo_residual;
780 }
781 
782 Real
784 {
785  return _initial_residual;
786 }
787 
788 void
790 {
791  _initial_residual = r;
792 }
793 
794 void
795 NonlinearSystemBase::zeroVectorForResidual(const std::string & vector_name)
796 {
797  for (unsigned int i = 0; i < _vecs_to_zero_for_residual.size(); ++i)
798  if (vector_name == _vecs_to_zero_for_residual[i])
799  return;
800 
801  _vecs_to_zero_for_residual.push_back(vector_name);
802 }
803 
804 void
806 {
807  _nl_vector_tags.clear();
808  _nl_vector_tags.insert(tag_id);
810 
812 
814 
816 }
817 
818 void
820 {
821  mooseDeprecated(" Please use computeResidualTag");
822 
823  computeResidualTag(residual, tag_id);
824 }
825 
826 void
827 NonlinearSystemBase::computeResidualTags(const std::set<TagID> & tags)
828 {
829  parallel_object_only();
830 
831  TIME_SECTION("nl::computeResidualTags", 5);
832 
835 
836  bool required_residual = tags.find(residualVectorTag()) == tags.end() ? false : true;
837 
839 
840  // not suppose to do anythin on matrix
842 
844 
845  for (const auto & numeric_vec : _vecs_to_zero_for_residual)
846  if (hasVector(numeric_vec))
847  {
848  NumericVector<Number> & vec = getVector(numeric_vec);
849  vec.close();
850  vec.zero();
851  }
852 
853  try
854  {
855  zeroTaggedVectors(tags);
857  closeTaggedVectors(tags);
858 
859  if (required_residual)
860  {
861  auto & residual = getVector(residualVectorTag());
862  if (!_time_integrators.empty())
863  {
864  for (auto & ti : _time_integrators)
865  ti->postResidual(residual);
866  }
867  else
868  residual += *_Re_non_time;
869  residual.close();
870  }
872  // We don't want to do nodal bcs or anything else
873  return;
874 
875  computeNodalBCs(tags);
876  closeTaggedVectors(tags);
877 
878  // If we are debugging residuals we need one more assignment to have the ghosted copy up to
879  // date
880  if (_need_residual_ghosted && _debugging_residuals && required_residual)
881  {
882  auto & residual = getVector(residualVectorTag());
883 
884  *_residual_ghosted = residual;
886  }
887  // Need to close and update the aux system in case residuals were saved to it.
890  if (hasSaveIn())
892  }
893  catch (MooseException & e)
894  {
895  // The buck stops here, we have already handled the exception by
896  // calling stopSolve(), it is now up to PETSc to return a
897  // "diverged" reason during the next solve.
898  }
899 
900  // not supposed to do anything on matrix
902 
904 }
905 
906 void
907 NonlinearSystemBase::computeResidualAndJacobianTags(const std::set<TagID> & vector_tags,
908  const std::set<TagID> & matrix_tags)
909 {
910  const bool required_residual =
911  vector_tags.find(residualVectorTag()) == vector_tags.end() ? false : true;
912 
913  try
914  {
915  zeroTaggedVectors(vector_tags);
916  computeResidualAndJacobianInternal(vector_tags, matrix_tags);
917  closeTaggedVectors(vector_tags);
918  closeTaggedMatrices(matrix_tags);
919 
920  if (required_residual)
921  {
922  auto & residual = getVector(residualVectorTag());
923  if (!_time_integrators.empty())
924  {
925  for (auto & ti : _time_integrators)
926  ti->postResidual(residual);
927  }
928  else
929  residual += *_Re_non_time;
930  residual.close();
931  }
932 
934  closeTaggedVectors(vector_tags);
935  closeTaggedMatrices(matrix_tags);
936  }
937  catch (MooseException & e)
938  {
939  // The buck stops here, we have already handled the exception by
940  // calling stopSolve(), it is now up to PETSc to return a
941  // "diverged" reason during the next solve.
942  }
943 }
944 
945 void
947 {
948  for (auto & ti : _time_integrators)
949  ti->preSolve();
950  if (_predictor.get())
951  _predictor->timestepSetup();
952 }
953 
954 void
956 {
958 
959  NumericVector<Number> & initial_solution(solution());
960  if (_predictor.get())
961  {
962  if (_predictor->shouldApply())
963  {
964  TIME_SECTION("applyPredictor", 2, "Applying Predictor");
965 
966  _predictor->apply(initial_solution);
967  _fe_problem.predictorCleanup(initial_solution);
968  }
969  else
970  _console << " Skipping predictor this step" << std::endl;
971  }
972 
973  // do nodal BC
974  {
975  TIME_SECTION("initialBCs", 2, "Applying BCs To Initial Condition");
976 
978  for (const auto & bnode : bnd_nodes)
979  {
980  BoundaryID boundary_id = bnode->_bnd_id;
981  Node * node = bnode->_node;
982 
983  if (node->processor_id() == processor_id())
984  {
985  bool has_preset_nodal_bcs = _preset_nodal_bcs.hasActiveBoundaryObjects(boundary_id);
986  bool has_ad_preset_nodal_bcs = _ad_preset_nodal_bcs.hasActiveBoundaryObjects(boundary_id);
987 
988  // reinit variables in nodes
989  if (has_preset_nodal_bcs || has_ad_preset_nodal_bcs)
990  _fe_problem.reinitNodeFace(node, boundary_id, 0);
991 
992  if (has_preset_nodal_bcs)
993  {
994  const auto & preset_bcs = _preset_nodal_bcs.getActiveBoundaryObjects(boundary_id);
995  for (const auto & preset_bc : preset_bcs)
996  preset_bc->computeValue(initial_solution);
997  }
998  if (has_ad_preset_nodal_bcs)
999  {
1000  const auto & preset_bcs_res = _ad_preset_nodal_bcs.getActiveBoundaryObjects(boundary_id);
1001  for (const auto & preset_bc : preset_bcs_res)
1002  preset_bc->computeValue(initial_solution);
1003  }
1004  }
1005  }
1006  }
1007 
1008 #ifdef MOOSE_KOKKOS_ENABLED
1011 #endif
1012 
1013  _sys.solution->close();
1014  update();
1015 
1016  // Set constraint secondary values
1017  setConstraintSecondaryValues(initial_solution, false);
1018 
1020  setConstraintSecondaryValues(initial_solution, true);
1021 }
1022 
1023 void
1024 NonlinearSystemBase::setPredictor(std::shared_ptr<Predictor> predictor)
1025 {
1026  _predictor = predictor;
1027 }
1028 
1029 void
1031 {
1033 
1034  _kernels.subdomainSetup(subdomain, tid);
1035  _nodal_kernels.subdomainSetup(subdomain, tid);
1036  _element_dampers.subdomainSetup(subdomain, tid);
1037  _nodal_dampers.subdomainSetup(subdomain, tid);
1038 }
1039 
1042 {
1043  if (!_Re_time)
1044  {
1046 
1047  // Most applications don't need the expense of ghosting
1049  _Re_time = &addVector(_Re_time_tag, false, ptype);
1050  }
1051  else if (_need_residual_ghosted && _Re_time->type() == PARALLEL)
1052  {
1053  const auto vector_name = _subproblem.vectorTagName(_Re_time_tag);
1054 
1055  // If an application changes its mind, the libMesh API lets us
1056  // change the vector.
1057  _Re_time = &system().add_vector(vector_name, false, GHOSTED);
1058  }
1059 
1060  return *_Re_time;
1061 }
1062 
1065 {
1066  if (!_Re_non_time)
1067  {
1069 
1070  // Most applications don't need the expense of ghosting
1072  _Re_non_time = &addVector(_Re_non_time_tag, false, ptype);
1073  }
1075  {
1076  const auto vector_name = _subproblem.vectorTagName(_Re_non_time_tag);
1077 
1078  // If an application changes its mind, the libMesh API lets us
1079  // change the vector.
1080  _Re_non_time = &system().add_vector(vector_name, false, GHOSTED);
1081  }
1082 
1083  return *_Re_non_time;
1084 }
1085 
1088 {
1089  mooseDeprecated("Please use getVector()");
1090  switch (tag)
1091  {
1092  case 0:
1093  return getResidualNonTimeVector();
1094 
1095  case 1:
1096  return getResidualTimeVector();
1097 
1098  default:
1099  mooseError("The required residual vector is not available");
1100  }
1101 }
1102 
1103 void
1105 {
1106  THREAD_ID tid = 0; // constraints are going to be done single-threaded
1107  residual.close();
1109  {
1110  const auto & ncs = _constraints.getActiveNodalConstraints();
1111  for (const auto & nc : ncs)
1112  {
1113  std::vector<dof_id_type> & secondary_node_ids = nc->getSecondaryNodeId();
1114  std::vector<dof_id_type> & primary_node_ids = nc->getPrimaryNodeId();
1115 
1116  if ((secondary_node_ids.size() > 0) && (primary_node_ids.size() > 0))
1117  {
1118  _fe_problem.reinitNodes(primary_node_ids, tid);
1119  _fe_problem.reinitNodesNeighbor(secondary_node_ids, tid);
1120  nc->computeResidual(residual);
1121  }
1122  }
1123  _fe_problem.addCachedResidualDirectly(residual, tid);
1124  residual.close();
1125  }
1126 }
1127 
1128 void
1130 {
1131  if (!hasMatrix(systemMatrixTag()))
1132  mooseError(" A system matrix is required");
1133 
1134  auto & jacobian = getMatrix(systemMatrixTag());
1135  THREAD_ID tid = 0; // constraints are going to be done single-threaded
1136 
1138  {
1139  const auto & ncs = _constraints.getActiveNodalConstraints();
1140  for (const auto & nc : ncs)
1141  {
1142  std::vector<dof_id_type> & secondary_node_ids = nc->getSecondaryNodeId();
1143  std::vector<dof_id_type> & primary_node_ids = nc->getPrimaryNodeId();
1144 
1145  if ((secondary_node_ids.size() > 0) && (primary_node_ids.size() > 0))
1146  {
1147  _fe_problem.reinitNodes(primary_node_ids, tid);
1148  _fe_problem.reinitNodesNeighbor(secondary_node_ids, tid);
1149  nc->computeJacobian(jacobian);
1150  }
1151  }
1153  }
1154 }
1155 
1156 void
1158  const BoundaryID secondary_boundary,
1159  const PenetrationInfo & info,
1160  const bool displaced)
1161 {
1162  auto & subproblem = displaced ? static_cast<SubProblem &>(*_fe_problem.getDisplacedProblem())
1163  : static_cast<SubProblem &>(_fe_problem);
1164 
1165  const Elem * primary_elem = info._elem;
1166  unsigned int primary_side = info._side_num;
1167  std::vector<Point> points;
1168  points.push_back(info._closest_point);
1169 
1170  // *These next steps MUST be done in this order!*
1171  // ADL: This is a Chesterton's fence situation. I don't know which calls exactly the above comment
1172  // is referring to. If I had to guess I would guess just the reinitNodeFace and prepareAssembly
1173  // calls since the former will size the variable's dof indices and then the latter will resize the
1174  // residual/Jacobian based off the variable's cached dof indices size
1175 
1176  // This reinits the variables that exist on the secondary node
1177  _fe_problem.reinitNodeFace(&secondary_node, secondary_boundary, 0);
1178 
1179  // This will set aside residual and jacobian space for the variables that have dofs on
1180  // the secondary node
1182 
1183  _fe_problem.setNeighborSubdomainID(primary_elem, 0);
1184 
1185  //
1186  // Reinit material on undisplaced mesh
1187  //
1188 
1189  const Elem * const undisplaced_primary_elem =
1190  displaced ? _mesh.elemPtr(primary_elem->id()) : primary_elem;
1191  const Point undisplaced_primary_physical_point =
1192  [&points, displaced, primary_elem, undisplaced_primary_elem]()
1193  {
1194  if (displaced)
1195  {
1196  const Point reference_point =
1197  FEMap::inverse_map(primary_elem->dim(), primary_elem, points[0]);
1198  return FEMap::map(primary_elem->dim(), undisplaced_primary_elem, reference_point);
1199  }
1200  else
1201  // If our penetration locator is on the reference mesh, then our undisplaced
1202  // physical point is simply the point coming from the penetration locator
1203  return points[0];
1204  }();
1205 
1207  undisplaced_primary_elem, primary_side, {undisplaced_primary_physical_point}, 0);
1208  // Stateful material properties are only initialized for neighbor material data for internal faces
1209  // for discontinuous Galerkin methods or for conforming interfaces for interface kernels. We don't
1210  // have either of those use cases here where we likely have disconnected meshes
1211  _fe_problem.reinitMaterialsNeighbor(primary_elem->subdomain_id(), 0, /*swap_stateful=*/false);
1212 
1213  // Reinit points for constraint enforcement
1214  if (displaced)
1215  subproblem.reinitNeighborPhys(primary_elem, primary_side, points, 0);
1216 }
1217 
1218 void
1220 {
1221 
1222  if (displaced)
1223  mooseAssert(_fe_problem.getDisplacedProblem(),
1224  "If we're calling this method with displaced = true, then we better well have a "
1225  "displaced problem");
1226  auto & subproblem = displaced ? static_cast<SubProblem &>(*_fe_problem.getDisplacedProblem())
1227  : static_cast<SubProblem &>(_fe_problem);
1228  const auto & penetration_locators = subproblem.geomSearchData()._penetration_locators;
1229 
1230  bool constraints_applied = false;
1231 
1232  for (const auto & it : penetration_locators)
1233  {
1234  PenetrationLocator & pen_loc = *(it.second);
1235 
1236  std::vector<dof_id_type> & secondary_nodes = pen_loc._nearest_node._secondary_nodes;
1237 
1238  BoundaryID secondary_boundary = pen_loc._secondary_boundary;
1239  BoundaryID primary_boundary = pen_loc._primary_boundary;
1240 
1241  if (_constraints.hasActiveNodeFaceConstraints(secondary_boundary, displaced))
1242  {
1243  const auto & constraints =
1244  _constraints.getActiveNodeFaceConstraints(secondary_boundary, displaced);
1245  std::unordered_set<unsigned int> needed_mat_props;
1246  for (const auto & constraint : constraints)
1247  {
1248  const auto & mp_deps = constraint->getMatPropDependencies();
1249  needed_mat_props.insert(mp_deps.begin(), mp_deps.end());
1250  }
1251  _fe_problem.setActiveMaterialProperties(needed_mat_props, /*tid=*/0);
1252 
1253  for (unsigned int i = 0; i < secondary_nodes.size(); i++)
1254  {
1255  dof_id_type secondary_node_num = secondary_nodes[i];
1256  Node & secondary_node = _mesh.nodeRef(secondary_node_num);
1257 
1258  if (secondary_node.processor_id() == processor_id())
1259  {
1260  if (pen_loc._penetration_info[secondary_node_num])
1261  {
1262  PenetrationInfo & info = *pen_loc._penetration_info[secondary_node_num];
1263 
1264  reinitNodeFace(secondary_node, secondary_boundary, info, displaced);
1265 
1266  for (const auto & nfc : constraints)
1267  {
1268  if (nfc->isExplicitConstraint())
1269  continue;
1270  // Return if this constraint does not correspond to the primary-secondary pair
1271  // prepared by the outer loops.
1272  // This continue statement is required when, e.g. one secondary surface constrains
1273  // more than one primary surface.
1274  if (nfc->secondaryBoundary() != secondary_boundary ||
1275  nfc->primaryBoundary() != primary_boundary)
1276  continue;
1277 
1278  if (nfc->shouldApply())
1279  {
1280  constraints_applied = true;
1281  nfc->computeSecondaryValue(solution);
1282  }
1283 
1284  if (nfc->hasWritableCoupledVariables())
1285  {
1286  Threads::spin_mutex::scoped_lock lock(Threads::spin_mtx);
1287  for (auto * var : nfc->getWritableCoupledVariables())
1288  {
1289  if (var->isNodalDefined())
1290  var->insert(_fe_problem.getAuxiliarySystem().solution());
1291  }
1292  }
1293  }
1294  }
1295  }
1296  }
1297  }
1298  }
1299 
1300  // go over NodeELemConstraints
1301  std::set<dof_id_type> unique_secondary_node_ids;
1302 
1303  for (const auto & secondary_id : _mesh.meshSubdomains())
1304  {
1305  for (const auto & primary_id : _mesh.meshSubdomains())
1306  {
1307  if (_constraints.hasActiveNodeElemConstraints(secondary_id, primary_id, displaced))
1308  {
1309  const auto & constraints =
1310  _constraints.getActiveNodeElemConstraints(secondary_id, primary_id, displaced);
1311 
1312  // get unique set of ids of all nodes on current block
1313  unique_secondary_node_ids.clear();
1314  const MeshBase & meshhelper = _mesh.getMesh();
1315  for (const auto & elem : as_range(meshhelper.active_subdomain_elements_begin(secondary_id),
1316  meshhelper.active_subdomain_elements_end(secondary_id)))
1317  {
1318  for (auto & n : elem->node_ref_range())
1319  unique_secondary_node_ids.insert(n.id());
1320  }
1321 
1322  for (auto secondary_node_id : unique_secondary_node_ids)
1323  {
1324  Node & secondary_node = _mesh.nodeRef(secondary_node_id);
1325 
1326  // check if secondary node is on current processor
1327  if (secondary_node.processor_id() == processor_id())
1328  {
1329  // This reinits the variables that exist on the secondary node
1330  _fe_problem.reinitNodeFace(&secondary_node, secondary_id, 0);
1331 
1332  // This will set aside residual and jacobian space for the variables that have dofs
1333  // on the secondary node
1335 
1336  for (const auto & nec : constraints)
1337  {
1338  if (nec->shouldApply())
1339  {
1340  constraints_applied = true;
1341  nec->computeSecondaryValue(solution);
1342  }
1343  }
1344  }
1345  }
1346  }
1347  }
1348  }
1349 
1350  // See if constraints were applied anywhere
1351  _communicator.max(constraints_applied);
1352 
1353  if (constraints_applied)
1354  {
1355  solution.close();
1356  update();
1357  }
1358 }
1359 
1360 void
1362 {
1363  // Make sure the residual is in a good state
1364  residual.close();
1365 
1366  if (displaced)
1367  mooseAssert(_fe_problem.getDisplacedProblem(),
1368  "If we're calling this method with displaced = true, then we better well have a "
1369  "displaced problem");
1370  auto & subproblem = displaced ? static_cast<SubProblem &>(*_fe_problem.getDisplacedProblem())
1371  : static_cast<SubProblem &>(_fe_problem);
1372  const auto & penetration_locators = subproblem.geomSearchData()._penetration_locators;
1373 
1374  bool constraints_applied;
1375  bool residual_has_inserted_values = false;
1377  constraints_applied = false;
1378  for (const auto & it : penetration_locators)
1379  {
1381  {
1382  // Reset the constraint_applied flag before each new constraint, as they need to be
1383  // assembled separately
1384  constraints_applied = false;
1385  }
1386  PenetrationLocator & pen_loc = *(it.second);
1387 
1388  std::vector<dof_id_type> & secondary_nodes = pen_loc._nearest_node._secondary_nodes;
1389 
1390  BoundaryID secondary_boundary = pen_loc._secondary_boundary;
1391  BoundaryID primary_boundary = pen_loc._primary_boundary;
1392 
1393  bool has_writable_variables(false);
1394 
1395  if (_constraints.hasActiveNodeFaceConstraints(secondary_boundary, displaced))
1396  {
1397  const auto & constraints =
1398  _constraints.getActiveNodeFaceConstraints(secondary_boundary, displaced);
1399 
1400  for (unsigned int i = 0; i < secondary_nodes.size(); i++)
1401  {
1402  dof_id_type secondary_node_num = secondary_nodes[i];
1403  Node & secondary_node = _mesh.nodeRef(secondary_node_num);
1404 
1405  if (secondary_node.processor_id() == processor_id())
1406  {
1407  if (pen_loc._penetration_info[secondary_node_num])
1408  {
1409  PenetrationInfo & info = *pen_loc._penetration_info[secondary_node_num];
1410 
1411  reinitNodeFace(secondary_node, secondary_boundary, info, displaced);
1412 
1413  for (const auto & nfc : constraints)
1414  {
1415  // Return if this constraint does not correspond to the primary-secondary pair
1416  // prepared by the outer loops.
1417  // This continue statement is required when, e.g. one secondary surface constrains
1418  // more than one primary surface.
1419  if (nfc->secondaryBoundary() != secondary_boundary ||
1420  nfc->primaryBoundary() != primary_boundary)
1421  continue;
1422 
1423  if (nfc->shouldApply())
1424  {
1425  constraints_applied = true;
1426  nfc->computeResidual();
1427 
1428  if (nfc->overwriteSecondaryResidual())
1429  {
1430  // The below will actually overwrite the residual for every single dof that
1431  // lives on the node. We definitely don't want to do that!
1432  // _fe_problem.setResidual(residual, 0);
1433 
1434  const auto & secondary_var = nfc->variable();
1435  const auto & secondary_dofs = secondary_var.dofIndices();
1436  mooseAssert(secondary_dofs.size() == secondary_var.count(),
1437  "We are on a node so there should only be one dof per variable (for "
1438  "an ArrayVariable we should have a number of dofs equal to the "
1439  "number of components");
1440 
1441  // Assume that if the user is overwriting the secondary residual, then they are
1442  // supplying residuals that do not correspond to their other physics
1443  // (e.g. Kernels), hence we should not apply a scalingFactor that is normally
1444  // based on the order of their other physics (e.g. Kernels)
1445  std::vector<Number> values = {nfc->secondaryResidual()};
1446  residual.insert(values, secondary_dofs);
1447  residual_has_inserted_values = true;
1448  }
1449  else
1452  }
1453  if (nfc->hasWritableCoupledVariables())
1454  {
1455  Threads::spin_mutex::scoped_lock lock(Threads::spin_mtx);
1456  has_writable_variables = true;
1457  for (auto * var : nfc->getWritableCoupledVariables())
1458  {
1459  if (var->isNodalDefined())
1460  var->insert(_fe_problem.getAuxiliarySystem().solution());
1461  }
1462  }
1463  }
1464  }
1465  }
1466  }
1467  }
1468  _communicator.max(has_writable_variables);
1469 
1470  if (has_writable_variables)
1471  {
1472  // Explicit contact dynamic constraints write to auxiliary variables and update the old
1473  // displacement solution on the constraint boundaries. Close solutions and update system
1474  // accordingly.
1477  solutionOld().close();
1478  }
1479 
1481  {
1482  // Make sure that secondary contribution to primary are assembled, and ghosts have been
1483  // exchanged, as current primaries might become secondaries on next iteration and will need to
1484  // contribute their former secondaries' contributions to the future primaries. See if
1485  // constraints were applied anywhere
1486  _communicator.max(constraints_applied);
1487 
1488  if (constraints_applied)
1489  {
1490  // If any of the above constraints inserted values in the residual, it needs to be
1491  // assembled before adding the cached residuals below.
1492  _communicator.max(residual_has_inserted_values);
1493  if (residual_has_inserted_values)
1494  {
1495  residual.close();
1496  residual_has_inserted_values = false;
1497  }
1499  residual.close();
1500 
1502  *_residual_ghosted = residual;
1503  }
1504  }
1505  }
1507  {
1508  _communicator.max(constraints_applied);
1509 
1510  if (constraints_applied)
1511  {
1512  // If any of the above constraints inserted values in the residual, it needs to be assembled
1513  // before adding the cached residuals below.
1514  _communicator.max(residual_has_inserted_values);
1515  if (residual_has_inserted_values)
1516  residual.close();
1517 
1519  residual.close();
1520 
1522  *_residual_ghosted = residual;
1523  }
1524  }
1525 
1526  // go over element-element constraint interface
1527  THREAD_ID tid = 0;
1528  const auto & element_pair_locators = subproblem.geomSearchData()._element_pair_locators;
1529  for (const auto & it : element_pair_locators)
1530  {
1531  ElementPairLocator & elem_pair_loc = *(it.second);
1532 
1533  if (_constraints.hasActiveElemElemConstraints(it.first, displaced))
1534  {
1535  // ElemElemConstraint objects
1536  const auto & element_constraints =
1537  _constraints.getActiveElemElemConstraints(it.first, displaced);
1538 
1539  // go over pair elements
1540  const std::list<std::pair<const Elem *, const Elem *>> & elem_pairs =
1541  elem_pair_loc.getElemPairs();
1542  for (const auto & pr : elem_pairs)
1543  {
1544  const Elem * elem1 = pr.first;
1545  const Elem * elem2 = pr.second;
1546 
1547  if (elem1->processor_id() != processor_id())
1548  continue;
1549 
1550  const ElementPairInfo & info = elem_pair_loc.getElemPairInfo(pr);
1551 
1552  // for each element process constraints on the
1553  for (const auto & ec : element_constraints)
1554  {
1555  _fe_problem.setCurrentSubdomainID(elem1, tid);
1556  subproblem.reinitElemPhys(elem1, info._elem1_constraint_q_point, tid);
1557  _fe_problem.setNeighborSubdomainID(elem2, tid);
1558  subproblem.reinitNeighborPhys(elem2, info._elem2_constraint_q_point, tid);
1559 
1560  ec->prepareShapes(ec->variable().number());
1561  ec->prepareNeighborShapes(ec->variable().number());
1562 
1563  ec->reinit(info);
1564  ec->computeResidual();
1567  }
1569  }
1570  }
1571  }
1572 
1573  // go over NodeElemConstraints
1574  std::set<dof_id_type> unique_secondary_node_ids;
1575 
1576  constraints_applied = false;
1577  residual_has_inserted_values = false;
1578  bool has_writable_variables = false;
1579  for (const auto & secondary_id : _mesh.meshSubdomains())
1580  {
1581  for (const auto & primary_id : _mesh.meshSubdomains())
1582  {
1583  if (_constraints.hasActiveNodeElemConstraints(secondary_id, primary_id, displaced))
1584  {
1585  const auto & constraints =
1586  _constraints.getActiveNodeElemConstraints(secondary_id, primary_id, displaced);
1587 
1588  // get unique set of ids of all nodes on current block
1589  unique_secondary_node_ids.clear();
1590  const MeshBase & meshhelper = _mesh.getMesh();
1591  for (const auto & elem : as_range(meshhelper.active_subdomain_elements_begin(secondary_id),
1592  meshhelper.active_subdomain_elements_end(secondary_id)))
1593  {
1594  for (auto & n : elem->node_ref_range())
1595  unique_secondary_node_ids.insert(n.id());
1596  }
1597 
1598  for (auto secondary_node_id : unique_secondary_node_ids)
1599  {
1600  Node & secondary_node = _mesh.nodeRef(secondary_node_id);
1601  // check if secondary node is on current processor
1602  if (secondary_node.processor_id() == processor_id())
1603  {
1604  // This reinits the variables that exist on the secondary node
1605  _fe_problem.reinitNodeFace(&secondary_node, secondary_id, 0);
1606 
1607  // This will set aside residual and jacobian space for the variables that have dofs
1608  // on the secondary node
1610 
1611  for (const auto & nec : constraints)
1612  {
1613  if (nec->shouldApply())
1614  {
1615  constraints_applied = true;
1616  nec->computeResidual();
1617 
1618  if (nec->overwriteSecondaryResidual())
1619  {
1620  _fe_problem.setResidual(residual, 0);
1621  residual_has_inserted_values = true;
1622  }
1623  else
1626  }
1627  if (nec->hasWritableCoupledVariables())
1628  {
1629  Threads::spin_mutex::scoped_lock lock(Threads::spin_mtx);
1630  has_writable_variables = true;
1631  for (auto * var : nec->getWritableCoupledVariables())
1632  {
1633  if (var->isNodalDefined())
1634  var->insert(_fe_problem.getAuxiliarySystem().solution());
1635  }
1636  }
1637  }
1639  }
1640  }
1641  }
1642  }
1643  }
1644  _communicator.max(constraints_applied);
1645 
1646  if (constraints_applied)
1647  {
1648  // If any of the above constraints inserted values in the residual, it needs to be assembled
1649  // before adding the cached residuals below.
1650  _communicator.max(residual_has_inserted_values);
1651  if (residual_has_inserted_values)
1652  residual.close();
1653 
1655  residual.close();
1656 
1658  *_residual_ghosted = residual;
1659  }
1660  _communicator.max(has_writable_variables);
1661 
1662  if (has_writable_variables)
1663  {
1664  // Explicit contact dynamic constraints write to auxiliary variables and update the old
1665  // displacement solution on the constraint boundaries. Close solutions and update system
1666  // accordingly.
1669  solutionOld().close();
1670  }
1671 
1672  // We may have additional tagged vectors that also need to be accumulated
1674 }
1675 
1676 void
1678 {
1679  // Overwrite results from integrator in case we have explicit dynamics contact constraints
1681  ? static_cast<SubProblem &>(*_fe_problem.getDisplacedProblem())
1682  : static_cast<SubProblem &>(_fe_problem);
1683  const auto & penetration_locators = subproblem.geomSearchData()._penetration_locators;
1684 
1685  for (const auto & it : penetration_locators)
1686  {
1687  PenetrationLocator & pen_loc = *(it.second);
1688 
1689  const auto & secondary_nodes = pen_loc._nearest_node._secondary_nodes;
1690  const BoundaryID secondary_boundary = pen_loc._secondary_boundary;
1691  const BoundaryID primary_boundary = pen_loc._primary_boundary;
1692 
1693  if (_constraints.hasActiveNodeFaceConstraints(secondary_boundary, true))
1694  {
1695  const auto & constraints =
1696  _constraints.getActiveNodeFaceConstraints(secondary_boundary, true);
1697  for (const auto i : index_range(secondary_nodes))
1698  {
1699  const auto secondary_node_num = secondary_nodes[i];
1700  const Node & secondary_node = _mesh.nodeRef(secondary_node_num);
1701 
1702  if (secondary_node.processor_id() == processor_id())
1703  if (pen_loc._penetration_info[secondary_node_num])
1704  for (const auto & nfc : constraints)
1705  {
1706  if (!nfc->isExplicitConstraint())
1707  continue;
1708 
1709  // Return if this constraint does not correspond to the primary-secondary pair
1710  // prepared by the outer loops.
1711  // This continue statement is required when, e.g. one secondary surface constrains
1712  // more than one primary surface.
1713  if (nfc->secondaryBoundary() != secondary_boundary ||
1714  nfc->primaryBoundary() != primary_boundary)
1715  continue;
1716 
1717  nfc->overwriteBoundaryVariables(soln, secondary_node);
1718  }
1719  }
1720  }
1721  }
1722  soln.close();
1723 }
1724 
1725 void
1727 {
1728  TIME_SECTION("residualSetup", 3);
1729 
1731 
1732  for (THREAD_ID tid = 0; tid < libMesh::n_threads(); tid++)
1733  {
1734  _kernels.residualSetup(tid);
1737  if (_doing_dg)
1743  }
1750 
1751 #ifdef MOOSE_KOKKOS_ENABLED
1756 #endif
1757 
1758  // Avoid recursion
1759  if (this == &_fe_problem.currentNonlinearSystem())
1762 }
1763 
1764 void
1766 {
1767  parallel_object_only();
1768 
1769  TIME_SECTION("computeResidualInternal", 3);
1770 
1771  residualSetup();
1772 
1773 #ifdef MOOSE_KOKKOS_ENABLED
1775  computeKokkosResidual(tags);
1776 #endif
1777 
1778  const auto vector_tag_data = _fe_problem.getVectorTags(tags);
1779 
1780  // Residual contributions from UOs - for now this is used for ray tracing
1781  // and ray kernels that contribute to the residual (think line sources)
1782  std::vector<UserObject *> uos;
1784  .query()
1785  .condition<AttribSystem>("UserObject")
1786  .condition<AttribExecOns>(EXEC_PRE_KERNELS)
1787  .queryInto(uos);
1788  for (auto & uo : uos)
1789  uo->residualSetup();
1790  for (auto & uo : uos)
1791  {
1792  uo->initialize();
1793  uo->execute();
1794  uo->finalize();
1795  }
1796 
1797  // reinit scalar variables
1798  for (unsigned int tid = 0; tid < libMesh::n_threads(); tid++)
1800 
1801  // residual contributions from the domain
1802  PARALLEL_TRY
1803  {
1804  TIME_SECTION("Kernels", 3 /*, "Computing Kernels"*/);
1805 
1807 
1809  Threads::parallel_reduce(elem_range, cr);
1810 
1811  // We pass face information directly to FV residual objects for their evaluation. Consequently
1812  // we must make sure to do separate threaded loops for 1) undisplaced face information objects
1813  // and undisplaced residual objects and 2) displaced face information objects and displaced
1814  // residual objects
1816  if (_fe_problem.haveFV())
1817  {
1819  _fe_problem, this->number(), tags, /*on_displaced=*/false);
1821  Threads::parallel_reduce(faces, fvr);
1822  }
1824  displaced_problem && displaced_problem->haveFV())
1825  {
1827  _fe_problem, this->number(), tags, /*on_displaced=*/true);
1828  FVRange faces(displaced_problem->mesh().ownedFaceInfoBegin(),
1829  displaced_problem->mesh().ownedFaceInfoEnd());
1830  Threads::parallel_reduce(faces, fvr);
1831  }
1832 
1833  unsigned int n_threads = libMesh::n_threads();
1834  for (unsigned int i = 0; i < n_threads;
1835  i++) // Add any cached residuals that might be hanging around
1837  }
1838  PARALLEL_CATCH;
1839 
1840  // residual contributions from the scalar kernels
1841  PARALLEL_TRY
1842  {
1843  // do scalar kernels (not sure how to thread this)
1845  {
1846  TIME_SECTION("ScalarKernels", 3 /*, "Computing ScalarKernels"*/);
1847 
1848  MooseObjectWarehouse<ScalarKernelBase> * scalar_kernel_warehouse;
1849  // This code should be refactored once we can do tags for scalar
1850  // kernels
1851  // Should redo this based on Warehouse
1852  if (!tags.size() || tags.size() == _fe_problem.numVectorTags(Moose::VECTOR_TAG_RESIDUAL))
1853  scalar_kernel_warehouse = &_scalar_kernels;
1854  else if (tags.size() == 1)
1855  scalar_kernel_warehouse =
1856  &(_scalar_kernels.getVectorTagObjectWarehouse(*(tags.begin()), 0));
1857  else
1858  // scalar_kernels is not threading
1859  scalar_kernel_warehouse = &(_scalar_kernels.getVectorTagsObjectWarehouse(tags, 0));
1860 
1861  bool have_scalar_contributions = false;
1862  const auto & scalars = scalar_kernel_warehouse->getActiveObjects();
1863  for (const auto & scalar_kernel : scalars)
1864  {
1865  scalar_kernel->reinit();
1866  const std::vector<dof_id_type> & dof_indices = scalar_kernel->variable().dofIndices();
1867  const DofMap & dof_map = scalar_kernel->variable().dofMap();
1868  const dof_id_type first_dof = dof_map.first_dof();
1869  const dof_id_type end_dof = dof_map.end_dof();
1870  for (dof_id_type dof : dof_indices)
1871  {
1872  if (dof >= first_dof && dof < end_dof)
1873  {
1874  scalar_kernel->computeResidual();
1875  have_scalar_contributions = true;
1876  break;
1877  }
1878  }
1879  }
1880  if (have_scalar_contributions)
1882  }
1883  }
1884  PARALLEL_CATCH;
1885 
1886  // residual contributions from Block NodalKernels
1887  PARALLEL_TRY
1888  {
1890  {
1891  TIME_SECTION("NodalKernels", 3 /*, "Computing NodalKernels"*/);
1892 
1894 
1896 
1897  if (range.begin() != range.end())
1898  {
1899  _fe_problem.reinitNode(*range.begin(), 0);
1900 
1901  Threads::parallel_reduce(range, cnk);
1902 
1903  unsigned int n_threads = libMesh::n_threads();
1904  for (unsigned int i = 0; i < n_threads;
1905  i++) // Add any cached residuals that might be hanging around
1907  }
1908  }
1909  }
1910  PARALLEL_CATCH;
1911 
1913  // We computed the volumetric objects. We can return now before we get into
1914  // any strongly enforced constraint conditions or penalty-type objects
1915  // (DGKernels, IntegratedBCs, InterfaceKernels, Constraints)
1916  return;
1917 
1918  // residual contributions from boundary NodalKernels
1919  PARALLEL_TRY
1920  {
1922  {
1923  TIME_SECTION("NodalKernelBCs", 3 /*, "Computing NodalKernelBCs"*/);
1924 
1926 
1928 
1929  Threads::parallel_reduce(bnd_node_range, cnk);
1930 
1931  unsigned int n_threads = libMesh::n_threads();
1932  for (unsigned int i = 0; i < n_threads;
1933  i++) // Add any cached residuals that might be hanging around
1935  }
1936  }
1937  PARALLEL_CATCH;
1938 
1940 
1941  if (_residual_copy.get())
1942  {
1943  _Re_non_time->close();
1945  }
1946 
1948  {
1949  _Re_non_time->close();
1952  }
1953 
1954  PARALLEL_TRY { computeDiracContributions(tags, false); }
1955  PARALLEL_CATCH;
1956 
1958  {
1959  PARALLEL_TRY { enforceNodalConstraintsResidual(*_Re_non_time); }
1960  PARALLEL_CATCH;
1961  _Re_non_time->close();
1962  }
1963 
1964  // Add in Residual contributions from other Constraints
1966  {
1967  PARALLEL_TRY
1968  {
1969  // Undisplaced Constraints
1971 
1972  // Displaced Constraints
1975 
1978  }
1979  PARALLEL_CATCH;
1980  _Re_non_time->close();
1981  }
1982 
1983  // Accumulate the occurrence of solution invalid warnings for the current iteration cumulative
1984  // counters
1987 }
1988 
1989 void
1991  const std::set<TagID> & matrix_tags)
1992 {
1993  TIME_SECTION("computeResidualAndJacobianInternal", 3);
1994 
1995  // Make matrix ready to use
1997 
1998  for (auto tag : matrix_tags)
1999  {
2000  if (!hasMatrix(tag))
2001  continue;
2002 
2003  auto & jacobian = getMatrix(tag);
2004  // Necessary for speed
2005  if (auto petsc_matrix = dynamic_cast<PetscMatrix<Number> *>(&jacobian))
2006  {
2007  LibmeshPetscCall(MatSetOption(petsc_matrix->mat(),
2008  MAT_KEEP_NONZERO_PATTERN, // This is changed in 3.1
2009  PETSC_TRUE));
2011  LibmeshPetscCall(
2012  MatSetOption(petsc_matrix->mat(), MAT_NEW_NONZERO_ALLOCATION_ERR, PETSC_FALSE));
2014  LibmeshPetscCall(MatSetOption(static_cast<PetscMatrix<Number> &>(jacobian).mat(),
2015  MAT_IGNORE_ZERO_ENTRIES,
2016  PETSC_TRUE));
2017  }
2018  }
2019 
2020  residualSetup();
2021 
2022  // Residual contributions from UOs - for now this is used for ray tracing
2023  // and ray kernels that contribute to the residual (think line sources)
2024  std::vector<UserObject *> uos;
2026  .query()
2027  .condition<AttribSystem>("UserObject")
2028  .condition<AttribExecOns>(EXEC_PRE_KERNELS)
2029  .queryInto(uos);
2030  for (auto & uo : uos)
2031  uo->residualSetup();
2032  for (auto & uo : uos)
2033  {
2034  uo->initialize();
2035  uo->execute();
2036  uo->finalize();
2037  }
2038 
2039  // reinit scalar variables
2040  for (unsigned int tid = 0; tid < libMesh::n_threads(); tid++)
2042 
2043  // residual contributions from the domain
2044  PARALLEL_TRY
2045  {
2046  TIME_SECTION("Kernels", 3 /*, "Computing Kernels"*/);
2047 
2049 
2050  ComputeResidualAndJacobianThread crj(_fe_problem, vector_tags, matrix_tags);
2051  Threads::parallel_reduce(elem_range, crj);
2052 
2054  if (_fe_problem.haveFV())
2055  {
2057  _fe_problem, this->number(), vector_tags, matrix_tags, /*on_displaced=*/false);
2059  Threads::parallel_reduce(faces, fvrj);
2060  }
2062  displaced_problem && displaced_problem->haveFV())
2063  {
2065  _fe_problem, this->number(), vector_tags, matrix_tags, /*on_displaced=*/true);
2066  FVRange faces(displaced_problem->mesh().ownedFaceInfoBegin(),
2067  displaced_problem->mesh().ownedFaceInfoEnd());
2068  Threads::parallel_reduce(faces, fvr);
2069  }
2070 
2072 
2073  unsigned int n_threads = libMesh::n_threads();
2074  for (unsigned int i = 0; i < n_threads;
2075  i++) // Add any cached residuals that might be hanging around
2076  {
2079  }
2080  }
2081  PARALLEL_CATCH;
2082 }
2083 
2084 void
2086 {
2087  _nl_vector_tags.clear();
2088 
2089  const auto & residual_vector_tags = _fe_problem.getVectorTags(Moose::VECTOR_TAG_RESIDUAL);
2090  for (const auto & residual_vector_tag : residual_vector_tags)
2091  _nl_vector_tags.insert(residual_vector_tag._id);
2092 
2094  computeNodalBCs(residual, _nl_vector_tags);
2096 }
2097 
2098 void
2099 NonlinearSystemBase::computeNodalBCs(NumericVector<Number> & residual, const std::set<TagID> & tags)
2100 {
2102 
2103  computeNodalBCs(tags);
2104 
2106 }
2107 
2108 void
2109 NonlinearSystemBase::computeNodalBCs(const std::set<TagID> & tags)
2110 {
2111  // We need to close the diag_save_in variables on the aux system before NodalBCBases clear the
2112  // dofs on boundary nodes
2113  if (_has_save_in)
2115 
2116  // Select nodal kernels
2117  MooseObjectWarehouse<NodalBCBase> * nbc_warehouse;
2118 
2119  if (tags.size() == _fe_problem.numVectorTags(Moose::VECTOR_TAG_RESIDUAL) || !tags.size())
2120  nbc_warehouse = &_nodal_bcs;
2121  else if (tags.size() == 1)
2122  nbc_warehouse = &(_nodal_bcs.getVectorTagObjectWarehouse(*(tags.begin()), 0));
2123  else
2124  nbc_warehouse = &(_nodal_bcs.getVectorTagsObjectWarehouse(tags, 0));
2125 
2126  // Return early if there is no nodal kernel
2127  if (!nbc_warehouse->size())
2128  return;
2129 
2130  PARALLEL_TRY
2131  {
2133 
2134  if (!bnd_nodes.empty())
2135  {
2136  TIME_SECTION("NodalBCs", 3 /*, "Computing NodalBCs"*/);
2137 
2138  for (const auto & bnode : bnd_nodes)
2139  {
2140  BoundaryID boundary_id = bnode->_bnd_id;
2141  Node * node = bnode->_node;
2142 
2143  if (node->processor_id() == processor_id() &&
2144  nbc_warehouse->hasActiveBoundaryObjects(boundary_id))
2145  {
2146  // reinit variables in nodes
2147  _fe_problem.reinitNodeFace(node, boundary_id, 0);
2148 
2149  const auto & bcs = nbc_warehouse->getActiveBoundaryObjects(boundary_id);
2150  for (const auto & nbc : bcs)
2151  if (nbc->shouldApply())
2152  nbc->computeResidual();
2153  }
2154  }
2155  }
2156  }
2157  PARALLEL_CATCH;
2158 
2159  if (_Re_time)
2160  _Re_time->close();
2161  _Re_non_time->close();
2162 }
2163 
2164 void
2166 {
2167  PARALLEL_TRY
2168  {
2170 
2171  if (!bnd_nodes.empty())
2172  {
2173  TIME_SECTION("NodalBCs", 3 /*, "Computing NodalBCs"*/);
2174 
2175  for (const auto & bnode : bnd_nodes)
2176  {
2177  BoundaryID boundary_id = bnode->_bnd_id;
2178  Node * node = bnode->_node;
2179 
2180  if (node->processor_id() == processor_id())
2181  {
2182  // reinit variables in nodes
2183  _fe_problem.reinitNodeFace(node, boundary_id, 0);
2184  if (_nodal_bcs.hasActiveBoundaryObjects(boundary_id))
2185  {
2186  const auto & bcs = _nodal_bcs.getActiveBoundaryObjects(boundary_id);
2187  for (const auto & nbc : bcs)
2188  if (nbc->shouldApply())
2189  nbc->computeResidualAndJacobian();
2190  }
2191  }
2192  }
2193  }
2194  }
2195  PARALLEL_CATCH;
2196 
2197  // Set the cached NodalBCBase values in the Jacobian matrix
2199 }
2200 
2201 void
2202 NonlinearSystemBase::getNodeDofs(dof_id_type node_id, std::vector<dof_id_type> & dofs)
2203 {
2204  const Node & node = _mesh.nodeRef(node_id);
2205  unsigned int s = number();
2206  if (node.has_dofs(s))
2207  {
2208  for (unsigned int v = 0; v < nVariables(); v++)
2209  for (unsigned int c = 0; c < node.n_comp(s, v); c++)
2210  dofs.push_back(node.dof_number(s, v, c));
2211  }
2212 }
2213 
2214 void
2216  GeometricSearchData & geom_search_data,
2217  std::unordered_map<dof_id_type, std::vector<dof_id_type>> & graph)
2218 {
2219  const auto & node_to_elem_map = _mesh.nodeToElemMap();
2220  const auto & nearest_node_locators = geom_search_data._nearest_node_locators;
2221  for (const auto & it : nearest_node_locators)
2222  {
2223  std::vector<dof_id_type> & secondary_nodes = it.second->_secondary_nodes;
2224 
2225  for (const auto & secondary_node : secondary_nodes)
2226  {
2227  std::set<dof_id_type> unique_secondary_indices;
2228  std::set<dof_id_type> unique_primary_indices;
2229 
2230  auto node_to_elem_pair = node_to_elem_map.find(secondary_node);
2231  if (node_to_elem_pair != node_to_elem_map.end())
2232  {
2233  const std::vector<dof_id_type> & elems = node_to_elem_pair->second;
2234 
2235  // Get the dof indices from each elem connected to the node
2236  for (const auto & cur_elem : elems)
2237  {
2238  std::vector<dof_id_type> dof_indices;
2239  dofMap().dof_indices(_mesh.elemPtr(cur_elem), dof_indices);
2240 
2241  for (const auto & dof : dof_indices)
2242  unique_secondary_indices.insert(dof);
2243  }
2244  }
2245 
2246  std::vector<dof_id_type> primary_nodes = it.second->_neighbor_nodes[secondary_node];
2247 
2248  for (const auto & primary_node : primary_nodes)
2249  {
2250  auto primary_node_to_elem_pair = node_to_elem_map.find(primary_node);
2251  mooseAssert(primary_node_to_elem_pair != node_to_elem_map.end(),
2252  "Missing entry in node to elem map");
2253  const std::vector<dof_id_type> & primary_node_elems = primary_node_to_elem_pair->second;
2254 
2255  // Get the dof indices from each elem connected to the node
2256  for (const auto & cur_elem : primary_node_elems)
2257  {
2258  std::vector<dof_id_type> dof_indices;
2259  dofMap().dof_indices(_mesh.elemPtr(cur_elem), dof_indices);
2260 
2261  for (const auto & dof : dof_indices)
2262  unique_primary_indices.insert(dof);
2263  }
2264  }
2265 
2266  for (const auto & secondary_id : unique_secondary_indices)
2267  for (const auto & primary_id : unique_primary_indices)
2268  {
2269  graph[secondary_id].push_back(primary_id);
2270  graph[primary_id].push_back(secondary_id);
2271  }
2272  }
2273  }
2274 
2275  // handle node-to-node constraints
2276  const auto & ncs = _constraints.getActiveNodalConstraints();
2277  for (const auto & nc : ncs)
2278  {
2279  std::vector<dof_id_type> primary_dofs;
2280  std::vector<dof_id_type> & primary_node_ids = nc->getPrimaryNodeId();
2281  for (const auto & node_id : primary_node_ids)
2282  {
2283  Node * node = _mesh.queryNodePtr(node_id);
2284  if (node && node->processor_id() == this->processor_id())
2285  {
2286  getNodeDofs(node_id, primary_dofs);
2287  }
2288  }
2289 
2290  _communicator.allgather(primary_dofs);
2291 
2292  std::vector<dof_id_type> secondary_dofs;
2293  std::vector<dof_id_type> & secondary_node_ids = nc->getSecondaryNodeId();
2294  for (const auto & node_id : secondary_node_ids)
2295  {
2296  Node * node = _mesh.queryNodePtr(node_id);
2297  if (node && node->processor_id() == this->processor_id())
2298  {
2299  getNodeDofs(node_id, secondary_dofs);
2300  }
2301  }
2302 
2303  _communicator.allgather(secondary_dofs);
2304 
2305  for (const auto & primary_id : primary_dofs)
2306  for (const auto & secondary_id : secondary_dofs)
2307  {
2308  graph[primary_id].push_back(secondary_id);
2309  graph[secondary_id].push_back(primary_id);
2310  }
2311  }
2312 
2313  // Make every entry sorted and unique
2314  for (auto & it : graph)
2315  {
2316  std::vector<dof_id_type> & row = it.second;
2317  std::sort(row.begin(), row.end());
2318  std::vector<dof_id_type>::iterator uit = std::unique(row.begin(), row.end());
2319  row.resize(uit - row.begin());
2320  }
2321 }
2322 
2323 void
2325 {
2326  if (!hasMatrix(systemMatrixTag()))
2327  mooseError("Need a system matrix ");
2328 
2329  // At this point, have no idea how to make
2330  // this work with tag system
2331  auto & jacobian = getMatrix(systemMatrixTag());
2332 
2333  std::unordered_map<dof_id_type, std::vector<dof_id_type>> graph;
2334 
2335  findImplicitGeometricCouplingEntries(geom_search_data, graph);
2336 
2337  for (const auto & it : graph)
2338  {
2339  dof_id_type dof = it.first;
2340  const auto & row = it.second;
2341 
2342  for (const auto & coupled_dof : row)
2343  jacobian.add(dof, coupled_dof, 0);
2344  }
2345 }
2346 
2347 void
2349  bool displaced)
2350 {
2351  if (!hasMatrix(systemMatrixTag()))
2352  mooseError("A system matrix is required");
2353 
2354  auto & jacobian = getMatrix(systemMatrixTag());
2355 
2357  LibmeshPetscCall(MatSetOption(static_cast<PetscMatrix<Number> &>(jacobian).mat(),
2358  MAT_NEW_NONZERO_ALLOCATION_ERR,
2359  PETSC_FALSE));
2361  LibmeshPetscCall(MatSetOption(
2362  static_cast<PetscMatrix<Number> &>(jacobian).mat(), MAT_IGNORE_ZERO_ENTRIES, PETSC_TRUE));
2363 
2364  std::vector<numeric_index_type> zero_rows;
2365 
2366  if (displaced)
2367  mooseAssert(_fe_problem.getDisplacedProblem(),
2368  "If we're calling this method with displaced = true, then we better well have a "
2369  "displaced problem");
2370  auto & subproblem = displaced ? static_cast<SubProblem &>(*_fe_problem.getDisplacedProblem())
2371  : static_cast<SubProblem &>(_fe_problem);
2372  const auto & penetration_locators = subproblem.geomSearchData()._penetration_locators;
2373 
2374  bool constraints_applied;
2376  constraints_applied = false;
2377  for (const auto & it : penetration_locators)
2378  {
2380  {
2381  // Reset the constraint_applied flag before each new constraint, as they need to be
2382  // assembled separately
2383  constraints_applied = false;
2384  }
2385  PenetrationLocator & pen_loc = *(it.second);
2386 
2387  std::vector<dof_id_type> & secondary_nodes = pen_loc._nearest_node._secondary_nodes;
2388 
2389  BoundaryID secondary_boundary = pen_loc._secondary_boundary;
2390  BoundaryID primary_boundary = pen_loc._primary_boundary;
2391 
2392  zero_rows.clear();
2393  if (_constraints.hasActiveNodeFaceConstraints(secondary_boundary, displaced))
2394  {
2395  const auto & constraints =
2396  _constraints.getActiveNodeFaceConstraints(secondary_boundary, displaced);
2397 
2398  for (const auto & secondary_node_num : secondary_nodes)
2399  {
2400  Node & secondary_node = _mesh.nodeRef(secondary_node_num);
2401 
2402  if (secondary_node.processor_id() == processor_id())
2403  {
2404  if (pen_loc._penetration_info[secondary_node_num])
2405  {
2406  PenetrationInfo & info = *pen_loc._penetration_info[secondary_node_num];
2407 
2408  reinitNodeFace(secondary_node, secondary_boundary, info, displaced);
2410 
2411  for (const auto & nfc : constraints)
2412  {
2413  if (nfc->isExplicitConstraint())
2414  continue;
2415  // Return if this constraint does not correspond to the primary-secondary pair
2416  // prepared by the outer loops.
2417  // This continue statement is required when, e.g. one secondary surface constrains
2418  // more than one primary surface.
2419  if (nfc->secondaryBoundary() != secondary_boundary ||
2420  nfc->primaryBoundary() != primary_boundary)
2421  continue;
2422 
2423  nfc->_jacobian = &jacobian_to_view;
2424 
2425  if (nfc->shouldApply())
2426  {
2427  constraints_applied = true;
2428 
2429  nfc->prepareShapes(nfc->variable().number());
2430  nfc->prepareNeighborShapes(nfc->variable().number());
2431 
2432  nfc->computeJacobian();
2433 
2434  if (nfc->overwriteSecondaryJacobian())
2435  {
2436  // Add this variable's dof's row to be zeroed
2437  zero_rows.push_back(nfc->variable().nodalDofIndex());
2438  }
2439 
2440  std::vector<dof_id_type> secondary_dofs(1, nfc->variable().nodalDofIndex());
2441 
2442  // Assume that if the user is overwriting the secondary Jacobian, then they are
2443  // supplying Jacobians that do not correspond to their other physics
2444  // (e.g. Kernels), hence we should not apply a scalingFactor that is normally
2445  // based on the order of their other physics (e.g. Kernels)
2446  Real scaling_factor =
2447  nfc->overwriteSecondaryJacobian() ? 1. : nfc->variable().scalingFactor();
2448 
2449  // Cache the jacobian block for the secondary side
2450  nfc->addJacobian(_fe_problem.assembly(0, number()),
2451  nfc->_Kee,
2452  secondary_dofs,
2453  nfc->_connected_dof_indices,
2454  scaling_factor);
2455 
2456  // Cache Ken, Kne, Knn
2457  if (nfc->addCouplingEntriesToJacobian())
2458  {
2459  // Make sure we use a proper scaling factor (e.g. don't use an interior scaling
2460  // factor when we're overwriting secondary stuff)
2461  nfc->addJacobian(_fe_problem.assembly(0, number()),
2462  nfc->_Ken,
2463  secondary_dofs,
2464  nfc->primaryVariable().dofIndicesNeighbor(),
2465  scaling_factor);
2466 
2467  // Use _connected_dof_indices to get all the correct columns
2468  nfc->addJacobian(_fe_problem.assembly(0, number()),
2469  nfc->_Kne,
2470  nfc->primaryVariable().dofIndicesNeighbor(),
2471  nfc->_connected_dof_indices,
2472  nfc->primaryVariable().scalingFactor());
2473 
2474  // We've handled Ken and Kne, finally handle Knn
2476  }
2477 
2478  // Do the off-diagonals next
2479  const std::vector<MooseVariableFEBase *> coupled_vars = nfc->getCoupledMooseVars();
2480  for (const auto & jvar : coupled_vars)
2481  {
2482  // Only compute jacobians for nonlinear variables
2483  if (jvar->kind() != Moose::VAR_SOLVER)
2484  continue;
2485 
2486  // Only compute Jacobian entries if this coupling is being used by the
2487  // preconditioner
2488  if (nfc->variable().number() == jvar->number() ||
2490  nfc->variable().number(), jvar->number(), this->number()))
2491  continue;
2492 
2493  // Need to zero out the matrices first
2495 
2496  nfc->prepareShapes(nfc->variable().number());
2497  nfc->prepareNeighborShapes(jvar->number());
2498 
2499  nfc->computeOffDiagJacobian(jvar->number());
2500 
2501  // Cache the jacobian block for the secondary side
2502  nfc->addJacobian(_fe_problem.assembly(0, number()),
2503  nfc->_Kee,
2504  secondary_dofs,
2505  nfc->_connected_dof_indices,
2506  scaling_factor);
2507 
2508  // Cache Ken, Kne, Knn
2509  if (nfc->addCouplingEntriesToJacobian())
2510  {
2511  // Make sure we use a proper scaling factor (e.g. don't use an interior scaling
2512  // factor when we're overwriting secondary stuff)
2513  nfc->addJacobian(_fe_problem.assembly(0, number()),
2514  nfc->_Ken,
2515  secondary_dofs,
2516  jvar->dofIndicesNeighbor(),
2517  scaling_factor);
2518 
2519  // Use _connected_dof_indices to get all the correct columns
2520  nfc->addJacobian(_fe_problem.assembly(0, number()),
2521  nfc->_Kne,
2522  nfc->variable().dofIndicesNeighbor(),
2523  nfc->_connected_dof_indices,
2524  nfc->variable().scalingFactor());
2525 
2526  // We've handled Ken and Kne, finally handle Knn
2528  }
2529  }
2530  }
2531  }
2532  }
2533  }
2534  }
2535  }
2537  {
2538  // See if constraints were applied anywhere
2539  _communicator.max(constraints_applied);
2540 
2541  if (constraints_applied)
2542  {
2543  LibmeshPetscCall(MatSetOption(static_cast<PetscMatrix<Number> &>(jacobian).mat(),
2544  MAT_KEEP_NONZERO_PATTERN, // This is changed in 3.1
2545  PETSC_TRUE));
2546 
2547  jacobian.close();
2548  jacobian.zero_rows(zero_rows, 0.0);
2549  jacobian.close();
2551  jacobian.close();
2552  }
2553  }
2554  }
2556  {
2557  // See if constraints were applied anywhere
2558  _communicator.max(constraints_applied);
2559 
2560  if (constraints_applied)
2561  {
2562  LibmeshPetscCall(MatSetOption(static_cast<PetscMatrix<Number> &>(jacobian).mat(),
2563  MAT_KEEP_NONZERO_PATTERN, // This is changed in 3.1
2564  PETSC_TRUE));
2565 
2566  jacobian.close();
2567  jacobian.zero_rows(zero_rows, 0.0);
2568  jacobian.close();
2570  jacobian.close();
2571  }
2572  }
2573 
2574  THREAD_ID tid = 0;
2575  // go over element-element constraint interface
2576  const auto & element_pair_locators = subproblem.geomSearchData()._element_pair_locators;
2577  for (const auto & it : element_pair_locators)
2578  {
2579  ElementPairLocator & elem_pair_loc = *(it.second);
2580 
2581  if (_constraints.hasActiveElemElemConstraints(it.first, displaced))
2582  {
2583  // ElemElemConstraint objects
2584  const auto & element_constraints =
2585  _constraints.getActiveElemElemConstraints(it.first, displaced);
2586 
2587  // go over pair elements
2588  const std::list<std::pair<const Elem *, const Elem *>> & elem_pairs =
2589  elem_pair_loc.getElemPairs();
2590  for (const auto & pr : elem_pairs)
2591  {
2592  const Elem * elem1 = pr.first;
2593  const Elem * elem2 = pr.second;
2594 
2595  if (elem1->processor_id() != processor_id())
2596  continue;
2597 
2598  const ElementPairInfo & info = elem_pair_loc.getElemPairInfo(pr);
2599 
2600  // for each element process constraints on the
2601  for (const auto & ec : element_constraints)
2602  {
2603  _fe_problem.setCurrentSubdomainID(elem1, tid);
2604  subproblem.reinitElemPhys(elem1, info._elem1_constraint_q_point, tid);
2605  _fe_problem.setNeighborSubdomainID(elem2, tid);
2606  subproblem.reinitNeighborPhys(elem2, info._elem2_constraint_q_point, tid);
2607 
2608  ec->prepareShapes(ec->variable().number());
2609  ec->prepareNeighborShapes(ec->variable().number());
2610 
2611  ec->reinit(info);
2612  ec->computeJacobian();
2615  }
2617  }
2618  }
2619  }
2620 
2621  // go over NodeElemConstraints
2622  std::set<dof_id_type> unique_secondary_node_ids;
2623  constraints_applied = false;
2624  for (const auto & secondary_id : _mesh.meshSubdomains())
2625  {
2626  for (const auto & primary_id : _mesh.meshSubdomains())
2627  {
2628  if (_constraints.hasActiveNodeElemConstraints(secondary_id, primary_id, displaced))
2629  {
2630  const auto & constraints =
2631  _constraints.getActiveNodeElemConstraints(secondary_id, primary_id, displaced);
2632 
2633  // get unique set of ids of all nodes on current block
2634  unique_secondary_node_ids.clear();
2635  const MeshBase & meshhelper = _mesh.getMesh();
2636  for (const auto & elem : as_range(meshhelper.active_subdomain_elements_begin(secondary_id),
2637  meshhelper.active_subdomain_elements_end(secondary_id)))
2638  {
2639  for (auto & n : elem->node_ref_range())
2640  unique_secondary_node_ids.insert(n.id());
2641  }
2642 
2643  for (auto secondary_node_id : unique_secondary_node_ids)
2644  {
2645  const Node & secondary_node = _mesh.nodeRef(secondary_node_id);
2646  // check if secondary node is on current processor
2647  if (secondary_node.processor_id() == processor_id())
2648  {
2649  // This reinits the variables that exist on the secondary node
2650  _fe_problem.reinitNodeFace(&secondary_node, secondary_id, 0);
2651 
2652  // This will set aside residual and jacobian space for the variables that have dofs
2653  // on the secondary node
2656 
2657  for (const auto & nec : constraints)
2658  {
2659  if (nec->shouldApply())
2660  {
2661  constraints_applied = true;
2662 
2663  nec->_jacobian = &jacobian_to_view;
2664  nec->prepareShapes(nec->variable().number());
2665  nec->prepareNeighborShapes(nec->variable().number());
2666 
2667  nec->computeJacobian();
2668 
2669  if (nec->overwriteSecondaryJacobian())
2670  {
2671  // Add this variable's dof's row to be zeroed
2672  zero_rows.push_back(nec->variable().nodalDofIndex());
2673  }
2674 
2675  std::vector<dof_id_type> secondary_dofs(1, nec->variable().nodalDofIndex());
2676 
2677  // Cache the jacobian block for the secondary side
2678  nec->addJacobian(_fe_problem.assembly(0, number()),
2679  nec->_Kee,
2680  secondary_dofs,
2681  nec->_connected_dof_indices,
2682  nec->variable().scalingFactor());
2683 
2684  // Cache the jacobian block for the primary side
2685  nec->addJacobian(_fe_problem.assembly(0, number()),
2686  nec->_Kne,
2687  nec->primaryVariable().dofIndicesNeighbor(),
2688  nec->_connected_dof_indices,
2689  nec->primaryVariable().scalingFactor());
2690 
2693 
2694  // Do the off-diagonals next
2695  const std::vector<MooseVariableFEBase *> coupled_vars = nec->getCoupledMooseVars();
2696  for (const auto & jvar : coupled_vars)
2697  {
2698  // Only compute jacobians for nonlinear variables
2699  if (jvar->kind() != Moose::VAR_SOLVER)
2700  continue;
2701 
2702  // Only compute Jacobian entries if this coupling is being used by the
2703  // preconditioner
2704  if (nec->variable().number() == jvar->number() ||
2706  nec->variable().number(), jvar->number(), this->number()))
2707  continue;
2708 
2709  // Need to zero out the matrices first
2711 
2712  nec->prepareShapes(nec->variable().number());
2713  nec->prepareNeighborShapes(jvar->number());
2714 
2715  nec->computeOffDiagJacobian(jvar->number());
2716 
2717  // Cache the jacobian block for the secondary side
2718  nec->addJacobian(_fe_problem.assembly(0, number()),
2719  nec->_Kee,
2720  secondary_dofs,
2721  nec->_connected_dof_indices,
2722  nec->variable().scalingFactor());
2723 
2724  // Cache the jacobian block for the primary side
2725  nec->addJacobian(_fe_problem.assembly(0, number()),
2726  nec->_Kne,
2727  nec->variable().dofIndicesNeighbor(),
2728  nec->_connected_dof_indices,
2729  nec->variable().scalingFactor());
2730 
2733  }
2734  }
2735  }
2736  }
2737  }
2738  }
2739  }
2740  }
2741  // See if constraints were applied anywhere
2742  _communicator.max(constraints_applied);
2743 
2744  if (constraints_applied)
2745  {
2746  LibmeshPetscCall(MatSetOption(static_cast<PetscMatrix<Number> &>(jacobian).mat(),
2747  MAT_KEEP_NONZERO_PATTERN, // This is changed in 3.1
2748  PETSC_TRUE));
2749 
2750  jacobian.close();
2751  jacobian.zero_rows(zero_rows, 0.0);
2752  jacobian.close();
2754  jacobian.close();
2755  }
2756 }
2757 
2758 void
2760 {
2761  MooseObjectWarehouse<ScalarKernelBase> * scalar_kernel_warehouse;
2762 
2763  if (!tags.size() || tags.size() == _fe_problem.numMatrixTags())
2764  scalar_kernel_warehouse = &_scalar_kernels;
2765  else if (tags.size() == 1)
2766  scalar_kernel_warehouse = &(_scalar_kernels.getMatrixTagObjectWarehouse(*(tags.begin()), 0));
2767  else
2768  scalar_kernel_warehouse = &(_scalar_kernels.getMatrixTagsObjectWarehouse(tags, 0));
2769 
2770  // Compute the diagonal block for scalar variables
2771  if (scalar_kernel_warehouse->hasActiveObjects())
2772  {
2773  const auto & scalars = scalar_kernel_warehouse->getActiveObjects();
2774 
2775  _fe_problem.reinitScalars(/*tid=*/0);
2776 
2777  _fe_problem.reinitOffDiagScalars(/*_tid*/ 0);
2778 
2779  bool have_scalar_contributions = false;
2780  for (const auto & kernel : scalars)
2781  {
2782  if (!kernel->computesJacobian())
2783  continue;
2784 
2785  kernel->reinit();
2786  const std::vector<dof_id_type> & dof_indices = kernel->variable().dofIndices();
2787  const DofMap & dof_map = kernel->variable().dofMap();
2788  const dof_id_type first_dof = dof_map.first_dof();
2789  const dof_id_type end_dof = dof_map.end_dof();
2790  for (dof_id_type dof : dof_indices)
2791  {
2792  if (dof >= first_dof && dof < end_dof)
2793  {
2794  kernel->computeJacobian();
2795  _fe_problem.addJacobianOffDiagScalar(kernel->variable().number());
2796  have_scalar_contributions = true;
2797  break;
2798  }
2799  }
2800  }
2801 
2802  if (have_scalar_contributions)
2804  }
2805 }
2806 
2807 void
2809 {
2811 
2812  for (THREAD_ID tid = 0; tid < libMesh::n_threads(); tid++)
2813  {
2814  _kernels.jacobianSetup(tid);
2817  if (_doing_dg)
2823  }
2830 
2831 #ifdef MOOSE_KOKKOS_ENABLED
2836 #endif
2837 
2838  // Avoid recursion
2839  if (this == &_fe_problem.currentNonlinearSystem())
2842 }
2843 
2844 void
2846 {
2847  TIME_SECTION("computeJacobianInternal", 3);
2848 
2850 
2851  // Make matrix ready to use
2853 
2854  for (auto tag : tags)
2855  {
2856  if (!hasMatrix(tag))
2857  continue;
2858 
2859  auto & jacobian = getMatrix(tag);
2860  // Necessary for speed
2861  if (auto petsc_matrix = dynamic_cast<PetscMatrix<Number> *>(&jacobian))
2862  {
2863  LibmeshPetscCall(MatSetOption(petsc_matrix->mat(),
2864  MAT_KEEP_NONZERO_PATTERN, // This is changed in 3.1
2865  PETSC_TRUE));
2867  LibmeshPetscCall(
2868  MatSetOption(petsc_matrix->mat(), MAT_NEW_NONZERO_ALLOCATION_ERR, PETSC_FALSE));
2870  LibmeshPetscCall(MatSetOption(static_cast<PetscMatrix<Number> &>(jacobian).mat(),
2871  MAT_IGNORE_ZERO_ENTRIES,
2872  PETSC_TRUE));
2873  }
2874  }
2875 
2876  jacobianSetup();
2877 
2878 #ifdef MOOSE_KOKKOS_ENABLED
2880  computeKokkosJacobian(tags);
2881 #endif
2882 
2883  // Jacobian contributions from UOs - for now this is used for ray tracing
2884  // and ray kernels that contribute to the Jacobian (think line sources)
2885  std::vector<UserObject *> uos;
2887  .query()
2888  .condition<AttribSystem>("UserObject")
2889  .condition<AttribExecOns>(EXEC_PRE_KERNELS)
2890  .queryInto(uos);
2891  for (auto & uo : uos)
2892  uo->jacobianSetup();
2893  for (auto & uo : uos)
2894  {
2895  uo->initialize();
2896  uo->execute();
2897  uo->finalize();
2898  }
2899 
2900  // reinit scalar variables
2901  for (unsigned int tid = 0; tid < libMesh::n_threads(); tid++)
2903 
2904  PARALLEL_TRY
2905  {
2906  // We would like to compute ScalarKernels, block NodalKernels, FVFluxKernels, and mortar objects
2907  // up front because we want these included whether we are computing an ordinary Jacobian or a
2908  // Jacobian for determining variable scaling factors
2910 
2911  // Block restricted Nodal Kernels
2913  {
2916  Threads::parallel_reduce(range, cnkjt);
2917 
2918  unsigned int n_threads = libMesh::n_threads();
2919  for (unsigned int i = 0; i < n_threads;
2920  i++) // Add any cached jacobians that might be hanging around
2922  }
2923 
2925  if (_fe_problem.haveFV())
2926  {
2927  // the same loop works for both residual and jacobians because it keys
2928  // off of FEProblem's _currently_computing_jacobian parameter
2930  _fe_problem, this->number(), tags, /*on_displaced=*/false);
2932  Threads::parallel_reduce(faces, fvj);
2933  }
2935  displaced_problem && displaced_problem->haveFV())
2936  {
2938  _fe_problem, this->number(), tags, /*on_displaced=*/true);
2939  FVRange faces(displaced_problem->mesh().ownedFaceInfoBegin(),
2940  displaced_problem->mesh().ownedFaceInfoEnd());
2941  Threads::parallel_reduce(faces, fvr);
2942  }
2943 
2945 
2946  // Get our element range for looping over
2948 
2950  {
2951  // Only compute Jacobians corresponding to the diagonals of volumetric compute objects
2952  // because this typically gives us a good representation of the physics. NodalBCs and
2953  // Constraints can introduce dramatically different scales (often order unity).
2954  // IntegratedBCs and/or InterfaceKernels may use penalty factors. DGKernels may be ok, but
2955  // they are almost always used in conjunction with Kernels
2957  Threads::parallel_reduce(elem_range, cj);
2958  unsigned int n_threads = libMesh::n_threads();
2959  for (unsigned int i = 0; i < n_threads;
2960  i++) // Add any Jacobian contributions still hanging around
2962 
2963  // Check whether any exceptions were thrown and propagate this information for parallel
2964  // consistency before
2965  // 1) we do parallel communication when closing tagged matrices
2966  // 2) early returning before reaching our PARALLEL_CATCH below
2968 
2969  closeTaggedMatrices(tags);
2970 
2971  return;
2972  }
2973 
2974  switch (_fe_problem.coupling())
2975  {
2976  case Moose::COUPLING_DIAG:
2977  {
2979  Threads::parallel_reduce(elem_range, cj);
2980 
2981  unsigned int n_threads = libMesh::n_threads();
2982  for (unsigned int i = 0; i < n_threads;
2983  i++) // Add any Jacobian contributions still hanging around
2985 
2986  // Boundary restricted Nodal Kernels
2988  {
2991 
2992  Threads::parallel_reduce(bnd_range, cnkjt);
2993  unsigned int n_threads = libMesh::n_threads();
2994  for (unsigned int i = 0; i < n_threads;
2995  i++) // Add any cached jacobians that might be hanging around
2997  }
2998  }
2999  break;
3000 
3001  default:
3003  {
3005  Threads::parallel_reduce(elem_range, cj);
3006  unsigned int n_threads = libMesh::n_threads();
3007 
3008  for (unsigned int i = 0; i < n_threads; i++)
3010 
3011  // Boundary restricted Nodal Kernels
3013  {
3016 
3017  Threads::parallel_reduce(bnd_range, cnkjt);
3018  unsigned int n_threads = libMesh::n_threads();
3019  for (unsigned int i = 0; i < n_threads;
3020  i++) // Add any cached jacobians that might be hanging around
3022  }
3023  }
3024  break;
3025  }
3026 
3027  computeDiracContributions(tags, true);
3028 
3029  static bool first = true;
3030 
3031  // This adds zeroes into geometric coupling entries to ensure they stay in the matrix
3032  if ((_fe_problem.restoreOriginalNonzeroPattern() || first) &&
3034  {
3035  first = false;
3037 
3040  }
3041  }
3042  PARALLEL_CATCH;
3043 
3044  // Have no idea how to have constraints work
3045  // with the tag system
3046  PARALLEL_TRY
3047  {
3048  // Add in Jacobian contributions from other Constraints
3049  if (_fe_problem._has_constraints && tags.count(systemMatrixTag()))
3050  {
3051  // Some constraints need to be able to read values from the Jacobian, which requires that it
3052  // be closed/assembled
3053  auto & system_matrix = getMatrix(systemMatrixTag());
3054 #if PETSC_RELEASE_GREATER_EQUALS(3, 23, 0)
3055  SparseMatrix<Number> * view_jac_ptr;
3056  std::unique_ptr<SparseMatrix<Number>> hash_copy;
3057  if (system_matrix.use_hash_table())
3058  {
3059  hash_copy = libMesh::cast_ref<PetscMatrix<Number> &>(system_matrix).copy_from_hash();
3060  view_jac_ptr = hash_copy.get();
3061  }
3062  else
3063  view_jac_ptr = &system_matrix;
3064  auto & jacobian_to_view = *view_jac_ptr;
3065 #else
3066  auto & jacobian_to_view = system_matrix;
3067 #endif
3068  if (&jacobian_to_view == &system_matrix)
3069  system_matrix.close();
3070 
3071  // Nodal Constraints
3073 
3074  // Undisplaced Constraints
3075  constraintJacobians(jacobian_to_view, false);
3076 
3077  // Displaced Constraints
3079  constraintJacobians(jacobian_to_view, true);
3080  }
3081  }
3082  PARALLEL_CATCH;
3083 
3084  // We need to close the save_in variables on the aux system before NodalBCBases clear the dofs
3085  // on boundary nodes
3086  if (_has_diag_save_in)
3088 
3089  PARALLEL_TRY
3090  {
3091  MooseObjectWarehouse<NodalBCBase> * nbc_warehouse;
3092  // Select nodal kernels
3093  if (tags.size() == _fe_problem.numMatrixTags() || !tags.size())
3094  nbc_warehouse = &_nodal_bcs;
3095  else if (tags.size() == 1)
3096  nbc_warehouse = &(_nodal_bcs.getMatrixTagObjectWarehouse(*(tags.begin()), 0));
3097  else
3098  nbc_warehouse = &(_nodal_bcs.getMatrixTagsObjectWarehouse(tags, 0));
3099 
3100  if (nbc_warehouse->hasActiveObjects())
3101  {
3102  // We may be switching from add to set. Moreover, we rely on a call to MatZeroRows to enforce
3103  // the nodal boundary condition constraints which requires that the matrix be truly assembled
3104  // as opposed to just flushed. Consequently we can't do the following despite any desire to
3105  // keep our initial sparsity pattern honored (see https://gitlab.com/petsc/petsc/-/issues/852)
3106  //
3107  // flushTaggedMatrices(tags);
3108  closeTaggedMatrices(tags);
3109 
3110  // Cache the information about which BCs are coupled to which
3111  // variables, so we don't have to figure it out for each node.
3112  std::map<std::string, std::set<unsigned int>> bc_involved_vars;
3113  const std::set<BoundaryID> & all_boundary_ids = _mesh.getBoundaryIDs();
3114  for (const auto & bid : all_boundary_ids)
3115  {
3116  // Get reference to all the NodalBCs for this ID. This is only
3117  // safe if there are NodalBCBases there to be gotten...
3118  if (nbc_warehouse->hasActiveBoundaryObjects(bid))
3119  {
3120  const auto & bcs = nbc_warehouse->getActiveBoundaryObjects(bid);
3121  for (const auto & bc : bcs)
3122  {
3123  const std::vector<MooseVariableFEBase *> & coupled_moose_vars =
3124  bc->getCoupledMooseVars();
3125 
3126  // Create the set of "involved" MOOSE nonlinear vars, which includes all coupled vars
3127  // and the BC's own variable
3128  std::set<unsigned int> & var_set = bc_involved_vars[bc->name()];
3129  for (const auto & coupled_var : coupled_moose_vars)
3130  if (coupled_var->kind() == Moose::VAR_SOLVER)
3131  var_set.insert(coupled_var->number());
3132 
3133  var_set.insert(bc->variable().number());
3134  }
3135  }
3136  }
3137 
3138  // reinit scalar variables again. This reinit does not re-fill any of the scalar variable
3139  // solution arrays because that was done above. It only will reorder the derivative
3140  // information for AD calculations to be suitable for NodalBC calculations
3141  for (unsigned int tid = 0; tid < libMesh::n_threads(); tid++)
3142  _fe_problem.reinitScalars(tid, true);
3143 
3144  // Get variable coupling list. We do all the NodalBCBase stuff on
3145  // thread 0... The couplingEntries() data structure determines
3146  // which variables are "coupled" as far as the preconditioner is
3147  // concerned, not what variables a boundary condition specifically
3148  // depends on.
3149  auto & coupling_entries = _fe_problem.couplingEntries(/*_tid=*/0, this->number());
3150 
3151  // Compute Jacobians for NodalBCBases
3153  for (const auto & bnode : bnd_nodes)
3154  {
3155  BoundaryID boundary_id = bnode->_bnd_id;
3156  Node * node = bnode->_node;
3157 
3158  if (nbc_warehouse->hasActiveBoundaryObjects(boundary_id) &&
3159  node->processor_id() == processor_id())
3160  {
3161  _fe_problem.reinitNodeFace(node, boundary_id, 0);
3162 
3163  const auto & bcs = nbc_warehouse->getActiveBoundaryObjects(boundary_id);
3164  for (const auto & bc : bcs)
3165  {
3166  // Get the set of involved MOOSE vars for this BC
3167  std::set<unsigned int> & var_set = bc_involved_vars[bc->name()];
3168 
3169  // Loop over all the variables whose Jacobian blocks are
3170  // actually being computed, call computeOffDiagJacobian()
3171  // for each one which is actually coupled (otherwise the
3172  // value is zero.)
3173  for (const auto & it : coupling_entries)
3174  {
3175  unsigned int ivar = it.first->number(), jvar = it.second->number();
3176 
3177  // We are only going to call computeOffDiagJacobian() if:
3178  // 1.) the BC's variable is ivar
3179  // 2.) jvar is "involved" with the BC (including jvar==ivar), and
3180  // 3.) the BC should apply.
3181  if ((bc->variable().number() == ivar) && var_set.count(jvar) && bc->shouldApply())
3182  bc->computeOffDiagJacobian(jvar);
3183  }
3184 
3185  const auto & coupled_scalar_vars = bc->getCoupledMooseScalarVars();
3186  for (const auto & jvariable : coupled_scalar_vars)
3187  if (hasScalarVariable(jvariable->name()))
3188  bc->computeOffDiagJacobianScalar(jvariable->number());
3189  }
3190  }
3191  } // end loop over boundary nodes
3192 
3193  // Set the cached NodalBCBase values in the Jacobian matrix
3195  }
3196  }
3197  PARALLEL_CATCH;
3198 
3199  closeTaggedMatrices(tags);
3200 
3201  // We need to close the save_in variables on the aux system before NodalBCBases clear the dofs
3202  // on boundary nodes
3205 
3206  if (hasDiagSaveIn())
3208 
3209  // Accumulate the occurrence of solution invalid warnings for the current iteration cumulative
3210  // counters
3213 }
3214 
3215 void
3217 {
3218  _nl_matrix_tags.clear();
3219 
3220  auto & tags = _fe_problem.getMatrixTags();
3221 
3222  for (auto & tag : tags)
3223  _nl_matrix_tags.insert(tag.second);
3224 
3225  computeJacobian(jacobian, _nl_matrix_tags);
3226 }
3227 
3228 void
3229 NonlinearSystemBase::computeJacobian(SparseMatrix<Number> & jacobian, const std::set<TagID> & tags)
3230 {
3232 
3233  computeJacobianTags(tags);
3234 
3236 }
3237 
3238 void
3239 NonlinearSystemBase::computeJacobianTags(const std::set<TagID> & tags)
3240 {
3241  TIME_SECTION("computeJacobianTags", 5);
3242 
3243  FloatingPointExceptionGuard fpe_guard(_app);
3244 
3245  try
3246  {
3248  }
3249  catch (MooseException & e)
3250  {
3251  // The buck stops here, we have already handled the exception by
3252  // calling stopSolve(), it is now up to PETSc to return a
3253  // "diverged" reason during the next solve.
3254  }
3255 }
3256 
3257 void
3259 {
3260  _nl_matrix_tags.clear();
3261 
3262  auto & tags = _fe_problem.getMatrixTags();
3263  for (auto & tag : tags)
3264  _nl_matrix_tags.insert(tag.second);
3265 
3267 }
3268 
3269 void
3271  const std::set<TagID> & tags)
3272 {
3273  TIME_SECTION("computeJacobianBlocks", 3);
3274  FloatingPointExceptionGuard fpe_guard(_app);
3275 
3276  for (unsigned int i = 0; i < blocks.size(); i++)
3277  {
3278  SparseMatrix<Number> & jacobian = blocks[i]->_jacobian;
3279 
3280  LibmeshPetscCall(MatSetOption(static_cast<PetscMatrix<Number> &>(jacobian).mat(),
3281  MAT_KEEP_NONZERO_PATTERN, // This is changed in 3.1
3282  PETSC_TRUE));
3284  LibmeshPetscCall(MatSetOption(static_cast<PetscMatrix<Number> &>(jacobian).mat(),
3285  MAT_NEW_NONZERO_ALLOCATION_ERR,
3286  PETSC_TRUE));
3287 
3288  jacobian.zero();
3289  }
3290 
3291  for (unsigned int tid = 0; tid < libMesh::n_threads(); tid++)
3293 
3294  PARALLEL_TRY
3295  {
3298  Threads::parallel_reduce(elem_range, cjb);
3299  }
3300  PARALLEL_CATCH;
3301 
3302  for (unsigned int i = 0; i < blocks.size(); i++)
3303  blocks[i]->_jacobian.close();
3304 
3305  for (unsigned int i = 0; i < blocks.size(); i++)
3306  {
3307  libMesh::System & precond_system = blocks[i]->_precond_system;
3308  SparseMatrix<Number> & jacobian = blocks[i]->_jacobian;
3309 
3310  unsigned int ivar = blocks[i]->_ivar;
3311  unsigned int jvar = blocks[i]->_jvar;
3312 
3313  // Dirichlet BCs
3314  std::vector<numeric_index_type> zero_rows;
3315  PARALLEL_TRY
3316  {
3318  for (const auto & bnode : bnd_nodes)
3319  {
3320  BoundaryID boundary_id = bnode->_bnd_id;
3321  Node * node = bnode->_node;
3322 
3323  if (_nodal_bcs.hasActiveBoundaryObjects(boundary_id))
3324  {
3325  const auto & bcs = _nodal_bcs.getActiveBoundaryObjects(boundary_id);
3326 
3327  if (node->processor_id() == processor_id())
3328  {
3329  _fe_problem.reinitNodeFace(node, boundary_id, 0);
3330 
3331  for (const auto & bc : bcs)
3332  if (bc->variable().number() == ivar && bc->shouldApply())
3333  {
3334  // The first zero is for the variable number... there is only one variable in
3335  // each mini-system The second zero only works with Lagrange elements!
3336  zero_rows.push_back(node->dof_number(precond_system.number(), 0, 0));
3337  }
3338  }
3339  }
3340  }
3341  }
3342  PARALLEL_CATCH;
3343 
3344  jacobian.close();
3345 
3346  // This zeroes the rows corresponding to Dirichlet BCs and puts a 1.0 on the diagonal
3347  if (ivar == jvar)
3348  jacobian.zero_rows(zero_rows, 1.0);
3349  else
3350  jacobian.zero_rows(zero_rows, 0.0);
3351 
3352  jacobian.close();
3353  }
3354 }
3355 
3356 void
3358 {
3365  _kernels.updateActive(tid);
3367 
3368  if (tid == 0)
3369  {
3377 
3378 #ifdef MOOSE_KOKKOS_ENABLED
3384 #endif
3385  }
3386 }
3387 
3388 Real
3390  const NumericVector<Number> & update)
3391 {
3392  // Default to no damping
3393  Real damping = 1.0;
3394  bool has_active_dampers = false;
3395 
3396  try
3397  {
3399  {
3400  PARALLEL_TRY
3401  {
3402  TIME_SECTION("computeDampers", 3, "Computing Dampers");
3403  has_active_dampers = true;
3406  Threads::parallel_reduce(_fe_problem.getCurrentAlgebraicElementRange(), cid);
3407  damping = std::min(cid.damping(), damping);
3408  }
3409  PARALLEL_CATCH;
3410  }
3411 
3413  {
3414  PARALLEL_TRY
3415  {
3416  TIME_SECTION("computeDamping::element", 3, "Computing Element Damping");
3417 
3418  has_active_dampers = true;
3421  Threads::parallel_reduce(_fe_problem.getCurrentAlgebraicNodeRange(), cndt);
3422  damping = std::min(cndt.damping(), damping);
3423  }
3424  PARALLEL_CATCH;
3425  }
3426 
3428  {
3429  PARALLEL_TRY
3430  {
3431  TIME_SECTION("computeDamping::general", 3, "Computing General Damping");
3432 
3433  has_active_dampers = true;
3434  const auto & gdampers = _general_dampers.getActiveObjects();
3435  for (const auto & damper : gdampers)
3436  {
3437  Real gd_damping = damper->computeDamping(solution, update);
3438  try
3439  {
3440  damper->checkMinDamping(gd_damping);
3441  }
3442  catch (MooseException & e)
3443  {
3445  }
3446  damping = std::min(gd_damping, damping);
3447  }
3448  }
3449  PARALLEL_CATCH;
3450  }
3451  }
3452  catch (MooseException & e)
3453  {
3454  // The buck stops here, we have already handled the exception by
3455  // calling stopSolve(), it is now up to PETSc to return a
3456  // "diverged" reason during the next solve.
3457  }
3458  catch (std::exception & e)
3459  {
3460  // Allow the libmesh error/exception on negative jacobian
3461  const std::string & message = e.what();
3462  if (message.find("Jacobian") == std::string::npos)
3463  throw e;
3464  }
3465 
3466  _communicator.min(damping);
3467 
3468  if (has_active_dampers && damping < 1.0)
3469  _console << " Damping factor: " << damping << std::endl;
3470 
3471  return damping;
3472 }
3473 
3474 void
3475 NonlinearSystemBase::computeDiracContributions(const std::set<TagID> & tags, bool is_jacobian)
3476 {
3478 
3479  std::set<const Elem *> dirac_elements;
3480 
3482  {
3483  TIME_SECTION("computeDirac", 3, "Computing DiracKernels");
3484 
3485  // TODO: Need a threading fix... but it's complicated!
3486  for (THREAD_ID tid = 0; tid < libMesh::n_threads(); ++tid)
3487  {
3488  const auto & dkernels = _dirac_kernels.getActiveObjects(tid);
3489  for (const auto & dkernel : dkernels)
3490  {
3491  dkernel->clearPoints();
3492  dkernel->addPoints();
3493  }
3494  }
3495 
3496  ComputeDiracThread cd(_fe_problem, tags, is_jacobian);
3497 
3498  _fe_problem.getDiracElements(dirac_elements);
3499 
3500  DistElemRange range(dirac_elements.begin(), dirac_elements.end(), 1);
3501  // TODO: Make Dirac work thread!
3502  // Threads::parallel_reduce(range, cd);
3503 
3504  cd(range);
3505  }
3506 }
3507 
3510 {
3511  if (!_residual_copy.get())
3513 
3514  return *_residual_copy;
3515 }
3516 
3519 {
3520  _need_residual_ghosted = true;
3521  if (!_residual_ghosted)
3522  {
3523  // The first time we realize we need a ghosted residual vector,
3524  // we add it.
3525  _residual_ghosted = &addVector("residual_ghosted", false, GHOSTED);
3526 
3527  // If we've already realized we need time and/or non-time
3528  // residual vectors, but we haven't yet realized they need to be
3529  // ghosted, fix that now.
3530  //
3531  // If an application changes its mind, the libMesh API lets us
3532  // change the vector.
3533  if (_Re_time)
3534  {
3535  const auto vector_name = _subproblem.vectorTagName(_Re_time_tag);
3536  _Re_time = &system().add_vector(vector_name, false, GHOSTED);
3537  }
3538  if (_Re_non_time)
3539  {
3540  const auto vector_name = _subproblem.vectorTagName(_Re_non_time_tag);
3541  _Re_non_time = &system().add_vector(vector_name, false, GHOSTED);
3542  }
3543  }
3544  return *_residual_ghosted;
3545 }
3546 
3547 void
3549  std::vector<dof_id_type> & n_nz,
3550  std::vector<dof_id_type> & n_oz)
3551 {
3553  {
3555 
3556  std::unordered_map<dof_id_type, std::vector<dof_id_type>> graph;
3557 
3559 
3562  graph);
3563 
3564  const dof_id_type first_dof_on_proc = dofMap().first_dof(processor_id());
3565  const dof_id_type end_dof_on_proc = dofMap().end_dof(processor_id());
3566 
3567  // The total number of dofs on and off processor
3568  const dof_id_type n_dofs_on_proc = dofMap().n_local_dofs();
3569  const dof_id_type n_dofs_not_on_proc = dofMap().n_dofs() - dofMap().n_local_dofs();
3570 
3571  for (const auto & git : graph)
3572  {
3573  dof_id_type dof = git.first;
3574  dof_id_type local_dof = dof - first_dof_on_proc;
3575 
3576  if (dof < first_dof_on_proc || dof >= end_dof_on_proc)
3577  continue;
3578 
3579  const auto & row = git.second;
3580 
3581  SparsityPattern::Row & sparsity_row = sparsity[local_dof];
3582 
3583  unsigned int original_row_length = sparsity_row.size();
3584 
3585  sparsity_row.insert(sparsity_row.end(), row.begin(), row.end());
3586 
3587  SparsityPattern::sort_row(
3588  sparsity_row.begin(), sparsity_row.begin() + original_row_length, sparsity_row.end());
3589 
3590  // Fix up nonzero arrays
3591  for (const auto & coupled_dof : row)
3592  {
3593  if (coupled_dof < first_dof_on_proc || coupled_dof >= end_dof_on_proc)
3594  {
3595  if (n_oz[local_dof] < n_dofs_not_on_proc)
3596  n_oz[local_dof]++;
3597  }
3598  else
3599  {
3600  if (n_nz[local_dof] < n_dofs_on_proc)
3601  n_nz[local_dof]++;
3602  }
3603  }
3604  }
3605  }
3606 }
3607 
3608 void
3610 {
3611  *_u_dot = u_dot;
3612 }
3613 
3614 void
3616 {
3617  *_u_dotdot = u_dotdot;
3618 }
3619 
3620 void
3622 {
3623  *_u_dot_old = u_dot_old;
3624 }
3625 
3626 void
3628 {
3629  *_u_dotdot_old = u_dotdot_old;
3630 }
3631 
3632 void
3633 NonlinearSystemBase::setPreconditioner(std::shared_ptr<MoosePreconditioner> pc)
3634 {
3635  if (_preconditioner.get() != nullptr)
3636  mooseError("More than one active Preconditioner detected");
3637 
3638  _preconditioner = pc;
3639 }
3640 
3641 MoosePreconditioner const *
3643 {
3644  return _preconditioner.get();
3645 }
3646 
3647 void
3649 {
3650  _increment_vec = &_sys.add_vector("u_increment", true, GHOSTED);
3651 }
3652 
3653 void
3655  const std::set<MooseVariable *> & damped_vars)
3656 {
3657  for (const auto & var : damped_vars)
3658  var->computeIncrementAtQps(*_increment_vec);
3659 }
3660 
3661 void
3663  const std::set<MooseVariable *> & damped_vars)
3664 {
3665  for (const auto & var : damped_vars)
3666  var->computeIncrementAtNode(*_increment_vec);
3667 }
3668 
3669 void
3670 NonlinearSystemBase::checkKernelCoverage(const std::set<SubdomainID> & mesh_subdomains) const
3671 {
3672  // Obtain all blocks and variables covered by all kernels
3673  std::set<SubdomainID> input_subdomains;
3674  std::set<std::string> kernel_variables;
3675 
3676  bool global_kernels_exist = false;
3677  global_kernels_exist |= _scalar_kernels.hasActiveObjects();
3678  global_kernels_exist |= _nodal_kernels.hasActiveObjects();
3679 
3680  _kernels.subdomainsCovered(input_subdomains, kernel_variables);
3681  _dg_kernels.subdomainsCovered(input_subdomains, kernel_variables);
3682  _nodal_kernels.subdomainsCovered(input_subdomains, kernel_variables);
3683  _scalar_kernels.subdomainsCovered(input_subdomains, kernel_variables);
3684  _constraints.subdomainsCovered(input_subdomains, kernel_variables);
3685 
3686 #ifdef MOOSE_KOKKOS_ENABLED
3687  _kokkos_kernels.subdomainsCovered(input_subdomains, kernel_variables);
3688  _kokkos_nodal_kernels.subdomainsCovered(input_subdomains, kernel_variables);
3689 #endif
3690 
3691  if (_fe_problem.haveFV())
3692  {
3693  std::vector<FVElementalKernel *> fv_elemental_kernels;
3695  .query()
3696  .template condition<AttribSystem>("FVElementalKernel")
3697  .queryInto(fv_elemental_kernels);
3698 
3699  for (auto fv_kernel : fv_elemental_kernels)
3700  {
3701  if (fv_kernel->blockRestricted())
3702  for (auto block_id : fv_kernel->blockIDs())
3703  input_subdomains.insert(block_id);
3704  else
3705  global_kernels_exist = true;
3706  kernel_variables.insert(fv_kernel->variable().name());
3707 
3708  // Check for lagrange multiplier
3709  if (dynamic_cast<FVScalarLagrangeMultiplierConstraint *>(fv_kernel))
3710  kernel_variables.insert(dynamic_cast<FVScalarLagrangeMultiplierConstraint *>(fv_kernel)
3711  ->lambdaVariable()
3712  .name());
3713  }
3714 
3715  std::vector<FVFluxKernel *> fv_flux_kernels;
3717  .query()
3718  .template condition<AttribSystem>("FVFluxKernel")
3719  .queryInto(fv_flux_kernels);
3720 
3721  for (auto fv_kernel : fv_flux_kernels)
3722  {
3723  if (fv_kernel->blockRestricted())
3724  for (auto block_id : fv_kernel->blockIDs())
3725  input_subdomains.insert(block_id);
3726  else
3727  global_kernels_exist = true;
3728  kernel_variables.insert(fv_kernel->variable().name());
3729  }
3730 
3731  std::vector<FVInterfaceKernel *> fv_interface_kernels;
3733  .query()
3734  .template condition<AttribSystem>("FVInterfaceKernel")
3735  .queryInto(fv_interface_kernels);
3736 
3737  for (auto fvik : fv_interface_kernels)
3738  if (auto scalar_fvik = dynamic_cast<FVScalarLagrangeMultiplierInterface *>(fvik))
3739  kernel_variables.insert(scalar_fvik->lambdaVariable().name());
3740 
3741  std::vector<FVFluxBC *> fv_flux_bcs;
3743  .query()
3744  .template condition<AttribSystem>("FVFluxBC")
3745  .queryInto(fv_flux_bcs);
3746 
3747  for (auto fvbc : fv_flux_bcs)
3748  if (auto scalar_fvbc = dynamic_cast<FVBoundaryScalarLagrangeMultiplierConstraint *>(fvbc))
3749  kernel_variables.insert(scalar_fvbc->lambdaVariable().name());
3750  }
3751 
3752  // Check kernel coverage of subdomains (blocks) in your mesh
3753  if (!global_kernels_exist)
3754  {
3755  std::set<SubdomainID> difference;
3756  std::set_difference(mesh_subdomains.begin(),
3757  mesh_subdomains.end(),
3758  input_subdomains.begin(),
3759  input_subdomains.end(),
3760  std::inserter(difference, difference.end()));
3761 
3762  // there supposed to be no kernels on this lower-dimensional subdomain
3763  for (const auto & id : _mesh.interiorLowerDBlocks())
3764  difference.erase(id);
3765  for (const auto & id : _mesh.boundaryLowerDBlocks())
3766  difference.erase(id);
3767 
3768  if (!difference.empty())
3769  {
3770  std::vector<SubdomainID> difference_vec =
3771  std::vector<SubdomainID>(difference.begin(), difference.end());
3772  std::vector<SubdomainName> difference_names = _mesh.getSubdomainNames(difference_vec);
3773  std::stringstream missing_block_names;
3774  std::copy(difference_names.begin(),
3775  difference_names.end(),
3776  std::ostream_iterator<std::string>(missing_block_names, " "));
3777  std::stringstream missing_block_ids;
3778  std::copy(difference.begin(),
3779  difference.end(),
3780  std::ostream_iterator<unsigned int>(missing_block_ids, " "));
3781 
3782  mooseError("Each subdomain must contain at least one Kernel.\nThe following block(s) lack an "
3783  "active kernel: " +
3784  missing_block_names.str(),
3785  " (ids: ",
3786  missing_block_ids.str(),
3787  ")");
3788  }
3789  }
3790 
3791  // Check kernel use of variables
3792  std::set<VariableName> variables(getVariableNames().begin(), getVariableNames().end());
3793 
3794  std::set<VariableName> difference;
3795  std::set_difference(variables.begin(),
3796  variables.end(),
3797  kernel_variables.begin(),
3798  kernel_variables.end(),
3799  std::inserter(difference, difference.end()));
3800 
3801  // skip checks for varaibles defined on lower-dimensional subdomain
3802  std::set<VariableName> vars(difference);
3803  for (auto & var_name : vars)
3804  {
3805  auto blks = getSubdomainsForVar(var_name);
3806  for (const auto & id : blks)
3807  if (_mesh.interiorLowerDBlocks().count(id) > 0 || _mesh.boundaryLowerDBlocks().count(id) > 0)
3808  difference.erase(var_name);
3809  }
3810 
3811  if (!difference.empty())
3812  {
3813  std::stringstream missing_kernel_vars;
3814  std::copy(difference.begin(),
3815  difference.end(),
3816  std::ostream_iterator<std::string>(missing_kernel_vars, " "));
3817  mooseError("Each variable must be referenced by at least one active Kernel.\nThe following "
3818  "variable(s) lack an active kernel: " +
3819  missing_kernel_vars.str());
3820  }
3821 }
3822 
3823 bool
3825 {
3826  auto & time_kernels = _kernels.getVectorTagObjectWarehouse(timeVectorTag(), 0);
3827 
3828  return time_kernels.hasActiveObjects();
3829 }
3830 
3831 std::vector<std::string>
3833 {
3834  std::vector<std::string> variable_names;
3835  const auto & time_kernels = _kernels.getVectorTagObjectWarehouse(timeVectorTag(), 0);
3836  if (time_kernels.hasActiveObjects())
3837  for (const auto & kernel : time_kernels.getObjects())
3838  variable_names.push_back(kernel->variable().name());
3839 
3840  return variable_names;
3841 }
3842 
3843 bool
3845 {
3846  // IntegratedBCs are for now the only objects we consider to be consuming
3847  // matprops on boundaries.
3848  if (_integrated_bcs.hasActiveBoundaryObjects(bnd_id, tid))
3849  for (const auto & bc : _integrated_bcs.getActiveBoundaryObjects(bnd_id, tid))
3850  if (std::static_pointer_cast<MaterialPropertyInterface>(bc)->getMaterialPropertyCalled())
3851  return true;
3852 
3853  // Thin layer heat transfer in the heat_transfer module is being used on a boundary even though
3854  // it's an interface kernel. That boundary is external, on both sides of a gap in a mesh
3856  for (const auto & ik : _interface_kernels.getActiveBoundaryObjects(bnd_id, tid))
3857  if (std::static_pointer_cast<MaterialPropertyInterface>(ik)->getMaterialPropertyCalled())
3858  return true;
3859 
3860  // Because MortarConstraints do not inherit from BoundaryRestrictable, they are not sorted
3861  // by boundary in the MooseObjectWarehouse. So for now, we return true for all boundaries
3862  // Note: constraints are not threaded at this time
3863  if (_constraints.hasActiveObjects(/*tid*/ 0))
3864  for (const auto & ct : _constraints.getActiveObjects(/*tid*/ 0))
3865  if (auto mpi = std::dynamic_pointer_cast<MaterialPropertyInterface>(ct);
3866  mpi && mpi->getMaterialPropertyCalled())
3867  return true;
3868  return false;
3869 }
3870 
3871 bool
3873 {
3874  // InterfaceKernels are for now the only objects we consider to be consuming matprops on internal
3875  // boundaries.
3877  for (const auto & ik : _interface_kernels.getActiveBoundaryObjects(bnd_id, tid))
3878  if (std::static_pointer_cast<MaterialPropertyInterface>(ik)->getMaterialPropertyCalled())
3879  return true;
3880  return false;
3881 }
3882 
3883 bool
3885 {
3886  // DGKernels are for now the only objects we consider to be consuming matprops on
3887  // internal sides.
3888  if (_dg_kernels.hasActiveBlockObjects(subdomain_id, tid))
3889  for (const auto & dg : _dg_kernels.getActiveBlockObjects(subdomain_id, tid))
3890  if (std::static_pointer_cast<MaterialPropertyInterface>(dg)->getMaterialPropertyCalled())
3891  return true;
3892  // NOTE:
3893  // HDG kernels do not require face material properties on internal sides at this time.
3894  // The idea is to have element locality of HDG for hybridization
3895  return false;
3896 }
3897 
3898 bool
3900 {
3901  return _doing_dg;
3902 }
3903 
3904 void
3906 {
3909 }
3910 
3911 void
3913  const std::set<TagID> & vector_tags,
3914  const std::set<TagID> & matrix_tags)
3915 {
3916  parallel_object_only();
3917 
3918  try
3919  {
3920  for (auto & map_pr : _undisplaced_mortar_functors)
3921  map_pr.second(compute_type, vector_tags, matrix_tags);
3922 
3923  for (auto & map_pr : _displaced_mortar_functors)
3924  map_pr.second(compute_type, vector_tags, matrix_tags);
3925  }
3926  catch (MetaPhysicL::LogicError &)
3927  {
3928  mooseError(
3929  "We caught a MetaPhysicL error in NonlinearSystemBase::mortarConstraints. This is very "
3930  "likely due to AD not having a sufficiently large derivative container size. Please run "
3931  "MOOSE configure with the '--with-derivative-size=<n>' option");
3932  }
3933 }
3934 
3935 void
3937 {
3938  if (_auto_scaling_initd)
3939  return;
3940 
3941  // Want the libMesh count of variables, not MOOSE, e.g. I don't care about array variable counts
3942  const auto n_vars = system().n_vars();
3943 
3944  if (_scaling_group_variables.empty())
3945  {
3946  _var_to_group_var.reserve(n_vars);
3948 
3949  for (const auto var_number : make_range(n_vars))
3950  _var_to_group_var.emplace(var_number, var_number);
3951  }
3952  else
3953  {
3954  std::set<unsigned int> var_numbers, var_numbers_covered, var_numbers_not_covered;
3955  for (const auto var_number : make_range(n_vars))
3956  var_numbers.insert(var_number);
3957 
3959 
3960  for (const auto group_index : index_range(_scaling_group_variables))
3961  for (const auto & var_name : _scaling_group_variables[group_index])
3962  {
3963  if (!hasVariable(var_name) && !hasScalarVariable(var_name))
3964  mooseError("'",
3965  var_name,
3966  "', provided to the 'scaling_group_variables' parameter, does not exist in "
3967  "the nonlinear system.");
3968 
3969  const MooseVariableBase & var =
3970  hasVariable(var_name)
3971  ? static_cast<MooseVariableBase &>(getVariable(0, var_name))
3972  : static_cast<MooseVariableBase &>(getScalarVariable(0, var_name));
3973  auto map_pair = _var_to_group_var.emplace(var.number(), group_index);
3974  if (!map_pair.second)
3975  mooseError("Variable ", var_name, " is contained in multiple scaling grouplings");
3976  var_numbers_covered.insert(var.number());
3977  }
3978 
3979  std::set_difference(var_numbers.begin(),
3980  var_numbers.end(),
3981  var_numbers_covered.begin(),
3982  var_numbers_covered.end(),
3983  std::inserter(var_numbers_not_covered, var_numbers_not_covered.begin()));
3984 
3985  _num_scaling_groups = _scaling_group_variables.size() + var_numbers_not_covered.size();
3986 
3987  auto index = static_cast<unsigned int>(_scaling_group_variables.size());
3988  for (auto var_number : var_numbers_not_covered)
3989  _var_to_group_var.emplace(var_number, index++);
3990  }
3991 
3992  _variable_autoscaled.resize(n_vars, true);
3993  const auto & number_to_var_map = _vars[0].numberToVariableMap();
3994 
3996  for (const auto i : index_range(_variable_autoscaled))
3999  libmesh_map_find(number_to_var_map, i)->name()) !=
4001  _variable_autoscaled[i] = false;
4002 
4003  _auto_scaling_initd = true;
4004 }
4005 
4006 bool
4008 {
4010  return true;
4011 
4012  _console << "\nPerforming automatic scaling calculation\n" << std::endl;
4013 
4014  TIME_SECTION("computeScaling", 3, "Computing Automatic Scaling");
4015 
4016  // It's funny but we need to assemble our vector of scaling factors here otherwise we will be
4017  // applying scaling factors of 0 during Assembly of our scaling Jacobian
4019 
4020  // container for repeated access of element global dof indices
4021  std::vector<dof_id_type> dof_indices;
4022 
4023  if (!_auto_scaling_initd)
4024  setupScalingData();
4025 
4026  std::vector<Real> inverse_scaling_factors(_num_scaling_groups, 0);
4027  std::vector<Real> resid_inverse_scaling_factors(_num_scaling_groups, 0);
4028  std::vector<Real> jac_inverse_scaling_factors(_num_scaling_groups, 0);
4029  auto & dof_map = dofMap();
4030 
4031  // what types of scaling do we want?
4032  bool jac_scaling = _resid_vs_jac_scaling_param < 1. - TOLERANCE;
4033  bool resid_scaling = _resid_vs_jac_scaling_param > TOLERANCE;
4034 
4035  const NumericVector<Number> & scaling_residual = RHS();
4036 
4037  if (jac_scaling)
4038  {
4039  // if (!_auto_scaling_initd)
4040  // We need to reinit this when the number of dofs changes
4041  // but there is no good way to track that
4042  // In theory, it is the job of libmesh system to track this,
4043  // but this special matrix is not owned by libMesh system
4044  // Let us reinit eveytime since it is not expensive
4045  {
4046  auto init_vector = NumericVector<Number>::build(this->comm());
4047  init_vector->init(system().n_dofs(), system().n_local_dofs(), /*fast=*/false, PARALLEL);
4048 
4049  _scaling_matrix->clear();
4050  _scaling_matrix->init(*init_vector);
4051  }
4052 
4054  // Dispatch to derived classes to ensure that we use the correct matrix tag
4057  }
4058 
4059  if (resid_scaling)
4060  {
4063  // Dispatch to derived classes to ensure that we use the correct vector tag
4067  }
4068 
4069  // Did something bad happen during residual/Jacobian scaling computation?
4071  return false;
4072 
4073  auto examine_dof_indices = [this,
4074  jac_scaling,
4075  resid_scaling,
4076  &dof_map,
4077  &jac_inverse_scaling_factors,
4078  &resid_inverse_scaling_factors,
4079  &scaling_residual](const auto & dof_indices, const auto var_number)
4080  {
4081  for (auto dof_index : dof_indices)
4082  if (dof_map.local_index(dof_index))
4083  {
4084  if (jac_scaling)
4085  {
4086  // For now we will use the diagonal for determining scaling
4087  auto mat_value = (*_scaling_matrix)(dof_index, dof_index);
4088  auto & factor = jac_inverse_scaling_factors[_var_to_group_var[var_number]];
4089  factor = std::max(factor, std::abs(mat_value));
4090  }
4091  if (resid_scaling)
4092  {
4093  auto vec_value = scaling_residual(dof_index);
4094  auto & factor = resid_inverse_scaling_factors[_var_to_group_var[var_number]];
4095  factor = std::max(factor, std::abs(vec_value));
4096  }
4097  }
4098  };
4099 
4100  // Compute our scaling factors for the spatial field variables
4101  for (const auto & elem : _fe_problem.getCurrentAlgebraicElementRange())
4102  for (const auto i : make_range(system().n_vars()))
4104  {
4105  dof_map.dof_indices(elem, dof_indices, i);
4106  examine_dof_indices(dof_indices, i);
4107  }
4108 
4109  for (const auto i : make_range(system().n_vars()))
4110  if (_variable_autoscaled[i] && system().variable_type(i).family == SCALAR)
4111  {
4112  dof_map.SCALAR_dof_indices(dof_indices, i);
4113  examine_dof_indices(dof_indices, i);
4114  }
4115 
4116  if (resid_scaling)
4117  _communicator.max(resid_inverse_scaling_factors);
4118  if (jac_scaling)
4119  _communicator.max(jac_inverse_scaling_factors);
4120 
4121  if (jac_scaling && resid_scaling)
4122  for (MooseIndex(inverse_scaling_factors) i = 0; i < inverse_scaling_factors.size(); ++i)
4123  {
4124  // Be careful not to take log(0)
4125  if (!resid_inverse_scaling_factors[i])
4126  {
4127  if (!jac_inverse_scaling_factors[i])
4128  inverse_scaling_factors[i] = 1;
4129  else
4130  inverse_scaling_factors[i] = jac_inverse_scaling_factors[i];
4131  }
4132  else if (!jac_inverse_scaling_factors[i])
4133  // We know the resid is not zero
4134  inverse_scaling_factors[i] = resid_inverse_scaling_factors[i];
4135  else
4136  inverse_scaling_factors[i] =
4137  std::exp(_resid_vs_jac_scaling_param * std::log(resid_inverse_scaling_factors[i]) +
4138  (1 - _resid_vs_jac_scaling_param) * std::log(jac_inverse_scaling_factors[i]));
4139  }
4140  else if (jac_scaling)
4141  inverse_scaling_factors = jac_inverse_scaling_factors;
4142  else if (resid_scaling)
4143  inverse_scaling_factors = resid_inverse_scaling_factors;
4144  else
4145  mooseError("We shouldn't be calling this routine if we're not performing any scaling");
4146 
4147  // We have to make sure that our scaling values are not zero
4148  for (auto & scaling_factor : inverse_scaling_factors)
4149  if (scaling_factor == 0)
4150  scaling_factor = 1;
4151 
4152  // Now flatten the group scaling factors to the individual variable scaling factors
4153  std::vector<Real> flattened_inverse_scaling_factors(system().n_vars());
4154  for (const auto i : index_range(flattened_inverse_scaling_factors))
4155  flattened_inverse_scaling_factors[i] = inverse_scaling_factors[_var_to_group_var[i]];
4156 
4157  // Now set the scaling factors for the variables
4158  applyScalingFactors(flattened_inverse_scaling_factors);
4160  displaced_problem->systemBaseNonlinear(number()).applyScalingFactors(
4161  flattened_inverse_scaling_factors);
4162 
4163  _computed_scaling = true;
4164  return true;
4165 }
4166 
4167 void
4169 {
4170  if (!hasVector("scaling_factors"))
4171  // No variables have indicated they need scaling
4172  return;
4173 
4174  auto & scaling_vector = getVector("scaling_factors");
4175 
4176  const auto & lm_mesh = _mesh.getMesh();
4177  const auto & dof_map = dofMap();
4178 
4179  const auto & field_variables = _vars[0].fieldVariables();
4180  const auto & scalar_variables = _vars[0].scalars();
4181 
4182  std::vector<dof_id_type> dof_indices;
4183 
4184  for (const Elem * const elem :
4185  as_range(lm_mesh.active_local_elements_begin(), lm_mesh.active_local_elements_end()))
4186  for (const auto * const field_var : field_variables)
4187  {
4188  const auto & factors = field_var->arrayScalingFactor();
4189  for (const auto i : make_range(field_var->count()))
4190  {
4191  dof_map.dof_indices(elem, dof_indices, field_var->number() + i);
4192  for (const auto dof : dof_indices)
4193  scaling_vector.set(dof, factors[i]);
4194  }
4195  }
4196 
4197  for (const auto * const scalar_var : scalar_variables)
4198  {
4199  mooseAssert(scalar_var->count() == 1,
4200  "Scalar variables should always have only one component.");
4201  dof_map.SCALAR_dof_indices(dof_indices, scalar_var->number());
4202  for (const auto dof : dof_indices)
4203  scaling_vector.set(dof, scalar_var->scalingFactor());
4204  }
4205 
4206  // Parallel assemble
4207  scaling_vector.close();
4208 
4209  if (auto * displaced_problem = _fe_problem.getDisplacedProblem().get())
4210  // copy into the corresponding displaced system vector because they should be the exact same
4211  displaced_problem->systemBaseNonlinear(number()).getVector("scaling_factors") = scaling_vector;
4212 }
4213 
4214 bool
4216 {
4217  // Clear the iteration counters
4218  _current_l_its.clear();
4219  _current_nl_its = 0;
4220 
4221  // Initialize the solution vector using a predictor and known values from nodal bcs
4223 
4224  // Now that the initial solution has ben set, potentially perform a residual/Jacobian evaluation
4225  // to determine variable scaling factors
4226  if (_automatic_scaling)
4227  {
4228  const bool scaling_succeeded = computeScaling();
4229  if (!scaling_succeeded)
4230  return false;
4231  }
4232 
4233  // We do not know a priori what variable a global degree of freedom corresponds to, so we need a
4234  // map from global dof to scaling factor. We just use a ghosted NumericVector for that mapping
4236 
4237  return true;
4238 }
4239 
4240 void
4242 {
4243  if (matrixFromColoring())
4244  LibmeshPetscCall(MatFDColoringDestroy(&_fdcoloring));
4245 }
4246 
4249 {
4250  if (!_fsp)
4251  mooseError("No field split preconditioner is present for this system");
4252 
4253  return *_fsp;
4254 }
std::string name(const ElemQuality q)
std::vector< std::shared_ptr< TimeIntegrator > > _time_integrators
Time integrator.
Definition: SystemBase.h:1049
virtual void setSolutionUDotDotOld(const NumericVector< Number > &u_dotdot_old)
virtual void reinitNeighborPhys(const Elem *neighbor, unsigned int neighbor_side, const std::vector< Point > &physical_points, const THREAD_ID tid)=0
virtual void residualSetup(THREAD_ID tid=0) const
NumericVector< Number > & getResidualTimeVector()
Return a numeric vector that is associated with the time tag.
void setActiveMaterialProperties(const std::unordered_set< unsigned int > &mat_prop_ids, const THREAD_ID tid)
Record and set the material properties required by the current computing thread.
MooseObjectTagWarehouse< NodalKernelBase > _nodal_kernels
NodalKernels for each thread.
NumericVector< Number > * _Re_time
residual vector for time contributions
void computeJacobianBlocks(std::vector< JacobianBlock *> &blocks)
Computes several Jacobian blocks simultaneously, summing their contributions into smaller preconditio...
virtual void insert(const T *v, const std::vector< numeric_index_type > &dof_indices)
MetaPhysicL::DualNumber< V, D, asd > abs(const MetaPhysicL::DualNumber< V, D, asd > &a)
Definition: EigenADReal.h:42
TagID _Re_time_tag
Tag for time contribution residual.
void allgather(const T &send_data, std::vector< T, A > &recv_data) const
unsigned int size(THREAD_ID tid=0) const
Return how many kernels we store in the current warehouse.
std::map< std::pair< BoundaryID, BoundaryID >, PenetrationLocator * > _penetration_locators
dof_id_type end_dof(const processor_id_type proc) const
bool empty() const
virtual void addKernel(const std::string &kernel_name, const std::string &name, InputParameters &parameters)
Adds a kernel.
virtual void setSolutionUDotDot(const NumericVector< Number > &udotdot)
Set transient term used by residual and Jacobian evaluation.
dof_id_type dof_number(const unsigned int s, const unsigned int var, const unsigned int comp) const
A kernel for hybridized finite element formulations.
Definition: HDGKernel.h:17
MooseObjectTagWarehouse< ResidualObject > _kokkos_nodal_kernels
KOKKOS_INLINE_FUNCTION const T * find(const T &target, const T *const begin, const T *const end)
Find a value in an array.
Definition: KokkosUtils.h:30
void reinitIncrementAtNodeForDampers(THREAD_ID tid, const std::set< MooseVariable *> &damped_vars)
Compute the incremental change in variables at nodes for dampers.
void overwriteNodeFace(NumericVector< Number > &soln)
Called from explicit time stepping to overwrite boundary positions (explicit dynamics).
Base class for deriving general dampers.
Definition: GeneralDamper.h:21
bool _use_pre_smo_residual
Whether to use the pre-SMO initial residual in the relative convergence check.
MoosePreconditioner const * getPreconditioner() const
virtual const char * what() const
Get out the error message.
std::vector< std::pair< MooseVariableFEBase *, MooseVariableFEBase * > > & couplingEntries(const THREAD_ID tid, const unsigned int nl_sys_num)
void findImplicitGeometricCouplingEntries(GeometricSearchData &geom_search_data, std::unordered_map< dof_id_type, std::vector< dof_id_type >> &graph)
Finds the implicit sparsity graph between geometrically related dofs.
void setupDampers()
Setup damping stuff (called before we actually start)
Real _initial_residual
The initial (i.e., 0th nonlinear iteration) residual, see setPreSMOResidual for a detailed explanatio...
unsigned int n_threads()
std::vector< bool > _variable_autoscaled
Container to hold flag if variable is to participate in autoscaling.
void zeroVectorForResidual(const std::string &vector_name)
virtual void cacheResidualNeighbor(const THREAD_ID tid) override
bool identifyVariableGroupsInNL() const
Whether to identify variable groups in nonlinear systems.
void zeroTaggedVectors(const std::set< TagID > &tags)
Zero all vectors for given tags.
Definition: SystemBase.C:693
const std::set< SubdomainID > & interiorLowerDBlocks() const
Definition: MooseMesh.h:1421
Base class for split-based preconditioners.
Definition: Split.h:25
bool hasActiveBlockObjects(THREAD_ID tid=0) const
std::shared_ptr< DisplacedProblem > displaced_problem
bool hasVector(const std::string &tag_name) const
Check if the named vector exists in the system.
Definition: SystemBase.C:924
virtual void checkExceptionAndStopSolve(bool print_message=true)
Check to see if an exception has occurred on any processor and, if possible, force the solve to fail...
unsigned int n_comp(const unsigned int s, const unsigned int var) const
void applyScalingFactors(const std::vector< Real > &inverse_scaling_factors)
Applies scaling factors to the system&#39;s variables.
Definition: SystemBase.C:1495
void computeNodalBCsResidualAndJacobian()
compute the residual and Jacobian for nodal boundary conditions
bool _debugging_residuals
true if debugging residuals
SCALAR
BoundaryID _secondary_boundary
NumericVector< Number > * _Re_non_time
residual vector for non-time contributions
const std::map< SubdomainID, std::vector< std::shared_ptr< T > > > & getActiveBlockObjects(THREAD_ID tid=0) const
MooseObjectTagWarehouse< ResidualObject > _kokkos_kernels
unsigned int TagID
Definition: MooseTypes.h:210
Real computeDamping(const NumericVector< Number > &solution, const NumericVector< Number > &update)
Compute damping.
virtual void reinitNode(const Node *node, const THREAD_ID tid) override
virtual void setPreviousNewtonSolution(const NumericVector< Number > &soln)
virtual void predictorCleanup(NumericVector< libMesh::Number > &ghosted_solution)
Perform cleanup tasks after application of predictor to solution vector.
bool _assemble_constraints_separately
Whether or not to assemble the residual and Jacobian after the application of each constraint...
virtual Elem * elemPtr(const dof_id_type i)
Definition: MooseMesh.C:3134
auto exp(const T &)
TagID systemMatrixTag() const override
Return the Matrix Tag ID for System.
MPI_Info info
NumericVector< Number > & solution()
Definition: SystemBase.h:196
bool hasObjects(THREAD_ID tid=0) const
Convenience functions for determining if objects exist.
MooseObjectTagWarehouse< DGKernelBase > _dg_kernels
virtual bool haveFV() const override
returns true if this problem includes/needs finite volume functionality.
char ** blocks
face_info_iterator ownedFaceInfoBegin()
Iterators to owned faceInfo objects.
Definition: MooseMesh.C:1548
PARALLEL
void reinitIncrementAtQpsForDampers(THREAD_ID tid, const std::set< MooseVariable *> &damped_vars)
Compute the incremental change in variables at QPs for dampers.
void computeResidualAndJacobianInternal(const std::set< TagID > &vector_tags, const std::set< TagID > &matrix_tags)
Compute residual and Jacobian from contributions not related to constraints, such as nodal boundary c...
void addImplicitGeometricCouplingEntriesToJacobian(bool add=true)
If called with true this will add entries into the jacobian to link together degrees of freedom that ...
void dof_indices(const Elem *const elem, std::vector< dof_id_type > &di) const
bool areCoupled(const unsigned int ivar, const unsigned int jvar, const unsigned int nl_sys_num) const
void mooseError(Args &&... args)
Emit an error message with the given stringified, concatenated args and terminate the application...
Definition: MooseError.h:323
unsigned int number() const
Get variable number coming from libMesh.
virtual void initialSetup() override
Setup Functions.
Data structure used to hold penetration information.
std::vector< std::pair< R1, R2 > > get(const std::string &param1, const std::string &param2) const
Combine two vector parameters into a single vector of pairs.
const std::vector< std::shared_ptr< NodalConstraint > > & getActiveNodalConstraints() const
Access methods for active objects.
NumericVector< Number > * _u_dot_old
old solution vector for u^dot
Definition: SystemBase.h:1011
bool _has_nodalbc_diag_save_in
If there is a nodal BC having diag_save_in.
virtual void reinitScalars(const THREAD_ID tid, bool reinit_for_derivative_reordering=false) override
fills the VariableValue arrays for scalar variables from the solution vector
Base class for automatic differentiation Dirichlet BCs.
void add(std::shared_ptr< MooseObject > obj)
add adds a new object to the warehouse and stores attributes/metadata about it for running queries/fi...
Definition: TheWarehouse.C:116
virtual void getDiracElements(std::set< const Elem *> &elems) override
Fills "elems" with the elements that should be looped over for Dirac Kernels.
void setupDM()
Setup the PETSc DM object (when appropriate)
void setCurrentlyComputingResidual(bool currently_computing_residual) final
Set whether or not the problem is in the process of computing the residual.
void checkKernelCoverage(const std::set< SubdomainID > &mesh_subdomains) const
void addDGKernel(std::string dg_kernel_name, const std::string &name, InputParameters &parameters)
Adds a DG kernel.
void computeJacobian(libMesh::SparseMatrix< Number > &jacobian, const std::set< TagID > &tags)
Associate jacobian to systemMatrixTag, and then form a matrix for all the tags.
const InputParameters & parameters() const
Get the parameters of the object.
Definition: MooseBase.h:131
char ** vars
virtual void reinitNeighborPhys(const Elem *neighbor, unsigned int neighbor_side, const std::vector< Point > &physical_points, const THREAD_ID tid) override
virtual TagID addVectorTag(const TagName &tag_name, const Moose::VectorTagType type=Moose::VECTOR_TAG_RESIDUAL)
Create a Tag.
Definition: SubProblem.C:93
std::vector< T * > & queryInto(std::vector< T *> &results, Args &&... args)
queryInto executes the query and stores the results in the given vector.
Definition: TheWarehouse.h:311
void getNodeDofs(dof_id_type node_id, std::vector< dof_id_type > &dofs)
Base class for all Constraint types.
Definition: Constraint.h:19
std::set< TagID > _nl_vector_tags
Vector tags to temporarily store all tags associated with the current system.
std::vector< std::string > _ignore_variables_for_autoscaling
A container for variables that do not partipate in autoscaling.
void residualSetup() override
Base boundary condition of a Dirichlet type.
virtual void associateVectorToTag(NumericVector< Number > &vec, TagID tag)
Associate a vector for a given tag.
Definition: SystemBase.C:981
virtual void reinitNodes(const std::vector< dof_id_type > &nodes, const THREAD_ID tid) override
MooseObjectTagWarehouse< NodalBCBase > _nodal_bcs
virtual void addJacobianOffDiagScalar(unsigned int ivar, const THREAD_ID tid=0)
virtual void setException(const std::string &message)
Set an exception, which is stored at this point by toggling a member variable in this class...
const std::vector< std::shared_ptr< NodeFaceConstraint > > & getActiveNodeFaceConstraints(BoundaryID boundary_id, bool displaced) const
virtual void customSetup(const ExecFlagType &exec_type, THREAD_ID tid=0) const
Factory & _factory
Definition: SystemBase.h:989
MooseObjectWarehouse< T > & getVectorTagsObjectWarehouse(const std::set< TagID > &tags, THREAD_ID tid)
Retrieve a moose object warehouse in which every moose object at least has one of the given vector ta...
virtual const Node * queryNodePtr(const dof_id_type i) const
Definition: MooseMesh.C:875
const ElementPairInfo & getElemPairInfo(std::pair< const Elem *, const Elem *> elem_pair) const
MooseObjectWarehouseBase< Split > _splits
Decomposition splits.
The main MOOSE class responsible for handling user-defined parameters in almost every MOOSE system...
NumericVector< Number > * _u_dotdot
solution vector for u^dotdot
Definition: SystemBase.h:1008
bool hasActiveMortarConstraints(const std::pair< BoundaryID, BoundaryID > &mortar_interface_key, bool displaced) const
const Parallel::Communicator & comm() const
NumericVector< Number > & add_vector(std::string_view vec_name, const bool projections=true, const ParallelType type=PARALLEL)
Solving a linear problem.
Definition: MooseTypes.h:849
bool _has_nodalbc_save_in
If there is a nodal BC having save_in.
bool hasActiveNodalConstraints() const
Deterimine if active objects exist.
virtual bool hasMatrix(TagID tag) const
Check if the tagged matrix exists in the system.
Definition: SystemBase.h:360
MooseObjectWarehouse< NodalDamper > _nodal_dampers
Nodal Dampers for each thread.
std::unique_ptr< T_DEST, T_DELETER > dynamic_pointer_cast(std::unique_ptr< T_SRC, T_DELETER > &src)
These are reworked from https://stackoverflow.com/a/11003103.
virtual libMesh::NonlinearSolver< Number > * nonlinearSolver()=0
dof_id_type n_dofs(const unsigned int vn) const
Real preSMOResidual() const
The pre-SMO residual.
std::unique_ptr< libMesh::DiagonalMatrix< Number > > _scaling_matrix
A diagonal matrix used for computing scaling.
virtual void setSolutionUDotOld(const NumericVector< Number > &u_dot_old)
virtual void associateMatrixToTag(libMesh::SparseMatrix< Number > &matrix, TagID tag)
Associate a matrix to a tag.
Definition: SystemBase.C:1076
bool hasDiagSaveIn() const
Weather or not the nonlinear system has diagonal Jacobian save-ins.
This class provides an interface for common operations on field variables of both FE and FV types wit...
const Parallel::Communicator & _communicator
void updateActive(THREAD_ID tid=0) override
Update the various active lists.
Real initialResidual() const
The initial residual.
The following methods are specializations for using the libMesh::Parallel::packed_range_* routines fo...
bool needInterfaceMaterialOnSide(BoundaryID bnd_id, THREAD_ID tid) const
Indicated whether this system needs material properties on interfaces.
const libMesh::ConstElemRange & getCurrentAlgebraicElementRange()
These are the element and nodes that contribute to the jacobian and residual for this local processor...
MooseObjectWarehouse< ResidualObject > _kokkos_preset_nodal_bcs
std::size_t _num_scaling_groups
The number of scaling groups.
const libMesh::ConstNodeRange & getCurrentAlgebraicNodeRange()
void computingScalingJacobian(bool computing_scaling_jacobian)
Setter for whether we&#39;re computing the scaling jacobian.
dof_id_type n_local_dofs(const unsigned int vn) const
virtual void updateActive(THREAD_ID tid=0) override
Update the active status of Kernels.
bool has_dofs(const unsigned int s=libMesh::invalid_uint) const
std::map< dof_id_type, PenetrationInfo * > & _penetration_info
Data structure of nodes and their associated penetration information.
bool hasActiveNodeElemConstraints(SubdomainID secondary_id, SubdomainID primary_id, bool displaced) const
Real _pre_smo_residual
The pre-SMO residual, see setPreSMOResidual for a detailed explanation.
bool hasDampers()
Whether or not this system has dampers.
const std::vector< std::shared_ptr< NodeElemConstraintBase > > & getActiveNodeElemConstraints(SubdomainID secondary_id, SubdomainID primary_id, bool displaced) const
bool _compute_scaling_once
Whether the scaling factors should only be computed once at the beginning of the simulation through a...
void computeResidualTags(const std::set< TagID > &tags)
Form multiple tag-associated residual vectors for all the given tags.
virtual void cacheJacobianNeighbor(const THREAD_ID tid) override
Specialization of SubProblem for solving nonlinear equations plus auxiliary equations.
bool _has_save_in
If there is any Kernel or IntegratedBC having save_in.
virtual const Node & nodeRef(const dof_id_type i) const
Definition: MooseMesh.C:849
TagID _Ke_system_tag
Tag for system contribution Jacobian.
virtual void setResidual(NumericVector< libMesh::Number > &residual, const THREAD_ID tid) override
dof_id_type n_dofs() const
virtual void disassociateMatrixFromTag(libMesh::SparseMatrix< Number > &matrix, TagID tag)
Disassociate a matrix from a tag.
Definition: SystemBase.C:1088
Scope guard for starting and stopping Floating Point Exception Trapping.
virtual void zero()=0
auto max(const L &left, const R &right)
Serves as a base class for DGKernel and ADDGKernel.
Definition: DGKernelBase.h:32
const Variable & variable(const unsigned int c) const override
void constraintResiduals(NumericVector< Number > &residual, bool displaced)
Add residual contributions from Constraints.
Base class for MOOSE preconditioners.
NumericVector< Number > & addVector(const std::string &vector_name, const bool project, const libMesh::ParallelType type)
Adds a solution length vector to the system.
virtual GeometricSearchData & geomSearchData() override
std::unordered_map< std::pair< BoundaryID, BoundaryID >, ComputeMortarFunctor > _undisplaced_mortar_functors
Functors for computing undisplaced mortar constraints.
Specialization for filling multiple "small" preconditioning matrices simulatenously.
virtual bool matrixFromColoring() const
Whether a system matrix is formed from coloring.
Definition: SolverSystem.h:102
void update()
Update the system (doing libMesh magic)
Definition: SystemBase.C:1243
virtual Assembly & assembly(const THREAD_ID tid, const unsigned int sys_num) override
void addScalarKernel(const std::string &kernel_name, const std::string &name, InputParameters &parameters)
Adds a scalar kernel.
void addBoundaryCondition(const std::string &bc_name, const std::string &name, InputParameters &parameters)
Adds a boundary condition.
bool hasActiveNodeFaceConstraints(BoundaryID boundary_id, bool displaced) const
void computeResidual(NumericVector< Number > &residual, TagID tag_id)
Form a residual vector for a given tag.
virtual void addCachedResidualDirectly(NumericVector< libMesh::Number > &residual, const THREAD_ID tid)
Allows for all the residual contributions that are currently cached to be added directly into the vec...
virtual unsigned int nVariables() const
Get the number of variables in this system.
Definition: SystemBase.C:891
EXTERN_C_BEGIN PetscErrorCode DMCreate_Moose(DM)
bool hasActiveBoundaryObjects(THREAD_ID tid=0) const
bool _need_residual_ghosted
Whether or not a ghosted copy of the residual needs to be made.
virtual void activateAllMatrixTags()
Make all existing matrices active.
Definition: SystemBase.C:1131
virtual const std::string & name() const
Definition: SystemBase.C:1340
virtual void jacobianSetup()
Definition: SystemBase.C:1606
virtual bool containsTimeKernel() override
If the system has a kernel that corresponds to a time derivative.
void onTimestepBegin()
Called at the beginning of the time step.
const ConstBndNodeRange & getCurrentAlgebraicBndNodeRange()
std::vector< dof_id_type, Threads::scalable_allocator< dof_id_type > > Row
MooseObjectTagWarehouse< DiracKernelBase > _dirac_kernels
Dirac Kernel storage for each thread.
void closeTaggedMatrices(const std::set< TagID > &tags)
Close all matrices associated the tags.
Definition: SystemBase.C:1060
TagID _Re_non_time_tag
Tag for non-time contribution residual.
std::set< TagID > _nl_matrix_tags
Matrix tags to temporarily store all tags associated with the current system.
void solutionInvalidAccumulation()
Pass the number of solution invalid occurrences from current iteration to cumulative counters...
void set_basic_system_only()
std::map< std::pair< BoundaryID, BoundaryID >, NearestNodeLocator * > _nearest_node_locators
void setPredictor(std::shared_ptr< Predictor > predictor)
Real _resid_vs_jac_scaling_param
The param that indicates the weighting of the residual vs the Jacobian in determining variable scalin...
bool _auto_scaling_initd
Whether we&#39;ve initialized the automatic scaling data structures.
virtual void computeScalingResidual()=0
Compute a "residual" for automatic scaling purposes.
bool _doing_dg
true if DG is active (optimization reasons)
virtual libMesh::DofMap & dofMap()
Gets writeable reference to the dof map.
Definition: SystemBase.C:1163
void syncIteration()
Sync iteration counts to main processor.
virtual void deactivateAllMatrixTags()
Make matrices inactive.
Definition: SystemBase.C:1119
unsigned int number() const
void computeResidualAndJacobianTags(const std::set< TagID > &vector_tags, const std::set< TagID > &matrix_tags)
Form possibly multiple tag-associated vectors and matrices.
SERIAL
bool needBoundaryMaterialOnSide(BoundaryID bnd_id, THREAD_ID tid) const
Indicated whether this system needs material properties on boundaries.
MooseObjectWarehouse< DirichletBCBase > _preset_nodal_bcs
std::unordered_map< unsigned int, unsigned int > _var_to_group_var
A map from variable index to group variable index and it&#39;s associated (inverse) scaling factor...
const std::vector< std::shared_ptr< T > > & getActiveObjects(THREAD_ID tid=0) const
Retrieve complete vector to the active all/block/boundary restricted objects for a given thread...
std::vector< unsigned int > _current_l_its
virtual void zero()=0
virtual std::unique_ptr< Base > create()=0
void setCurrentNonlinearSystem(const unsigned int nl_sys_num)
std::shared_ptr< T > getActiveObject(const std::string &name, THREAD_ID tid=0) const
This is the common base class for the three main kernel types implemented in MOOSE, Kernel, VectorKernel and ArrayKernel.
Definition: KernelBase.h:23
dof_id_type id() const
std::vector< dof_id_type > _secondary_nodes
MeshBase & getMesh()
Accessor for the underlying libMesh Mesh object.
Definition: MooseMesh.C:3469
void min(const T &r, T &o, Request &req) const
void computeDiracContributions(const std::set< TagID > &tags, bool is_jacobian)
void updateActive(THREAD_ID tid)
Update active objects of Warehouses owned by NonlinearSystemBase.
TheWarehouse & theWarehouse() const
unsigned int n_vars
void reinitMaterialsNeighbor(SubdomainID blk_id, const THREAD_ID tid, bool swap_stateful=true, const std::deque< MaterialBase *> *reinit_mats=nullptr)
reinit materials on the neighboring element face
const ElementPairList & getElemPairs() const
FieldSplitPreconditionerBase & getFieldSplitPreconditioner()
std::unordered_map< std::pair< BoundaryID, BoundaryID >, ComputeMortarFunctor > _displaced_mortar_functors
Functors for computing displaced mortar constraints.
const std::unordered_map< std::pair< BoundaryID, BoundaryID >, std::unique_ptr< AutomaticMortarGeneration > > & getMortarInterfaces(bool on_displaced) const
GHOSTED
boundary_id_type BoundaryID
Real referenceResidual() const
The reference residual used in relative convergence check.
void addSplit(const std::string &split_name, const std::string &name, InputParameters &parameters)
Adds a split.
bool _automatic_scaling
Whether to automatically scale the variables.
Definition: SystemBase.h:1055
void computeJacobianTags(const std::set< TagID > &tags)
Computes multiple (tag associated) Jacobian matricese.
SolutionInvalidity & solutionInvalidity()
Get the SolutionInvalidity for this app.
Definition: MooseApp.h:179
std::shared_ptr< MoosePreconditioner > _preconditioner
Preconditioner.
virtual void timestepSetup(THREAD_ID tid=0) const
NonlinearSystemBase & currentNonlinearSystem()
SimpleRange< IndexType > as_range(const std::pair< IndexType, IndexType > &p)
void computingScalingResidual(bool computing_scaling_residual)
Setter for whether we&#39;re computing the scaling residual.
virtual void setupDM()=0
setup the data management data structure that manages the field split
SubProblem & subproblem()
Definition: SystemBase.h:101
std::vector< std::string > _vecs_to_zero_for_residual
vectors that will be zeroed before a residual computation
virtual TagID addMatrixTag(TagName tag_name)
Create a Tag.
Definition: SubProblem.C:312
std::shared_ptr< Split > getSplit(const std::string &name)
Retrieves a split by name.
std::unique_ptr< NumericVector< Number > > solution
SubProblem & _subproblem
The subproblem for whom this class holds variable data, etc; this can either be the governing finite ...
Definition: SystemBase.h:983
void subdomainsCovered(std::set< SubdomainID > &subdomains_covered, std::set< std::string > &unique_variables, THREAD_ID tid=0) const
Populates a set of covered subdomains and the associated variable names.
void addImplicitGeometricCouplingEntries(GeometricSearchData &geom_search_data)
Adds entries to the Jacobian in the correct positions for couplings coming from dofs being coupled th...
virtual void addHDGKernel(const std::string &kernel_name, const std::string &name, InputParameters &parameters)
Adds a hybridized discontinuous Galerkin (HDG) kernel.
virtual void zero_rows(std::vector< numeric_index_type > &rows, T diag_value=0.0)
MooseObjectTagWarehouse< KernelBase > _kernels
MooseObjectWarehouse< T > & getMatrixTagObjectWarehouse(TagID tag_id, THREAD_ID tid)
Retrieve a moose object warehouse in which every moose object has the given matrix tag...
bool shouldEvaluatePreSMOResidual() const
We offer the option to check convergence against the pre-SMO residual.
std::vector< VectorTag > getVectorTags(const std::set< TagID > &tag_ids) const
Definition: SubProblem.C:173
Moose::SolveType _type
Definition: SolverParams.h:19
Base class for deriving nodal dampers.
Definition: NodalDamper.h:27
virtual void disassociateVectorFromTag(NumericVector< Number > &vec, TagID tag)
Disassociate a given vector from a given tag.
virtual void cacheResidual(const THREAD_ID tid) override
void mooseDeprecated(Args &&... args)
Emit a deprecated code/feature message with the given stringified, concatenated args.
Definition: MooseError.h:374
virtual void computeScalingJacobian()=0
Compute a "Jacobian" for automatic scaling purposes.
bool errorOnJacobianNonzeroReallocation() const
Will return True if the user wants to get an error when a nonzero is reallocated in the Jacobian by P...
This is the ElementPairLocator class.
This is the ElementPairInfo class.
std::map< BoundaryID, std::shared_ptr< ElementPairLocator > > _element_pair_locators
Moose::CouplingType coupling() const
unsigned int number() const
Gets the number of this system.
Definition: SystemBase.C:1157
void setupScalingData()
Setup group scaling containers.
virtual GeometricSearchData & geomSearchData()=0
const bool & usePreSMOResidual() const
Whether we are using pre-SMO residual in relative convergence checks.
MooseObjectTagWarehouse< HDGKernel > _hybridized_kernels
const std::set< SubdomainID > & boundaryLowerDBlocks() const
Definition: MooseMesh.h:1425
std::vector< SubdomainName > getSubdomainNames(const std::vector< SubdomainID > &subdomain_ids) const
Get the associated subdomainNames for the subdomain ids that are passed in.
Definition: MooseMesh.C:1807
auto log(const T &)
AuxiliarySystem & getAuxiliarySystem()
virtual void initialSetup(THREAD_ID tid=0) const
Convenience methods for calling object setup methods.
virtual void updateGeomSearch(GeometricSearchData::GeometricSearchType type=GeometricSearchData::ALL) override
void closeTaggedVectors(const std::set< TagID > &tags)
Close all vectors for given tags.
Definition: SystemBase.C:667
Base class for deriving any boundary condition that works at nodes.
Definition: NodalBCBase.h:26
void computeResidualInternal(const std::set< TagID > &tags)
Compute the residual for a given tag.
virtual void prepareAssembly(const THREAD_ID tid) override
bool computeScaling()
Method used to obtain scaling factors for variables.
Interface for objects interacting with the PerfGraph.
virtual std::map< TagName, TagID > & getMatrixTags()
Return all matrix tags in the system, where a tag is represented by a map from name to ID...
Definition: SubProblem.h:253
virtual bool hasVariable(const std::string &var_name) const
Query a system for a variable.
Definition: SystemBase.C:851
virtual void setCurrentSubdomainID(const Elem *elem, const THREAD_ID tid) override
virtual void clearDiracInfo() override
Gets called before Dirac Kernels are asked to add the points they are supposed to be evaluated in...
void destroyColoring()
Destroy the coloring object if it exists.
virtual void reinitNodesNeighbor(const std::vector< dof_id_type > &nodes, const THREAD_ID tid) override
bool identify_variable_groups() const
virtual void close()=0
virtual void jacobianSetup(THREAD_ID tid=0) const
ConstraintWarehouse _constraints
Constraints storage object.
virtual void turnOffJacobian()
Turn off the Jacobian (must be called before equation system initialization)
void computeKokkosJacobian(const std::set< TagID > &tags)
Compute Jacobian with Kokkos objects.
TagID residualVectorTag() const override
const std::map< BoundaryID, std::vector< std::shared_ptr< T > > > & getActiveBoundaryObjects(THREAD_ID tid=0) const
void computingNonlinearResid(bool computing_nonlinear_residual) final
Set whether or not the problem is in the process of computing the nonlinear residual.
virtual void customSetup(const ExecFlagType &exec_type) override
ComputeType
The type of nonlinear computation being performed.
Definition: MooseTypes.h:781
Base class for deriving element dampers.
Definition: ElementDamper.h:33
Base interface for field split preconditioner.
bool _add_implicit_geometric_coupling_entries_to_jacobian
Whether or not to add implicit geometric couplings to the Jacobian for FDP.
const_iterator end() const
MooseObjectTagWarehouse< ResidualObject > _kokkos_integrated_bcs
virtual void addNodalKernel(const std::string &kernel_name, const std::string &name, InputParameters &parameters)
Adds a NodalKernel.
virtual unsigned int numMatrixTags() const
The total number of tags.
Definition: SubProblem.h:248
Base class for creating new types of boundary conditions.
tbb::split split
MooseApp & _app
Definition: SystemBase.h:988
FEProblemBase & _fe_problem
the governing finite element/volume problem
Definition: SystemBase.h:986
virtual void reinitElemPhys(const Elem *elem, const std::vector< Point > &phys_points_in_elem, const THREAD_ID tid)=0
virtual MooseVariableScalar & getScalarVariable(THREAD_ID tid, const std::string &var_name) const
Gets a reference to a scalar variable with specified number.
Definition: SystemBase.C:145
ParallelType type() const
virtual NumericVector< Number > & RHS()=0
std::vector< VariableWarehouse > _vars
Variable warehouses (one for each thread)
Definition: SystemBase.h:996
Provides a way for users to bail out of the current solve.
unsigned int _n_residual_evaluations
Total number of residual evaluations that have been performed.
virtual void update()
void addDiracKernel(const std::string &kernel_name, const std::string &name, InputParameters &parameters)
Adds a Dirac kernel.
virtual void close()=0
void computeJacobianInternal(const std::set< TagID > &tags)
Form multiple matrices for all the tags.
bool _has_diag_save_in
If there is any Kernel or IntegratedBC having diag_save_in.
Base class for creating new types of nodal kernels.
const FEType & variable_type(const unsigned int i) const
DIE A HORRIBLE DEATH HERE typedef LIBMESH_DEFAULT_SCALAR_TYPE Real
virtual std::shared_ptr< const DisplacedProblem > getDisplacedProblem() const
std::unique_ptr< NumericVector< Number > > _residual_copy
Copy of the residual vector, or nullptr if a copy is not needed.
virtual void subdomainSetup()
Definition: SystemBase.C:1592
bool hasSaveIn() const
Weather or not the nonlinear system has save-ins.
Generic class for solving transient nonlinear problems.
Definition: SubProblem.h:78
TagID timeVectorTag() const override
Ideally, we should not need this API.
Class for containing MooseEnum item information.
Definition: MooseEnumItem.h:18
MooseMesh & _mesh
Definition: SystemBase.h:991
virtual std::vector< std::string > timeKernelVariableNames() override
Returns the names of the variables that have time derivative kernels in the system.
bool hasActiveObjects(THREAD_ID tid=0) const
NonlinearSystemBase(FEProblemBase &problem, libMesh::System &sys, const std::string &name)
void max(const T &r, T &o, Request &req) const
NumericVector< Number > * _u_dot
solution vector for u^dot
Definition: SystemBase.h:1006
MooseObjectWarehouse< ElementDamper > _element_dampers
Element Dampers for each thread.
void addObject(std::shared_ptr< Constraint > object, THREAD_ID tid=0, bool recurse=true) override
Add Constraint object to the warehouse.
virtual void setSolutionUDot(const NumericVector< Number > &udot)
Set transient term used by residual and Jacobian evaluation.
virtual libMesh::SparseMatrix< Number > & getMatrix(TagID tag)
Get a raw SparseMatrix.
Definition: SystemBase.C:1024
bool hasKokkosResidualObjects() const
virtual void augmentSparsity(libMesh::SparsityPattern::Graph &sparsity, std::vector< dof_id_type > &n_nz, std::vector< dof_id_type > &n_oz) override
Will modify the sparsity pattern to add logical geometric connections.
const std::vector< std::shared_ptr< MortarConstraintBase > > & getActiveMortarConstraints(const std::pair< BoundaryID, BoundaryID > &mortar_interface_key, bool displaced) const
MooseObjectWarehouse< T > & getMatrixTagsObjectWarehouse(const std::set< TagID > &tags, THREAD_ID tid)
Retrieve a moose object warehouse in which every moose object has one of the given matrix tags...
void computeScalarKernelsJacobians(const std::set< TagID > &tags)
void setInitialResidual(Real r)
Record the initial residual (for later relative convergence check)
Query query()
query creates and returns an initialized a query object for querying objects from the warehouse...
Definition: TheWarehouse.h:466
virtual Assembly & assembly(const THREAD_ID tid, const unsigned int sys_num)=0
const_iterator begin() const
bool needInternalNeighborSideMaterial(SubdomainID subdomain_id, THREAD_ID tid) const
Indicates whether this system needs material properties on internal sides.
void computeKokkosResidual(const std::set< TagID > &tags)
Compute residual with Kokkos objects.
Base class for deriving dampers.
Definition: Damper.h:24
bool getFailNextNonlinearConvergenceCheck() const
Whether it will skip further residual evaluations and fail the next nonlinear convergence check(s) ...
Base class shared by AD and non-AD scalar kernels.
void addDamper(const std::string &damper_name, const std::string &name, InputParameters &parameters)
Adds a damper.
IntRange< T > make_range(T beg, T end)
virtual MooseMesh & mesh() override
virtual void postAddResidualObject(ResidualObject &)
Called after any ResidualObject-derived objects are added to the system.
const std::vector< VariableName > & getVariableNames() const
Definition: SystemBase.h:860
virtual void preInit() override
This is called prior to the libMesh system has been init&#39;d.
Definition: SolverSystem.C:32
virtual void updateActive(THREAD_ID tid=0) override
Update the active status of Kernels.
virtual void timestepSetup() override
bool hasActiveElemElemConstraints(const InterfaceID interface_id, bool displaced) const
bool _off_diagonals_in_auto_scaling
Whether to include off diagonals when determining automatic scaling factors.
virtual unsigned int numVectorTags(const Moose::VectorTagType type=Moose::VECTOR_TAG_ANY) const
The total number of tags, which can be limited to the tag type.
Definition: SubProblem.C:196
void reinitNodeFace(const Node &secondary_node, const BoundaryID secondary_boundary, const PenetrationInfo &info, const bool displaced)
Reinitialize quantities such as variables, residuals, Jacobians, materials for node-face constraints...
Base class for deriving any boundary condition of a integrated type.
SolverParams & solverParams(unsigned int solver_sys_num=0)
Get the solver parameters.
NumericVector< Number > * _residual_ghosted
ghosted form of the residual
void setCachedJacobian(GlobalDataKey)
Sets previously-cached Jacobian values via SparseMatrix::set() calls.
Definition: Assembly.C:4492
TagID _Re_tag
Used for the residual vector from PETSc.
virtual void customSetup(const ExecFlagType &exec_type)
Definition: SystemBase.C:1585
MooseObjectWarehouse< T > & getVectorTagObjectWarehouse(TagID tag_id, THREAD_ID tid)
Retrieve a moose object warehouse in which every moose object has the given vector tag...
void computeNodalBCs(NumericVector< Number > &residual)
Enforces nodal boundary conditions.
const std::set< SubdomainID > & getSubdomainsForVar(unsigned int var_number) const
Definition: SystemBase.h:761
virtual void addObject(std::shared_ptr< T > object, THREAD_ID tid=0, bool recurse=true)
Adds an object to the storage structure.
bool ignoreZerosInJacobian() const
Will return true if zeros in the Jacobian are to be dropped from the sparsity pattern.
const ExecFlagType EXEC_PRE_KERNELS
Definition: Moose.C:56
libMesh::System & _sys
void mortarConstraints(Moose::ComputeType compute_type, const std::set< TagID > &vector_tags, const std::set< TagID > &matrix_tags)
Do mortar constraint residual/jacobian computations.
NumericVector< Number > * _increment_vec
increment vector
InterfaceKernelBase is the base class for all InterfaceKernel type classes.
QueryCache & condition(Args &&... args)
Adds a new condition to the query.
Definition: TheWarehouse.h:284
bool doingDG() const
Getter for _doing_dg.
void computeResidualTag(NumericVector< Number > &residual, TagID tag_id)
Computes residual for a given tag.
void addConstraint(const std::string &c_name, const std::string &name, InputParameters &parameters)
Adds a Constraint.
virtual TagName vectorTagName(const TagID tag) const
Retrieve the name associated with a TagID.
Definition: SubProblem.C:222
MOOSE now contains C++17 code, so give a reasonable error message stating what the user can do to add...
const ConsoleStream _console
An instance of helper class to write streams to the Console objects.
bool hasKokkosObjects() const
bool restoreOriginalNonzeroPattern() const
face_info_iterator ownedFaceInfoEnd()
Definition: MooseMesh.C:1557
void constraintJacobians(const SparseMatrix< Number > &jacobian_to_view, bool displaced)
Add jacobian contributions from Constraints.
virtual libMesh::System & system() override
Get the reference to the libMesh system.
MooseVariableFieldBase & getVariable(THREAD_ID tid, const std::string &var_name) const
Gets a reference to a variable of with specified name.
Definition: SystemBase.C:90
std::vector< BoundaryID > getBoundaryIDs(const Elem *const elem, const unsigned short int side) const
Returns a vector of boundary IDs for the requested element on the requested side. ...
bool preSolve()
Perform some steps to get ready for the solver.
void full_sparsity_pattern_needed()
bool _has_constraints
Whether or not this system has any Constraints.
dof_id_type first_dof(const processor_id_type proc) const
bool _computed_scaling
Flag used to indicate whether we have already computed the scaling Jacobian.
void remove_algebraic_ghosting_functor(GhostingFunctor &evaluable_functor)
MooseObjectTagWarehouse< InterfaceKernelBase > _interface_kernels
unsigned int n_vars() const
void resetSolutionInvalidCurrentIteration()
Reset the number of solution invalid occurrences back to zero.
NumericVector< Number > & residualVector(TagID tag)
Return a residual vector that is associated with the residual tag.
NumericVector< Number > & solutionOld()
Definition: SystemBase.h:197
NumericVector< Number > & getResidualNonTimeVector()
Return a numeric vector that is associated with the nontime tag.
virtual bool hasScalarVariable(const std::string &var_name) const
Definition: SystemBase.C:876
const TagName PREVIOUS_NL_SOLUTION_TAG
Definition: MooseTypes.C:28
processor_id_type processor_id() const
std::shared_ptr< Predictor > _predictor
If predictor is active, this is non-NULL.
void enforceNodalConstraintsResidual(NumericVector< Number > &residual)
Enforce nodal constraints.
void subdomainsCovered(std::set< SubdomainID > &subdomains_covered, std::set< std::string > &unique_variables, THREAD_ID tid=0) const
Update supplied subdomain and variable coverate containters.
MooseObjectWarehouse< GeneralDamper > _general_dampers
General Dampers.
const std::vector< std::shared_ptr< ElemElemConstraint > > & getActiveElemElemConstraints(InterfaceID interface_id, bool displaced) const
virtual void initialSetup()
Setup Functions.
Definition: SystemBase.C:1558
bool defaultGhosting()
Whether or not the user has requested default ghosting ot be on.
Definition: SubProblem.h:144
void addCachedJacobian(GlobalDataKey)
Adds the values that have been cached by calling cacheJacobian() and or cacheJacobianNeighbor() to th...
Definition: Assembly.C:3815
virtual void cacheJacobian(const THREAD_ID tid) override
auto min(const L &left, const R &right)
void jacobianSetup() override
void setKokkosInitialSolution()
virtual NumericVector< Number > & getVector(const std::string &name)
Get a raw NumericVector by name.
Definition: SystemBase.C:933
MooseObjectTagWarehouse< IntegratedBCBase > _integrated_bcs
const DofMap & get_dof_map() const
virtual void reinitNodeFace(const Node *node, BoundaryID bnd_id, const THREAD_ID tid) override
virtual void residualSetup()
Definition: SystemBase.C:1599
virtual void reinitOffDiagScalars(const THREAD_ID tid) override
processor_id_type processor_id() const
virtual void addObject(std::shared_ptr< T > object, THREAD_ID tid=0, bool recurse=true) override
Adds an object to the storage structure.
virtual void setNeighborSubdomainID(const Elem *elem, unsigned int side, const THREAD_ID tid) override
virtual void addResidualScalar(const THREAD_ID tid=0)
virtual void subdomainSetup(THREAD_ID tid=0) const
virtual void jacobianSetup() override
virtual void addCachedResidual(const THREAD_ID tid) override
std::vector< std::vector< std::string > > _scaling_group_variables
A container of variable groupings that can be used in scaling calculations.
void addInterfaceKernel(std::string interface_kernel_name, const std::string &name, InputParameters &parameters)
Adds an interface kernel.
FieldSplitPreconditionerBase * _fsp
The field split preconditioner if this sytem is using one.
auto index_range(const T &sizable)
virtual NumericVector< Number > & residualCopy() override
void reinitMortarFunctors()
Update the mortar functors if the mesh has changed.
virtual NumericVector< Number > & residualGhosted() override
DiracKernelBase is the base class for all DiracKernel type classes.
virtual void updateActive(THREAD_ID tid=0)
Updates the active objects storage.
MooseObjectTagWarehouse< ResidualObject > _kokkos_nodal_bcs
void assembleScalingVector()
Assemble the numeric vector of scaling factors such that it can be used during assembly of the system...
BoundaryID _primary_boundary
NumericVector< Number > * _u_dotdot_old
old solution vector for u^dotdot
Definition: SystemBase.h:1013
Base variable class.
MooseObjectTagWarehouse< ScalarKernelBase > _scalar_kernels
virtual void residualEnd(THREAD_ID tid=0) const
void setConstraintSecondaryValues(NumericVector< Number > &solution, bool displaced)
Sets the value of constrained variables in the solution vector.
unsigned int THREAD_ID
Definition: MooseTypes.h:209
uint8_t dof_id_type
virtual void addJacobianScalar(const THREAD_ID tid=0)
NearestNodeLocator & _nearest_node
const std::map< dof_id_type, std::vector< dof_id_type > > & nodeToElemMap()
If not already created, creates a map from every node to all elements to which they are connected...
Definition: MooseMesh.C:1216
const std::set< SubdomainID > & meshSubdomains() const
Returns a read-only reference to the set of subdomains currently present in the Mesh.
Definition: MooseMesh.C:3192
ParallelType
virtual void addCachedJacobian(const THREAD_ID tid) override
virtual void timestepSetup()
Definition: SystemBase.C:1578
virtual ~NonlinearSystemBase()
virtual void preInit() override
This is called prior to the libMesh system has been init&#39;d.
virtual void residualSetup() override
void setPreconditioner(std::shared_ptr< MoosePreconditioner > pc)
Sets a preconditioner.
virtual void localize(std::vector< T > &v_local) const=0
MooseObjectWarehouse< ADDirichletBCBase > _ad_preset_nodal_bcs
virtual libMesh::System & system() override
Get the reference to the libMesh system.
Key structure for APIs manipulating global vectors/matrices.
Definition: Assembly.h:844