Skip to content

Hypre3D Laplacian diagnostics, parallel boundary default fix#3055

Closed
bendudson wants to merge 7 commits intonextfrom
next-hypre3d-laplace
Closed

Hypre3D Laplacian diagnostics, parallel boundary default fix#3055
bendudson wants to merge 7 commits intonextfrom
next-hypre3d-laplace

Conversation

@bendudson
Copy link
Contributor

  • Hypre3D Laplacian solver now saves diagnostic data if Laplacian::savePerformance() is called.
  • The default parallel boundary condition "parallel_dirichlet" is not implemented. Changed to "parallel_dirichlet_o2".
  • Updated and fixed the elm-pb example: Added phiSolver->savePerformance call, and fixed the data-hypre input.

bendudson and others added 7 commits January 9, 2025 10:44
Saved to dmp files by the Laplacian savePerformance monitor.
If the savePerformance() method is called to enable outputs,
the mean solver and AMG iterations, and the final residual,
are saved as time-varying scalars.
Default was "parallel_dirichlet", but no boundary is registered
with that name. Changed to "parallel_dirichlet_o2".
Calling `phiSolver->savePerformance(...)` causes performance (e.g. iteration count)
diagnostics to be saved to the output file.

hypre input did have the wrong parallel transform, boundary conditions
that have not been implemented. Now uses hypre3d as the phiSolver.
Add `override` keyword to squash warning.
- Mesh variable names are case sensitive
- mpirun needs ./ to find executable on some systems
@bendudson
Copy link
Contributor Author

Merged into #3022

@bendudson bendudson closed this Jan 17, 2025
@bendudson bendudson deleted the next-hypre3d-laplace branch January 17, 2025 23:42
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant