Reproducibility strategies for parallel preconditioned conjugate gradient
DOI10.1016/J.CAM.2019.112697zbMATH Open1433.65368OpenAlexW2997514602WikidataQ126412216 ScholiaQ126412216MaRDI QIDQ2297141FDOQ2297141
Authors: Roman Iakymchuk, M. Wiesenberger, Enrique S. Quintana-Ortí, María Barreda, José I. Aliaga
Publication date: 18 February 2020
Published in: Journal of Computational and Applied Mathematics (Search for Journal in Brave)
Full work available at URL: https://doi.org/10.1016/j.cam.2019.112697
Recommendations
- First steps towards more numerical reproducibility
- Accuracy and effectiveness of preconditioned conjugate gradient algorithms for large and ill-conditioned problems
- Reproducible and accurate matrix multiplication
- Using accurate arithmetics to improve numerical reproducibility and stability in parallel applications
- Efficiency of reproducible level 1 BLAS
reproducibilityaccuracyhigh-performance computingpreconditioned conjugate gradientfloating-point expansionlong accumulator
Parallel numerical computation (65Y05) Complexity and performance of numerical algorithms (65Y20) Modes of computation (nondeterministic, parallel, interactive, probabilistic, etc.) (68Q10)
Cites Work
- Parallel iterative methods for sparse linear systems
- MPFR
- Accuracy and Stability of Numerical Algorithms
- Basic Linear Algebra Subprograms for Fortran Usage
- A set of level 3 basic linear algebra subprograms
- Title not available (Why is that?)
- Accurate Sum and Dot Product
- Handbook of Floating-Point Arithmetic
- Accurate floating-point summation. II: Sign, \(K\)-fold faithful and rounding to nearest
- Error-free transformations of matrix multiplication by using fast routines of matrix multiplication and its applications
- The exact dot product as basic tool for long interval arithmetic
- Accelerating the solution of linear systems by iterative refinement in three precisions
- Reproducible and accurate matrix multiplication
- Emulation of a FMA and Correctly Rounded Sums: Proved Algorithms Using Rounding to Odd
- Parallel Reproducible Summation
- High-Precision Anchored Accumulators for Reproducible Floating-Point Summation
Cited In (3)
- Infinite-precision inner product and sparse matrix-vector multiplication using Ozaki scheme with Dot2 on manycore processors
- General framework for deriving reproducible Krylov subspace algorithms: BiCGStab case
- Compensated summation and dot product algorithms for floating-point vectors on parallel architectures: error bounds, implementation and application in the Krylov subspace methods
Uses Software
This page was built for publication: Reproducibility strategies for parallel preconditioned conjugate gradient
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q2297141)