Abstract
Most numerical algorithms are designed for single or double precision floating point arithmetic, and their complexity is measured in terms of the total number of floating point operations. The resolution of problems with high condition numbers (e.g. when approaching a singularity or degeneracy) may require higher working precisions, in which case it is important to take the precision into account when doing complexity analyses. In this paper, we propose a new "ultimate complexity"model, which focuses on analyzing the cost of numerical algorithms for "sufficiently large"precisions. As an example application we will present an ultimately softly linear time algorithm for modular composition of univariate polynomials.
| Original language | English |
|---|---|
| Pages (from-to) | 1-13 |
| Number of pages | 13 |
| Journal | ACM Communications in Computer Algebra |
| Volume | 54 |
| Issue number | 1 |
| DOIs | |
| Publication status | Published - 19 Aug 2020 |
Fingerprint
Dive into the research topics of 'Ultimate complexity for numerical algorithms'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver