Version 3.8.0 release notes

by Pavel Holoborodko on February 12, 2015

Over last several months we have been focusing on making toolbox fast in processing the large matrices. Although many updates were released along the way (see full changelog) here we outline only the most important changes:

  1. High-level dense linear algebra routines have been updated to use modern performance-oriented algorithms:
    • Divide & Conquer for SVD [1, 2].
    • Multi-shift QR for nonsymmetric eigenproblems [3, 4, 7].
    • MRRR for Hermitian, symmetric and tridiagonal symmetric eigenproblems [5, 6].
  2. All algorithms are implemented in quadruple as well as in arbitrary precision modes.

  3. Most of the dense matrix algorithms have been enabled with parallel execution on multi-core CPUs. In future we plan to add support for GPU (at least for quadruple precision).
  4. Many significant improvements have been made in toolbox architecture: low-level kernel modules are re-designed to minimize memory fragmentation, cache misses, copy-overhead in data exchange with MATLAB, etc.

All combined allows toolbox to reach the following results:

UPDATE (April 22, 2017): Timings for Mathematica 11.1 have been added to the table, thanks to test script contributed by Bor Plestenjak.

Matrix factorization in quadruple precision (500×500, 34 digits)

Factorization Timing (sec)Speed-up (times)
MATLAB Maple MathematicaAdvanpixOver MATLABOver MapleOver Mathematica
A & B are pseudo-random real matrices (500×500):
[L,U] = lu(A) 249.13 85.1615.120.47534.38 182.6732.42
[Q,R] = qr(A) 514.34 458.8644.393.08167.25 149.2114.43
[U,S,V] = svd(A) 5595.26 4317.40376.949.57584.62 451.1139.38
S = svd(A) 2765.11 645.6056.962.98927.17 216.4819.10
[V,D] = eig(A) 21806.30 6060.90584.8633.75646.05 179.5717.33
lambda = eig(A) 3384.58 7822.30205.4623.55143.70 332.118.72
[V,D] = eig(A,B) n/a 11358.00928.89112.05101.36 8.29
lambda = eig(A,B) n/a 5273.00510.8060.3787.34 8.46

Advanpix toolbox outperforms VPA by 500 times, Maple by 210 times and Wolfram Mathematica by 18 times in average.

VPA requires 6 hours to do the full eigen-decomposition alone!

Maple was unable to work with 500×500 matrix in GUI mode (choked with "mserver stopped responding"). Eventually we switched to Maple’s command-line interface. In complex case (not shown) Maple wasn’t able to provide accurate eigenvectors for generalized eigen-problem at all and it required 16 hours to compute the eigenvalues…

Our test environment and test scripts:

Future Work

As always, we will continue working on performance improvement of dense matrix computations focusing on: more parallelism (multi-core CPU, GPU) and advanced algorithms (e.g. multi-shift QZ [7, 8] for generalized eigenproblems, MRRR for SVD).

However for the next release we will prioritize long-awaited (and delayed) Krulov-Schur eigensolver [9] for sparse matrices. Iterative solvers GMRES and BiCGSTAB will be introduced as well.


I am grateful to Andre Van Moer (Université libre de Bruxelles) who not only ran all the comparison tests on his powerful computer but also spent a lot of time and efforts on making tests run properly in all software packages (re-writing scripts for Maple CLI, etc.). Also for his tremendous help with ongoing testing of alpha-beta versions of toolbox in various configurations and settings.

I also thank Gerhard Wittreich (University of Delaware), Michael Klanner (University of Technology Graz), Brandon Langley (University of Illinois at Urbana-Champaign), Adam Rancon (Ecole Normale Superieure de Lyon), Steffen Hofmann (Technische Universität Berlin) and many others who provided detailed feedback, bug reports and ideas for improvement.

It would be impossible to move forward without such supportive users. Thank you all.


  1. Jessup E. R., Sorensen D. C., A Parallel Algorithm for Computing the Singular Value Decomposition of a Matrix. SIAM Journal on Matrix Analysis and Applications 15:530-548, 1994.^
  2. Gu M., Eisenstat S.C., A divide-and-conquer algorithm for the bidiagonal SVD, SIAM Journal on Matrix Analysis and Applications, 16:79–92, 1995.^
  3. Braman K., Byers R., Mathias R., The multi-shift QR algorithm Part I: Maintaining well focused shifts, and level 3 performance, SIAM Journal on Matrix Analysis and Applications, 23:929–947, 2002.^
  4. Braman K., Byers R., Mathias R., The multi-shift QR algorithm Part II: Aggressive early deflation, SIAM Journal on Matrix Analysis and Applications, 23:948–973, 2002.^
  5. Dhillon I.S., Parlett B.N., Vömel C., The design and implementation of the MRRR algorithm, ACM Transactions on Mathematical Software 32:533-560, 2006.^
  6. Dhillon I.S., and Parlett B.N., Multiple representations to compute orthogonal eigenvectors of symmetric tridiagonal matrices, Linear Algebra and its Applications, 387(1), pp. 1-28, August 2004.^
  7. Kreßner D., Numerical methods and software for general and structured eigenvalue problems, PhD thesis, TU Berlin, Institut für Mathematik, 2004.^
  8. Kågström B., Kreßner D., Multishift variants of the QZ algorithm with aggressive early deflation, SIAM Journal on Matrix Analysis and Applications, 2006.^
  9. Stewart G. W., A Krylov–Schur Algorithm for Large Eigenproblems. SIAM Journal on Matrix Analysis and Applications, 23:601–614, 2001.^

{ 0 comments… add one now }

Leave a Comment

Previous post:

Next post: