Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
Skip to main content

Using graphics processors to accelerate the computation of the matrix inverse

  • Published:
The Journal of Supercomputing Aims and scope Submit manuscript

Abstract

We study the use of massively parallel architectures for computing a matrix inverse. Two different algorithms are reviewed, the traditional approach based on Gaussian elimination and the Gauss–Jordan elimination alternative, and several high performance implementations are presented and evaluated. The target architecture is a current general-purpose multicore processor (CPU) connected to a graphics processor (GPU). Numerical experiments show the efficiency attained by the proposed implementations and how the computation of large-scale inverses, which only a few years ago would have required a distributed-memory cluster, take only a few minutes on a hybrid architecture formed by a multicore CPU and a GPU.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Similar content being viewed by others

References

  1. Agullo E, Demmel J, Dongarra J, Hadri B, Kurzak J, Langou J, Ltaief H, Luszczek P, Tomov S (2009) Numerical linear algebra on emerging architectures: the PLASMA and MAGMA projects. J Phys Conf Ser 180(1):012037

    Article  Google Scholar 

  2. Anderson E, Bai Z, Bischof C, Demmel J, Dongarra J, Du Croz J, Greenbaum A, Hammarling S, McKenney A, Sorensen D (1999) LAPACK users’ guide, 3rd edn. SIAM, Philadelphia

    Book  Google Scholar 

  3. Benner P, Ezzatti P, Quintana ES, Remón A (2009) Using hybrid cpu-gpu platforms to accelerate the computation of the matrix sign function. In: LNCS, 7th int workshop on algorithms, models and tools for parallel computing on heterogeneous networks. Springer, Berlin

    Google Scholar 

  4. Bientinesi P, Gunnels JA, Myers ME, Quintana-Ortí ES, van de Geijn RA (2005) The science of deriving dense linear algebra algorithms. ACM Trans Math Softw 31(1):1–26

    Article  Google Scholar 

  5. Gerbessiotis AV (1997) Algorithmic and practical considerations for dense matrix computations on the BSP model. PRG-TR 32, Oxford University Computing Laboratory

  6. Golub G, Loan CV (1996) Matrix computations, 3rd edn. The Johns Hopkins University Press, Baltimore

    MATH  Google Scholar 

  7. Gunnels JA, Gustavson FG, Henry GM, van de Geijn RA (2001) FLAME: formal linear algebra methods environment. ACM Trans Math Softw 27(4):422–455

    Article  MATH  Google Scholar 

  8. Quintana-Ortí E, Quintana-Ortí G, Sun X, van de Geijn R (2001) A note on parallel matrix inversion. SIAM J Sci Comput 22:1762–1771

    Article  Google Scholar 

  9. Strazdins A (1998) A comparison of lookahead and algorithmic blocking techniques for parallel matrix factorization. TR-CS-98-07 07, The Australian National University

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to A. Remón.

Rights and permissions

Reprints and permissions

About this article

Cite this article

Ezzatti, P., Quintana-Ortí, E.S. & Remón, A. Using graphics processors to accelerate the computation of the matrix inverse. J Supercomput 58, 429–437 (2011). https://doi.org/10.1007/s11227-011-0606-4

Download citation

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11227-011-0606-4

Keywords