Parallelizing dense and banded linear algebra libraries using SMPSs
Ver/ Abrir
Impacto
Scholar |
Otros documentos de la autoría: Badía Sala, Rosa María; Herrero, Josep R.; Labarta Mancho, Jesús; Pérez, Josep M.; Quintana-Orti, Enrique S.; Quintana-Ortí, Gregorio
Metadatos
Mostrar el registro completo del ítemcomunitat-uji-handle:10234/9
comunitat-uji-handle2:10234/7036
comunitat-uji-handle3:10234/8620
comunitat-uji-handle4:
INVESTIGACIONMetadatos
Título
Parallelizing dense and banded linear algebra libraries using SMPSsAutoría
Fecha de publicación
2009-12-25Editor
John WileyISSN
1532-0626Cita bibliográfica
BADIA, Rosa M., et al. Parallelizing dense and banded linear algebra libraries using SMPSs. Concurrency and Computation: Practice and Experience, 2009, vol. 21, no 18, p. 2438-2456.Tipo de documento
info:eu-repo/semantics/articlePalabras clave / Materias
Resumen
The promise of future many-core processors, with hundreds of threads running concurrently, has led the developers of linear algebra libraries to rethink their design in order to extract more parallelism, further exploit ... [+]
The promise of future many-core processors, with hundreds of threads running concurrently, has led the developers of linear algebra libraries to rethink their design in order to extract more parallelism, further exploit data locality, attain better load balance, and pay careful attention to the critical path of computation. In this paper we describe how existing serial libraries such as (C)LAPACK and FLAME can be easily parallelized using the SMPSs tools, consisting of a few OpenMP-like pragmas and a runtime system. In the LAPACK case, this usually requires the development of blocked algorithms for simple BLAS-level operations, which expose concurrency at a finer grain. For better performance, our experimental results indicate that column-major order, as employed by this library, needs to be abandoned in benefit of a block data layout. This will require a deeper rewrite of LAPACK or, alternatively, a dynamic conversion of the storage pattern at run-time. The parallelization of FLAME routines using SMPSs is simpler as this library includes blocked algorithms (or algorithms-by-blocks in the FLAME argot) for most operations and storage-by-blocks (or block data layout) is already in place [-]
Publicado en
Concurrency and computation : practice & experience, 2009 vol. 21, no. 18Derechos de acceso
http://rightsstatements.org/vocab/CNE/1.0/
info:eu-repo/semantics/openAccess
info:eu-repo/semantics/openAccess
Aparece en las colecciones
- ICC_Articles [417]