Mostrar el registro sencillo del ítem

dc.contributor.authorBarrachina Mir, Sergio
dc.contributor.authorDolz, Manuel F.
dc.contributor.authorSan Juan, Pablo
dc.contributor.authorQuintana-Orti, Enrique S.
dc.date.accessioned2022-06-22T12:01:16Z
dc.date.available2022-06-22T12:01:16Z
dc.date.issued2022-05-30
dc.identifier.citationBarrachina, S., Dolz, M. F., San Juan, P., & Quintana-Ortí, E. S. (2022). Efficient and Portable GEMM-based Convolution Operators for Deep Neural Network Training on Multicore Processors. Journal of Parallel and Distributed Computing.ca_CA
dc.identifier.issn0743-7315
dc.identifier.urihttp://hdl.handle.net/10234/198101
dc.description.abstractConvolutional Neural Networks (CNNs) play a crucial role in many image recognition and classification tasks, recommender systems, brain-computer interfaces, etc. As a consequence, there is a notable interest in developing high performance realizations of the convolution operators, which concentrate a significant portion of the computational cost of this type of neural networks. In a previous work, we introduced a portable, high performance convolution algorithm, based on the BLIS realization of matrix multiplication, which eliminates most of the runtime and memory overheads that impair the performance of the convolution operators appearing in the forward training pass, when performed via explicit im2col transform. In this paper, we extend our ideas to the full training process of CNNs on multicore processors, proposing new high performance strategies to tackle the convolution operators that are present in the more complex backward pass of the training process, while maintaining the portability of the realizations. In addition, we conduct a full integration of these algorithms into a framework for distributed training of CNNs on clusters of computers, providing a complete experimental evaluation of the actual benefits in terms of both performance and memory consumption. Compared with baseline implementation, the use of the new convolution operators using pre-allocated memory can accelerate the training by a factor of about 6%–25%, provided there is sufficient memory available. In comparison, the operator variants that do not rely on persistent memory can save up to 70% of memory.ca_CA
dc.description.sponsorShipFunding for open access charge: CRUE-Universitat Jaume I
dc.format.extent15 p.ca_CA
dc.format.mimetypeapplication/pdfca_CA
dc.language.isoengca_CA
dc.publisherElsevierca_CA
dc.publisherAcademic Pressca_CA
dc.relation.isPartOfJournal of Parallel and Distributed Computing 167 (2022) 240–254ca_CA
dc.rights0743-7315/© 2022 The Author(s). Published by Elsevier Inc. This is an open access article under the CC BY-NC-ND license (http://creativecommons.org/licenses/by-nc-nd/4.0/).ca_CA
dc.rights.urihttp://creativecommons.org/licenses/by-nc-nd/4.0/ca_CA
dc.subjectconvolutional neural networksca_CA
dc.subjectdistributed trainingca_CA
dc.subjecthigh performanceca_CA
dc.subjectPythonca_CA
dc.subjectclusters of multicore processorsca_CA
dc.titleEfficient and portable GEMM-based convolution operators for deep neural network training on multicore processorsca_CA
dc.typeinfo:eu-repo/semantics/articleca_CA
dc.identifier.doihttps://doi.org/10.1016/j.jpdc.2022.05.009
dc.rights.accessRightsinfo:eu-repo/semantics/openAccessca_CA
dc.type.versioninfo:eu-repo/semantics/publishedVersionca_CA
project.funder.nameGeneralitat Valencianaca_CA
oaire.awardNumberPID2020-113656RB-C21/C22ca_CA
oaire.awardNumberMCIN/AEI/10.13039/501100011033ca_CA
oaire.awardNumberPrometeo/2019/109ca_CA
oaire.awardNumberCDEIGENT/2018/014ca_CA


Ficheros en el ítem

Thumbnail

Este ítem aparece en la(s) siguiente(s) colección(ones)

Mostrar el registro sencillo del ítem

0743-7315/© 2022 The Author(s). Published by Elsevier Inc. This is an open access article under the CC BY-NC-ND license
(http://creativecommons.org/licenses/by-nc-nd/4.0/).
Excepto si se señala otra cosa, la licencia del ítem se describe como: 0743-7315/© 2022 The Author(s). Published by Elsevier Inc. This is an open access article under the CC BY-NC-ND license (http://creativecommons.org/licenses/by-nc-nd/4.0/).