Identificador persistente para citar o vincular este elemento: http://hdl.handle.net/10662/20373
Registro completo de Metadatos
Campo DCValoridioma
dc.contributor.authorMoreno Álvarez, Sergio-
dc.contributor.authorHaut Hurtado, Juan Mario-
dc.contributor.authorPaoletti Ávila, Mercedes Eugenia-
dc.contributor.authorRico Gallego, Juan Antonio-
dc.date.accessioned2024-02-07T18:55:51Z-
dc.date.available2024-02-07T18:55:51Z-
dc.date.issued2021-
dc.identifier.citationSergio Moreno-Alvarez, Juan M. Haut, Mercedes E. Paoletti, Juan A. Rico-Gallego, Heterogeneous model parallelism for deep neural networks, Neurocomputing, Volume 441, 2021, Pages 1-12, ISSN 0925-2312, https://doi.org/10.1016/j.neucom.2021.01.125 (https://www.sciencedirect.com/science/article/pii/S0925231221002320)-
dc.identifier.issn0925-2312-
dc.identifier.urihttp://hdl.handle.net/10662/20373-
dc.description.abstractDeep neural networks (DNNs) have transformed computer vision, establishing themselves as the current state-of-the-art for image processing. Nevertheless, the training of current large DNN models is one of the main challenges to be solved. In this sense, data-parallelism has been the most widespread distributed training strategy since it is easy to program and can be applied to almost all cases. However, this solution suffers from several limitations, such as its high communication requirements and the memory con- straints when training very large models. To overcome these limitations model-parallelism has been pro- posed, solving the most substantial problems of the former strategy. However, describing and implementing the parallelization of the training of a DNN model across a set of processes deployed on several devices is a challenging task. Current proposed solutions assume a homogeneous distribution, being impractical when working with devices of different computational capabilities, which is quite com- mon on high performance computing platforms. To address previous shortcomings, this work proposes a novel model-parallelism technique considering heterogeneous platforms, where a load balancing mech- anism between uneven devices of an HPC platform has been implemented. Our proposal takes advantage of the Google Brain’s Mesh-TensorFlow for convolutional networks, splitting computing tensors across filter dimension in order to balance the computational load of the available devices. Conducted experi- ments show an improvement in the exploitation of heterogeneous computational resources, enhancing the training performance. The code is available on: https://github.com/mhaut/HeterogeneusModelDNN.es_ES
dc.description.sponsorshipSupported by (1) The European Regional Development Fund ’A way to achieve Europe’ (ERDF) and the Extremadura Local Government (Ref. IB16118); (2) The Spanish Ministry of Science and Innovation (Ref. PID2019-110315RB-I00 APRISA); and (3) The computing facilities of Extremadura Research Centre for Advanced Technologies (CETA-CIEMAT), funded by the European Regional Development Fund (ERDF). CETA-CIEMAT belongs to CIEMAT and the Government of Spain.-
dc.format.extent12 p.es_ES
dc.format.mimetypeapplication/pdfen_US
dc.language.isoenges_ES
dc.publisherElsevieres_ES
dc.subjectDeep learninges_ES
dc.subjectHigh performance computinges_ES
dc.subjectDistributed traininges_ES
dc.subjectHeterogeneous platformses_ES
dc.subjectModel parallelismes_ES
dc.subjectHPC-
dc.subjectComputación de alto rendimiento-
dc.subjectEntrenamiento distribuido-
dc.subjectPlataformas de computación heterogénea-
dc.titleHeterogeneous model parallelism for deep neural networkses_ES
dc.typearticlees_ES
dc.description.versionpeerReviewedes_ES
europeana.typeTEXTen_US
dc.rights.accessRightsclosedAccesses_ES
dc.subject.unesco1203 Ciencia de Los Ordenadores-
europeana.dataProviderUniversidad de Extremadura. Españaes_ES
dc.type.versionpublishedVersiones_ES
dc.contributor.affiliationUniversidad de Extremadura. Departamento de Ingeniería de Sistemas Informáticos y Telemáticoses_ES
dc.contributor.affiliationUniversidad Nacional de Educación a Distancia-
dc.contributor.affiliationUniversidad de Extremadura. Departamento de Tecnología de los Computadores y de las Comunicaciones-
dc.relation.publisherversionhttps://www.sciencedirect.com/science/article/pii/S0925231221002320?via%3Dihubes_ES
dc.identifier.doihttps://doi.org/10.1016/j.neucom.2021.01.125-
dc.identifier.publicationtitleNeurocomputinges_ES
dc.identifier.publicationissue44es_ES
dc.identifier.publicationfirstpage1es_ES
dc.identifier.publicationlastpage12es_ES
dc.identifier.publicationvolume441es_ES
dc.identifier.orcid0000-0002-4264-7473es_ES
dc.identifier.orcid0000-0002-1858-9920-
dc.identifier.orcid0000-0001-6701-961X-
dc.identifier.orcid0000-0003-1030-3729-
Colección:DIEEA - Artículos
DISIT - Artículos
DTCYC - Artículos

Archivos
Archivo Descripción TamañoFormato 
j_neucom_2021_01_125.pdf
???org.dspace.app.webui.jsptag.ItemTag.accessRestricted???
1,78 MBAdobe PDFDescargar    Pide una copia


Este elemento está sujeto a una licencia Licencia Creative Commons Creative Commons