Identificador persistente para citar o vincular este elemento: http://hdl.handle.net/10662/21063
Registro completo de Metadatos
Campo DCValoridioma
dc.contributor.authorMoreno Álvarez, Sergio-
dc.contributor.authorPaoletti Ávila, Mercedes Eugenia-
dc.contributor.authorCavallaro, Gabriele-
dc.contributor.authorHaut Hurtado, Juan Mario-
dc.date.accessioned2024-04-19T10:23:15Z-
dc.date.available2024-04-19T10:23:15Z-
dc.date.issued2023-
dc.descriptionVersión aceptada del artículo publicado en IEEE Transactions on Neural Networks and Learning Systems, pp. 1-15.2023. ISSN 2162237X-
dc.description.abstractThe amount of data needed to effectively train modern deep neural architectures has grown significantly, leading to increased computational requirements. These intensive computations are tackled by the combination of last generation computing resources, such as accelerators, or classic processing units. Nevertheless, gradient communication remains as the major bottleneck, hindering the efficiency notwithstanding the improvements in runtimes obtained through data parallelism strategies. Data parallelism involves all processes in a global exchange of potentially high amount of data, which may impede the achievement of the desired speedup and the elimination of noticeable delays or bottlenecks. As a result, communication latency issues pose a significant challenge that profoundly impacts the performance on distributed platforms. This research presents node-based optimization steps to significantly reduce the gradient exchange between model replicas whilst ensuring model convergence. The proposal serves as a versatile communication scheme, suitable for integration into a wide range of general-purpose deep neural network (DNN) algorithms. The optimization takes into consideration the specific location of each replica within the platform. To demonstrate the effectiveness, different neural network approaches and datasets with disjoint properties are used. In addition, multiple types of applications are considered to demonstrate the robustness and versatility of our proposal. The experimental results show a global training time reduction whilst slightly improving accuracy. Code: https://github.com/mhaut/eDNNcom.es_ES
dc.description.sponsorshipThis work was supported in part by the Consejería de Economía, Ciencia y Agenda Digital of the Junta de Extremadura, in part by the European Regional Development Fund (ERDF) of the European Union under Grant GR21040, Grant GR21099 and Grant IB20040, in part by the Spanish Ministerio de Ciencia e Innovacion under Project PID2019-110315RB-I00 (APRISA), in part by the DEEP-EST Project (computing resources), and in part by the European Union’s Horizon 2020 Research and Innovation Programme under Grant 754304.es_ES
dc.format.extent15 p.es_ES
dc.format.mimetypeapplication/pdfen_US
dc.language.isoenges_ES
dc.rightsAttribution-NonCommercial-NoDerivatives 4.0 International*
dc.rights.urihttp://creativecommons.org/licenses/by-nc-nd/4.0/*
dc.subjectParalelismo de datoses_ES
dc.subjectEnseñanza profundaes_ES
dc.subjectComputación de alto rendimientoes_ES
dc.subjectRedes neuronaleses_ES
dc.subjectComunicación síncronaes_ES
dc.subjectData parallelismes_ES
dc.subjectDeep learninges_ES
dc.subjectHigh-performance computinges_ES
dc.subjectNeural networkses_ES
dc.subjectSynchronous communicationses_ES
dc.titleEnhancing Distributed Neural Network Training through Node-Based Communicationses_ES
dc.typepreprintes_ES
dc.description.versionpeerReviewedes_ES
europeana.typeTEXTen_US
dc.rights.accessRightsopenAccesses_ES
dc.subject.unesco2490.02 Neuroquímicaes_ES
dc.subject.unesco2490 Neurocienciases_ES
europeana.dataProviderUniversidad de Extremadura. Españaes_ES
dc.identifier.bibliographicCitationS. Moreno-Álvarez, M. E. Paoletti, G. Cavallaro and J. M. Haut, "Enhancing Distributed Neural Network Training Through Node-Based Communications," in IEEE Transactions on Neural Networks and Learning Systems, doi: 10.1109/TNNLS.2023.3309735es_ES
dc.type.versionacceptedVersiones_ES
dc.contributor.affiliationUniversidad de Extremadura. Departamento de Ingeniería de Sistemas Informáticos y Telemáticoses_ES
dc.contributor.affiliationUniversidad de Extremadura. Departamento de Tecnología de los Computadores y de las Comunicacioneses_ES
dc.relation.publisherversionhttps://ieeexplore.ieee.org/document/10254237es_ES
dc.identifier.doi10.1109/TNNLS.2023.3309735-
dc.identifier.publicationfirstpage1es_ES
dc.identifier.publicationlastpage15es_ES
dc.identifier.orcid0000-0002-1858-9920es_ES
dc.identifier.orcid0000-0003-1030-3729es_ES
dc.identifier.orcid0000-0002-3239-9904es_ES
dc.identifier.orcid0000-0001-6701-961Xes_ES
Colección:DISIT - Artículos
DTCYC - Artículos

Archivos
Archivo Descripción TamañoFormato 
TNNLS_2023_3309735.pdf3,11 MBAdobe PDFDescargar


Este elemento está sujeto a una licencia Licencia Creative Commons Creative Commons