Parallelization of Recurrent Neural Network training algorithm with implicit aggregation on multi-core architectures - Unité de modélisation mathématique et informatique des systèmes complexes Accéder directement au contenu
Pré-Publication, Document De Travail Année : 2024

Parallelization of Recurrent Neural Network training algorithm with implicit aggregation on multi-core architectures

Résumé

Recent work has shown that deep learning algorithms are efficient for various tasks, whether in Natural Language Processing (NLP) or in Computer Vision (CV). One of the particularities of these algorithms is that they are so efficient as the amount of data used is large. However, sequential execution of these algorithms on large amounts of data can take a very long time. In this paper, we consider the problem of training Recurrent Neural Network (RNN) for hate (aggressive) messages detection task. We first compared the sequential execution of three variants of RNN, we have shown that Long Short Time Memory (LSTM) provides better metric performance, but implies more important execution time in comparison with Gated Recurrent Unit (GRU) and standard RNN. To have both good metric performance and reduced execution time, we proceeded to a parallel implementation of the training algorithms. We proposed a parallel algorithm based on an implicit aggregation strategy in comparison to the existing approach which is based on a strategy with an aggregation function. We have shown that the convergence of this proposed parallel algorithm is close to that of the sequential algorithm. The experimental results on an 32-core machine at 1.5 GHz and 62 Go of RAM show that better results are obtained with the parallelization strategy that we proposed. For example, with an LSTM on a dataset having more than 100k comments, we obtained an f-measure of 0.922 and a speedup of 7 with our approach, compared to a f-measure of 0.874 and a speedup of 5 with an explicit aggregation between workers.
Fichier principal
Vignette du fichier
ARIMA_messi_nzekon_onana_11_04_2024.pdf (506.21 Ko) Télécharger le fichier
Origine : Fichiers produits par l'(les) auteur(s)

Dates et versions

hal-04542984 , version 1 (11-04-2024)

Licence

Paternité

Identifiants

  • HAL Id : hal-04542984 , version 1

Citer

Thomas Messi Nguelé, Armel Jacques Nzekon Nzeko'o, Damase Donald Onana. Parallelization of Recurrent Neural Network training algorithm with implicit aggregation on multi-core architectures. 2024. ⟨hal-04542984⟩
0 Consultations
0 Téléchargements

Partager

Gmail Facebook X LinkedIn More