Aktemur, Ege and Zorlutuna, Ege and Bilgili, Kaan and Bök, Tacettin Emre and Yanıkoğlu, Berrin and Mutluergil, Süha Orhun (2025) Going forward-forward in distributed deep learning. In: 13th International Conference on Networked Systems (NETYS 2025), Rabat, Morocco

going_forward_forward.pdf
Download (1MB)
Official URL: http://dx.doi.org/10.1007/978-3-032-00347-8_11
Abstract
We introduce a new approach in distributed learning, build- ing on Hinton’s Forward-Forward (FF) algorithm to speed up the train- ing of neural networks in distributed environments without losing ac- curacy. Unlike traditional methods that rely on forward and backward passes, the FF algorithm employs a dual forward pass strategy, eliminat- ing the dependency among layers required during the backpropagation period, which prevents efficient parallelization of the training process. Although the original FF algorithm focused on its ability to match the performance of the backpropagation algorithm, this work aims to re- duce the training time with pipeline parallelism. We propose three novel pipelined FF algorithms that speed up training 3.75 times on the MNIST dataset while maintaining accuracy when training a four-layer network with four compute nodes. These results show that FF is highly paral- lelizable and its potential in large-scale distributed/federated systems to enable faster training for larger and more complex models.
Item Type: | Papers in Conference Proceedings |
---|---|
Subjects: | Q Science > QA Mathematics > QA075 Electronic computers. Computer science Q Science > QA Mathematics > QA076 Computer software |
Divisions: | Faculty of Engineering and Natural Sciences > Academic programs > Biological Sciences & Bio Eng. Faculty of Engineering and Natural Sciences > Academic programs > Computer Science & Eng. Faculty of Engineering and Natural Sciences |
Depositing User: | Berrin Yanıkoğlu |
Date Deposited: | 01 Oct 2025 11:40 |
Last Modified: | 01 Oct 2025 11:40 |
URI: | https://research.sabanciuniv.edu/id/eprint/52458 |