On the Performance of Network Parallel Training in Artificial Neural
Networks
L. Ericson, и R. Mbuvha. (2017)cite arxiv:1701.05130Comment: 4 Pages, 4 Figures, 1 Table.
Аннотация
Artificial Neural Networks (ANNs) have received increasing attention in
recent years with applications that span a wide range of disciplines including
vital domains such as medicine, network security and autonomous transportation.
However, neural network architectures are becoming increasingly complex and
with an increasing need to obtain real-time results from such models, it has
become pivotal to use parallelization as a mechanism for speeding up network
training and deployment. In this work we propose an implementation of Network
Parallel Training through Cannon's Algorithm for matrix multiplication. We show
that increasing the number of processes speeds up training until the point
where process communication costs become prohibitive; this point varies by
network complexity. We also show through empirical efficiency calculations that
the speedup obtained is superlinear.
Описание
On the Performance of Network Parallel Training in Artificial Neural Networks
%0 Generic
%1 ericson2017performance
%A Ericson, Ludvig
%A Mbuvha, Rendani
%D 2017
%K deep dl large-scale networks neural
%T On the Performance of Network Parallel Training in Artificial Neural
Networks
%U http://arxiv.org/abs/1701.05130
%X Artificial Neural Networks (ANNs) have received increasing attention in
recent years with applications that span a wide range of disciplines including
vital domains such as medicine, network security and autonomous transportation.
However, neural network architectures are becoming increasingly complex and
with an increasing need to obtain real-time results from such models, it has
become pivotal to use parallelization as a mechanism for speeding up network
training and deployment. In this work we propose an implementation of Network
Parallel Training through Cannon's Algorithm for matrix multiplication. We show
that increasing the number of processes speeds up training until the point
where process communication costs become prohibitive; this point varies by
network complexity. We also show through empirical efficiency calculations that
the speedup obtained is superlinear.
@misc{ericson2017performance,
abstract = {Artificial Neural Networks (ANNs) have received increasing attention in
recent years with applications that span a wide range of disciplines including
vital domains such as medicine, network security and autonomous transportation.
However, neural network architectures are becoming increasingly complex and
with an increasing need to obtain real-time results from such models, it has
become pivotal to use parallelization as a mechanism for speeding up network
training and deployment. In this work we propose an implementation of Network
Parallel Training through Cannon's Algorithm for matrix multiplication. We show
that increasing the number of processes speeds up training until the point
where process communication costs become prohibitive; this point varies by
network complexity. We also show through empirical efficiency calculations that
the speedup obtained is superlinear.},
added-at = {2019-06-04T16:01:07.000+0200},
author = {Ericson, Ludvig and Mbuvha, Rendani},
biburl = {https://www.bibsonomy.org/bibtex/2128057d513a16fa4411883aa54fad3e0/alrigazzi},
description = {On the Performance of Network Parallel Training in Artificial Neural Networks},
interhash = {3f82773416e51ed9115bfeafbccacb2a},
intrahash = {128057d513a16fa4411883aa54fad3e0},
keywords = {deep dl large-scale networks neural},
note = {cite arxiv:1701.05130Comment: 4 Pages, 4 Figures, 1 Table},
timestamp = {2019-06-04T16:01:07.000+0200},
title = {On the Performance of Network Parallel Training in Artificial Neural
Networks},
url = {http://arxiv.org/abs/1701.05130},
year = 2017
}