Результаты исследований: Публикации в книгах, отчётах, сборниках, трудах конференций › статья в сборнике материалов конференции › научная
Speedup of deep neural network learning on the MIC-architecture. / Milova, E.; Sveshnikova, S.; Gankevich, I.
International Conference on High Performance Computing Simulation (HPCS'16). Institute of Electrical and Electronics Engineers Inc., 2016. стр. 989-992.Результаты исследований: Публикации в книгах, отчётах, сборниках, трудах конференций › статья в сборнике материалов конференции › научная
}
TY - GEN
T1 - Speedup of deep neural network learning on the MIC-architecture
AU - Milova, E.
AU - Sveshnikova, S.
AU - Gankevich, I.
PY - 2016
Y1 - 2016
N2 - Deep neural networks are more accurate, but require more computational power in the learning process. Moreover, it is an iterative process. The goal of the research is to investigate efficiency of solving this problem on MIC architecture without changing baseline algorithm. Well-known code vectorization and parallelization methods are used to increase the effectiveness of the program on MIC architecture. In the course of the experiments we test two coprocessor data transfer models: explicit and implicit one. We show that implicit memory copying is more efficient than explicit one, because only modified memory blocks are copied. MIC architecture shows competitive performance compared to multi-core ×86 processor.
AB - Deep neural networks are more accurate, but require more computational power in the learning process. Moreover, it is an iterative process. The goal of the research is to investigate efficiency of solving this problem on MIC architecture without changing baseline algorithm. Well-known code vectorization and parallelization methods are used to increase the effectiveness of the program on MIC architecture. In the course of the experiments we test two coprocessor data transfer models: explicit and implicit one. We show that implicit memory copying is more efficient than explicit one, because only modified memory blocks are copied. MIC architecture shows competitive performance compared to multi-core ×86 processor.
KW - many-core architecture
KW - DNN
KW - optimisation
KW - parallel computing
KW - vectorization
KW - offload
KW - Xeon Phi
KW - coprocessor
UR - https://www.scopus.com/inward/record.uri?eid=2-s2.0-84991734860&doi=10.1109%2fHPCSim.2016.7568443&partnerID=40&md5=46fb21a706001b7a2fc2402d08dd0e81
U2 - 10.1109/HPCSim.2016.7568443
DO - 10.1109/HPCSim.2016.7568443
M3 - Conference contribution
SN - 978-1-5090-2088-1
SP - 989
EP - 992
BT - International Conference on High Performance Computing Simulation (HPCS'16)
PB - Institute of Electrical and Electronics Engineers Inc.
ER -
ID: 7632736