hpcs-16-mic-v2

Speedup of deep neural network learning on the MIC-architecture
git clone https://git.igankevich.com/hpcs-16-mic-v2.git
Log | Files | Refs

abstract.tex (943B)


      1 \begin{abstract}
      2 Deep neural networks are more accurate, but require more computational power in the learning process. Moreover, it is an iterative process. The goal of the research is to investigate  efficiency of solving this problem on MIC architecture without changing baseline algorithm. Well-known code vectorization and parallelization methods are used to increase the effectiveness of the program on MIC architecture. In the course of the experiments we test two coprocessor data transfer models: explicit and implicit one. We show that implicit memory copying is more efficient than explicit one, because only modified memory blocks are copied. MIC architecture shows competitive performance compared to multi-core x86 processor.
      3 \end{abstract}
      4 \vspace{0.1in}
      5 \begin{IEEEkeywords}
      6 DNN, optimisation, parallel computing, vectorization, offload, Xeon Phi, coprocessor, many-core architecture
      7 \end{IEEEkeywords}
      8 
      9 \IEEEpeerreviewmaketitle