• English
  • Deutsch
  • Log In
    Password Login
    Research Outputs
    Fundings & Projects
    Researchers
    Institutes
    Statistics
Repository logo
Fraunhofer-Gesellschaft
  1. Home
  2. Fraunhofer-Gesellschaft
  3. Scopus
  4. Demonstrating NADA: A Workflow for Distributed CNN Training on FPGA Clusters
 
  • Details
  • Full
Options
2023
Conference Paper
Title

Demonstrating NADA: A Workflow for Distributed CNN Training on FPGA Clusters

Abstract
We introduce our Network Attached Deep learning Accelerator called NADA, which consists of a novel and flexible HW/SW framework for efficient training of deep neural networks on FPGA clusters. NADA is centered around layer parallelism, instantiating a specific implementation for each layer. These implementations are placed across the desired number of network attached FPGAs in the cluster. The NADA hardware architecture relies on a high-speed UDP/IP pure hardware network stack. We demonstrate the usability of our approach by training a couple of demo networks on an Arria 10 FPGA cluster.
Author(s)
Knapheide, Justin
Fraunhofer-Institut für Nachrichtentechnik, Heinrich-Hertz-Institut HHI  
Kreowsky, Philipp
Fraunhofer-Institut für Nachrichtentechnik, Heinrich-Hertz-Institut HHI  
Stabernack, Benno  
Fraunhofer-Institut für Nachrichtentechnik, Heinrich-Hertz-Institut HHI  
Mainwork
33rd International Conference on Field-Programmable Logic and Applications, FPL 2023  
Conference
International Conference on Field-Programmable Logic and Applications 2023  
DOI
10.1109/FPL60245.2023.00068
Language
English
Fraunhofer-Institut für Nachrichtentechnik, Heinrich-Hertz-Institut HHI  
Keyword(s)
  • CNN

  • DNN

  • FPGA

  • HW/SW framework

  • Network Attached Accelerator

  • Online Normalization

  • SpinalHDL

  • UDP

  • Cookie settings
  • Imprint
  • Privacy policy
  • Api
  • Contact
© 2024