:::

TIGP (AIoT) -- Enabling Model Parallelism for Neural Network Training Based on Decoupled Local Losses


  • 講者 : 陳弘軒 教授
  • 日期 : 2024/03/01 (Fri.) 14:00~16:00
  • 地點 : 資訊所新館106演講廳
  • 邀請人 : TIGP (AIoT)
Abstract
Backpropagation (BP) is foundational in deep learning. However, its inefficiency is partially caused by backward locking, making simultaneous gradient computation across layers difficult and reducing training efficiency. In this talk, I will introduce our recent research on simultaneously computing parameter gradients in different layers through pipelining. This approach improves the training efficiency while preserving testing accuracies comparable to BP-trained models.