Tolerating Defects in Low-Power Neural Network Accelerators Via Retraining-Free Weight Approximation
| Author(s) | Hosseini, Fateme S. | |
| Author(s) | Meng, Fanruo | |
| Author(s) | Yang, Chengmo | |
| Author(s) | Wen, Wujie | |
| Author(s) | Cammarota, Rosario | |
| Date Accessioned | 2022-01-12T18:28:32Z | |
| Date Available | 2022-01-12T18:28:32Z | |
| Publication Date | 2021-09-23 | |
| Description | This article was originally published in ACM Transactions on Embedded Computing Systems. The version of record is available at: https://doi.org/10.1145/3477016 | en_US |
| Abstract | Hardware accelerators are essential to the accommodation of ever-increasing Deep Neural Network (DNN) workloads on the resource-constrained embedded devices. While accelerators facilitate fast and energy-efficient DNN operations, their accuracy is threatened by faults in their on-chip and off-chip memories, where millions of DNN weights are held. The use of emerging Non-Volatile Memories (NVM) further exposes DNN accelerators to a non-negligible rate of permanent defects due to immature fabrication, limited endurance, and aging. To tolerate defects in NVM-based DNN accelerators, previous work either requires extra redundancy in hardware or performs defect-aware retraining, imposing significant overhead. In comparison, this paper proposes a set of algorithms that exploit the flexibility in setting the fault-free bits in weight memory to effectively approximate weight values, so as to mitigate defect-induced accuracy drop. These algorithms can be applied as a one-step solution when loading the weights to embedded devices. They only require trivial hardware support and impose negligible run-time overhead. Experiments on popular DNN models show that the proposed techniques successfully boost inference accuracy even in the face of elevated defect rates in the weight memory. | en_US |
| Citation | Fateme S. Hosseini, Fanruo Meng, Chengmo Yang, Wujie Wen, and Rosario Cammarota. 2021. Tolerating Defects in Low-Power Neural Network Accelerators Via Retraining-Free Weight Approximation. ACM Trans. Embedd. Comput. Syst. 20, 5s, Article 85 (September 2021), 21 pages. https://doi.org/10.1145/3477016 | en_US |
| ISSN | 1558-3465 | |
| URL | https://udspace.udel.edu/handle/19716/29964 | |
| Language | en_US | en_US |
| Publisher | ACM Transactions on Embedded Computing Systems | en_US |
| Keywords | Computer systems organization | en_US |
| Keywords | Reliability | en_US |
| Keywords | Neural networks | en_US |
| Keywords | Embedded software | en_US |
| Keywords | Hardware | en_US |
| Keywords | Error detection | en_US |
| Keywords | error correction | en_US |
| Keywords | Neural network accelerator | en_US |
| Keywords | defect tolerance | en_US |
| Keywords | memory faults | en_US |
| Keywords | approximation | en_US |
| Title | Tolerating Defects in Low-Power Neural Network Accelerators Via Retraining-Free Weight Approximation | en_US |
| Type | Article | en_US |
Files
Original bundle
1 - 1 of 1
Loading...
- Name:
- Tolerating Defects in Low-Power Neural Network.pdf
- Size:
- 5.68 MB
- Format:
- Adobe Portable Document Format
- Description:
- Main article
License bundle
1 - 1 of 1
No Thumbnail Available
- Name:
- license.txt
- Size:
- 124 B
- Format:
- Item-specific license agreed upon to submission
- Description:
