The evolution of Deep Learning (DL) algorithms, coupled with the availability of vast amounts of data, has enabled superior accuracy across multiple Artificial Intelligence (AI) tasks. However, such accuracy incurs substantial computational and energy costs, limiting deployment on resource-constrained embedded devices. Approximate Computing has emerged as a promising paradigm to mitigate this challenge by trading off tolerable accuracy loss for significant gains in energy, power, area, and performance. Leveraging the inherent error resilience of DL models, approximation techniques have been extensively explored at multiple layers of the DL computing stack, and several surveys also exist. However, to the best of our knowledge, no prior Systematic Literature Review (SLR) exists that is specifically focused on Approximate Computing for DL in Embedded Systems. This article presents SLR by selecting 51 studies published between January 2019 and May 2025. These studies are organized into four categories corresponding to layers of the DL computing stack, including cross-layer approximations. Furthermore, 12 approximation techniques, 11 benchmark datasets, 26 prominent DL architectures, and three embedded platforms for the energy-efficient DL deployment are summarized, along with the leading tools, frameworks, and optimization/evaluation metrics. Additionally, the academic and industry-wide global adoption has also been highlighted. This SLR reveals that Approximate Multipliers, CIFAR-10 and MNIST datasets, ResNet-50 and LeNet models, and ASIC-based 45 nm technology designs are most frequently targeted, whereas Synopsys DC, PyTorch, and Verilog emerged as the predominant tools and frameworks supporting approximations in DL. Finally, this SLR highlights the current challenges, research gaps, and introduces a conceptual cross-layer approximate DL framework to address them. The findings of this SLR offer a comprehensive reference for the researchers and practitioners to select suitable approximation techniques along with the underlying tools and hardware platforms, aligned with their application requirements.

Rasheed, Y., Anwar, M.w., Gillani, G.a., Ottavi, M. (2025). Toward approximate computing for deep learning in embedded systems: a systematic literature review. IEEE ACCESS, 13, 210863-210891 [10.1109/ACCESS.2025.3641821].

Toward approximate computing for deep learning in embedded systems: a systematic literature review

Ottavi, Marco
2025-01-01

Abstract

The evolution of Deep Learning (DL) algorithms, coupled with the availability of vast amounts of data, has enabled superior accuracy across multiple Artificial Intelligence (AI) tasks. However, such accuracy incurs substantial computational and energy costs, limiting deployment on resource-constrained embedded devices. Approximate Computing has emerged as a promising paradigm to mitigate this challenge by trading off tolerable accuracy loss for significant gains in energy, power, area, and performance. Leveraging the inherent error resilience of DL models, approximation techniques have been extensively explored at multiple layers of the DL computing stack, and several surveys also exist. However, to the best of our knowledge, no prior Systematic Literature Review (SLR) exists that is specifically focused on Approximate Computing for DL in Embedded Systems. This article presents SLR by selecting 51 studies published between January 2019 and May 2025. These studies are organized into four categories corresponding to layers of the DL computing stack, including cross-layer approximations. Furthermore, 12 approximation techniques, 11 benchmark datasets, 26 prominent DL architectures, and three embedded platforms for the energy-efficient DL deployment are summarized, along with the leading tools, frameworks, and optimization/evaluation metrics. Additionally, the academic and industry-wide global adoption has also been highlighted. This SLR reveals that Approximate Multipliers, CIFAR-10 and MNIST datasets, ResNet-50 and LeNet models, and ASIC-based 45 nm technology designs are most frequently targeted, whereas Synopsys DC, PyTorch, and Verilog emerged as the predominant tools and frameworks supporting approximations in DL. Finally, this SLR highlights the current challenges, research gaps, and introduces a conceptual cross-layer approximate DL framework to address them. The findings of this SLR offer a comprehensive reference for the researchers and practitioners to select suitable approximation techniques along with the underlying tools and hardware platforms, aligned with their application requirements.
2025
Pubblicato
Rilevanza internazionale
Articolo
Esperti anonimi
Settore IINF-01/A - Elettronica
English
Approximate CNN; Approximate computing; Approximate DL; Approximate multipliers; Deep learning; Energy-efficient DL; Inexact computing; Systematic literature review
Rasheed, Y., Anwar, M.w., Gillani, G.a., Ottavi, M. (2025). Toward approximate computing for deep learning in embedded systems: a systematic literature review. IEEE ACCESS, 13, 210863-210891 [10.1109/ACCESS.2025.3641821].
Rasheed, Y; Anwar, Mw; Gillani, Ga; Ottavi, M
Articolo su rivista
File in questo prodotto:
Non ci sono file associati a questo prodotto.

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/2108/453266
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 0
  • ???jsp.display-item.citation.isi??? 0
social impact