Enabling Efficient Inference of Convolutional Neural Networks via Approximation

0Citations
Citations of this article
1Readers
Mendeley users who have this article in their library.
Get full text

Abstract

With the rapid advancements in the Artificial Intelligence area, Neural Networks (NNs) became the driving force both in general purpose and embedded computing domains. Especially, resource constrained embedded systems progressively rely on multiple NNs to provide on the spot sophisticated services. Nevertheless, supporting NN-based workloads is challenging due to the enormous computational and energy requirements. Exploiting the inherent error resiliency of NNs, significant research focuses on designing approximate Convolutional NN (CNN) inference accelerators, demonstrating that, for negligible accuracy loss, they satisfy tight latency, power, and temperature constraints. This chapter provides a comprehensive discussion of different aspects of approximate CNN implementations.

Cite

CITATION STYLE

APA

Zervakis, G., Anagnostopoulos, I., Amrouch, H., & Henkel, J. (2022). Enabling Efficient Inference of Convolutional Neural Networks via Approximation. In Approximate Computing (pp. 429–450). Springer International Publishing. https://doi.org/10.1007/978-3-030-98347-5_17

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free