With the rapid advancements in the Artificial Intelligence area, Neural Networks (NNs) became the driving force both in general purpose and embedded computing domains. Especially, resource constrained embedded systems progressively rely on multiple NNs to provide on the spot sophisticated services. Nevertheless, supporting NN-based workloads is challenging due to the enormous computational and energy requirements. Exploiting the inherent error resiliency of NNs, significant research focuses on designing approximate Convolutional NN (CNN) inference accelerators, demonstrating that, for negligible accuracy loss, they satisfy tight latency, power, and temperature constraints. This chapter provides a comprehensive discussion of different aspects of approximate CNN implementations.
CITATION STYLE
Zervakis, G., Anagnostopoulos, I., Amrouch, H., & Henkel, J. (2022). Enabling Efficient Inference of Convolutional Neural Networks via Approximation. In Approximate Computing (pp. 429–450). Springer International Publishing. https://doi.org/10.1007/978-3-030-98347-5_17
Mendeley helps you to discover research relevant for your work.