Optimizing Deep Learning Acceleration on FPGA for Real-Time and Resource-Efficient Image Classification
Deep learning (DL) has revolutionized image classification, yet deploying convolutional neural networks (CNNs) on edge devices for real-time applications remains a significant challenge due to constraints in computation, memory, and power efficiency. This work presents an optimized implementation of...
Saved in:
Main Authors: | Ahmad Mouri Zadeh Khaki, Ahyoung Choi |
---|---|
Format: | Article |
Language: | English |
Published: |
MDPI AG
2025-01-01
|
Series: | Applied Sciences |
Subjects: | |
Online Access: | https://www.mdpi.com/2076-3417/15/1/422 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Similar Items
-
Efficient Hardware Implementation of a Multi-Layer Gradient-Free Online-Trainable Spiking Neural Network on FPGA
by: Ali Mehrabi, et al.
Published: (2024-01-01) -
An efficient loop tiling framework for convolutional neural network inference accelerators
by: Hongmin Huang, et al.
Published: (2022-01-01) -
Accelerating Deep Neural Networks implementation: A survey
by: Meriam Dhouibi, et al.
Published: (2021-03-01) -
A Hardware Accelerator for the Inference of a Convolutional Neural network
by: Edwin González, et al.
Published: (2019-11-01) -
FPGA-QNN: Quantized Neural Network Hardware Acceleration on FPGAs
by: Mustafa Tasci, et al.
Published: (2025-01-01)