JOURNAL ARTICLE

An Efficient Convolutional Neural Network Accelerator on FPGA Platform

Abstract

In conventional neural network computation, data movement constitutes a substantial part of the dataflow. Rather than primarily relying on time-consuming MAC operations, this study emphasizes data inference within the memory, significantly reducing the data volumes that must be repeatedly read by computational units, which often imposes a processing burden on the network. The hardware architecture in this study is purposefully designed to minimize the time consumed by data transfer, achieved through strategic buffer configurations, efficient data transfer between computational units, and meticulous dataflow organization. Notably, this research places a strong emphasis on optimizing data reuse post-reading. The implementation results detail the comprehensive steps of object detection, offering a thorough analysis of the time allocation for each image processing stage. Finally, the study concludes with a comparative evaluation against several well-known papers in the field.

Keywords:
Field-programmable gate array Convolutional neural network Computer science Embedded system Computer architecture Artificial intelligence

Metrics

0
Cited By
0.00
FWCI (Field Weighted Citation Impact)
8
Refs
0.02
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Topics

CCD and CMOS Imaging Sensors
Physical Sciences →  Engineering →  Electrical and Electronic Engineering
Image Processing Techniques and Applications
Physical Sciences →  Engineering →  Media Technology
Advanced Neural Network Applications
Physical Sciences →  Computer Science →  Computer Vision and Pattern Recognition
© 2026 ScienceGate Book Chapters — All rights reserved.