Joseph Peter.VR. AnithaS. AnusooyaP. K. JawaharE. NitheshS. SairamsivaS. K.
Deep learning models increasing computational requirements have increased the demand for specialized hardware architectures that can provide high performance while using less energy. Because of their high-power consumption, low throughput, and incapacity to handle real-time processing demands, general-purpose processors frequently fall short. In order to overcome these obstacles, this work introduces a hardware-efficient multiplier design for deep learning processing unit (DPU). To improve performance and energy efficiency, the suggested architecture combines low-power arithmetic circuits, parallel processing units, and optimized dataflow mechanisms. Neural network core operations, such as matrix computations and activation functions, are performed by dedicated hardware blocks. By minimizing data movement, an effective on-chip memory hierarchy lowers latency and power consumption. According to simulation results using industry-standard very large-scale integration (VLSI) tools, compared to traditional processors, there is a 25% decrease in latency, a 40% increase in computational throughput, and a 30% reduction in power consumption. Architecture’s scalability and modularity guarantee compatibility with a variety of deep learning applications, such as edge computing, autonomous systems, and internet of things devices.
Anakhi HazarikaSoumyajit PoddarHafizur Rahaman
Chiraz MassoudAnissa SghaierMedien ZeghidMohssen Machhout
Nithish SindheSafdar AhmedAman RaoArshiya Anjum
Uppugunduru Anil KumarSyed Ershad Ahmed
Shravani ChandakaBalaji Narayanam