Near-Memory acceleration of Convolutional Neural Networks by exploiting Parallelism, Sparsity, and Redundancy

dc.contributor.authorDas, Palash
dc.date.accessioned2023-03-01T12:55:44Z
dc.date.accessioned2023-10-20T04:37:20Z
dc.date.available2023-03-01T12:55:44Z
dc.date.available2023-10-20T04:37:20Z
dc.date.issued2022
dc.descriptionSupervisor: Kapoor, Hemangee Ken_US
dc.description.abstractThe gap between the processing speed of the CPU and the access speed of the memory is becoming a bottleneck for many emerging applications. This gap can be reduced if the computation can be taken closer to the memory through near-memory processing (NMP). Among the logic options, application-specific integrated circuits (ASICs) are highly efficient in terms of power and area overhead for NMP logic integration. In this thesis, we aim to accelerate Convolutional Neural Networks (CNNs) by integrating custom hardware near the memory. As CNNs are widely used in several emerging applications, the designed hardware can be extensively used in all such cases. To design an NMP-based system with high performance and energy efficiency, we explore various techniques such as leveraging parallelism, exploiting data sparsity, and utilizing computation redundancy to reduce the number of operations. All such techniques result in hardware designs that implement the appropriate data flow and data-parallel algorithm. The designs have positively impacted the system's performance and energy efficiency. To examine the deployability of the NMP approach, we perform experiments on various memory technologies like 3D memory, hybrid memory, and the commodity DRAM. Additionally, we also measure the efficacy of NMP for other applications like database operations. The proposed systems have performed substantially well while comparing them with various baselines and state-of-the-art works.en_US
dc.identifier.otherROLL NO.156101001
dc.identifier.urihttps://gyan.iitg.ac.in/handle/123456789/2306
dc.language.isoenen_US
dc.relation.ispartofseriesTH-2870;
dc.subjectNear-memory Processingen_US
dc.subjectConvolutional Neural Networksen_US
dc.subjectAccelerated Architecturesen_US
dc.subjectCNN Acceleratorsen_US
dc.titleNear-Memory acceleration of Convolutional Neural Networks by exploiting Parallelism, Sparsity, and Redundancyen_US
dc.typeThesisen_US
Files
Original bundle
Now showing 1 - 2 of 2
No Thumbnail Available
Name:
Abstract-TH-2870_156101001.pdf
Size:
104.85 KB
Format:
Adobe Portable Document Format
Description:
ABSTRACT
No Thumbnail Available
Name:
TH-2870_156101001.pdf
Size:
11.06 MB
Format:
Adobe Portable Document Format
Description:
THESIS
License bundle
Now showing 1 - 1 of 1
No Thumbnail Available
Name:
license.txt
Size:
1.71 KB
Format:
Plain Text
Description: