webinaire à la demande

Cornell University: Building Sparse Linear Algebra Accelerators with HLS

Durée estimée : 72 minutes

Partager

Cornell intros HiSparse: accelerator on sparse-matrix dense-vector multiplication. Using both HLS implementation and simulation, their sparse accelerators deliver promising speedup with increased bandwidth and energy efficiency.

Sparse linear algebra (SLA) operations are essential in many applications such as data analytics, graph processing, machine learning, and scientific computing. However, it is challenging to build efficient hardware accelerators for SLA operations since they typically exhibit low operational intensity and irregular compute and data access patterns. In particular, some of these challenges are not well studied in the context of High-Level Synthesis (HLS). 
 
In this talk, we first introduce HiSparse, an accelerator on sparse-matrix dense-vector multiplication (SpMV). To achieve a high bandwidth utilization, we co-design the sparse storage format and the accelerator architecture. We further demonstrate the use of Catapult HLS to build a high-throughput pipeline that can handle irregular data dependencies and access patterns. Building on our SpMV accelerator, we further develop a versatile sparse accelerator that can support multiple SLA operations with run-time configurability to support different compute patterns. Our architecture design is guided by a novel analytical model which enables rapid exploration of the design configuration search space. According to our evaluation using both HLS implementation and simulation, our sparse accelerators deliver promising speedup with increased bandwidth and energy efficiency when compared to CPU and GPU executions.

À propos de l'intervenant

Cornell University

Yixiao Du

Ph.D. student

Yixiao is a second-year Ph.D. student at Cornell Computer Systems Lab, advised by Prof. Zhiru Zhang. His current research interests include efficient acceleration of sparse workloads with high-level synthesis. He received his bachelor’s degree in Microelectronics Science and Engineering from the University of Electronic Science and Technology of China.

Ressources associées

Infineon & Coseda: Facelifting a SystemC System Level Model Towards Physical Prototype – Adoption of High-Level-Synthesis
Webinar

Infineon & Coseda: Facelifting a SystemC System Level Model Towards Physical Prototype – Adoption of High-Level-Synthesis

Infineon & Coseda present on the adoption of High-Level-Synthesis at an existing SystemC system level model.

STMicroelectronics: A Common C++ and UVM Verification Flow of High-Level IP
Webinar

STMicroelectronics: A Common C++ and UVM Verification Flow of High-Level IP

STMicro presents a unified way to integrate the definition of RTL and C functional coverage and assertion (reducing the coding effort) and a method to add constraints to the random values generated in UVMF.

CEA: Bridging the Gap Between Neural Network Exploration and Hardware Implementation
Webinar

CEA: Bridging the Gap Between Neural Network Exploration and Hardware Implementation

CEA presents a methodology that bridges the open-source DL framework N2D2 and Catapult HLS to help reducing the design process of hardware accelerators, making it possible to keep pace with new AI algorithms.

High-Level Synthesis & Advanced RTL Power Optimization – Are you still missing out?
Webinar

High-Level Synthesis & Advanced RTL Power Optimization – Are you still missing out?

Discover how C++ & SystemC/MatchLib HLS is more than just converting SystemC to RTL. In the RTL Design space, we will cover our technology for Power Optimization with PowerPro Designer & Optimizer.

Alibaba: Innovating Agile Hardware Development with Catapult HLS
Webinar

Alibaba: Innovating Agile Hardware Development with Catapult HLS

At the IP level, an ISP was created within a year using Catapult, a task impossible using traditional RTL. To reduce dependency on designer experience, Alibaba introduced an AI-assisted DSE tool.