Files
hugo-cwpearson/content/publication/20190926_huang_hpec/index.md
2021-01-27 17:40:20 -07:00

2.5 KiB

+++ title = "Accelerating Sparse Deep Neural Networks on FPGAs" date = 2019-09-26T00:00:00 # Schedule page publish date. draft = false

Authors. Comma separated list, e.g. ["Bob Smith", "David Jones"].

authors = ["Sitao Huang", "Carl Pearson", "Rakesh Nagi", "Jinjun Xiong", "Deming Chen", "Wen-Mei Hwu"]

Publication type.

Legend:

0 = Uncategorized

1 = Conference paper

2 = Journal article

3 = Manuscript

4 = Report

5 = Book

6 = Book section

publication_types = ["1"]

Publication name and optional abbreviated version.

publication = "2019 IEEE High Performance Extreme Computing Conference" publication_short = "In HPEC'19"

Does this page contain LaTeX math? (true/false)

math = false

Does this page require source code highlighting? (true/false)

highlight = false

Featured image thumbnail (optional)

image_preview = ""

Is this a selected publication? (true/false)

selected = true

Projects (optional).

Associate this publication with one or more of your projects.

Simply enter your project's folder or file name without extension.

E.g. projects = ["deep-learning"] references

content/project/deep-learning/index.md.

Otherwise, set projects = [].

projects = []

Links (optional)

url_pdf = "pdf/2019_huang_hpec.pdf" url_preprint = "" url_code = "" url_dataset = "" url_project = "" url_slides = "" url_video = "" url_poster = "" url_source = ""

Featured image

To use, add an image named featured.jpg/png to your page's folder.

[image]

Caption (optional)

caption = ""

Focal point (optional)

Options: Smart, Center, TopLeft, Top, TopRight, Left, Right, BottomLeft, Bottom, BottomRight

focal_point = "" +++ Deep neural networks (DNNs) have been widely adopted in many domains, including computer vision, natural language processing, and medical care. Recent research revealsthat sparsity in DNN parameters can be exploited to reduce inference computational complexity and improve network quality. However, sparsity also introduces irregularity and extra complexity in data processing, which make the accelerator design challenging. This work presents the design and implementation of a highly flexible sparse DNN inference accelerator on FPGA.Our proposed inference engine can be easily configured to beused in both mobile computing and high-performance computing scenarios. Evaluation shows our proposed inference engine effectively accelerates sparse DNNs and outperforms CPU solution by up to 4.7x in terms of energy efficiency.