Acceleration for ML 论文导读

Energy efficient parallel neuromorphic architectures with approximate arithmetic on FPGA

Motivation

To address the slow operation and high energy and resource consumption problem caused by realizing spiking neural network (SNN) using software.

Problem

  1. software : slow operation, high energy consumption and space resources
  2. analog circuits: hard to reconfigure and intrinsically sensitive to process, voltage and temperature (PVT) Var.
  3. FPGA: most of works focus on the acceleration of SNN without considering energy consumption and efficiency of resource utilization.
  4. This work presented the parallel neuromorphic processor architectures with approximate arithmetic for SNN on FPGA.

There is no related work part in this paper.


In-Datacenter Performance Analysis of a Tensor Processing Unit

Motivation

This paper evaluates a custom ASIC - called a Tensor Processing Unit (TPU) to accelerates the inference phase of neural networks (NN).

Problem

Many NN applications have hard response time deadline. Hence, inference phase must response quickly when user do some action. While CPU and GPU are poor in response.

All works are focus on hardware processing, such as DRAM, hardware protocol and so on.

posted @ 2019-03-21 21:04  考拉小无  阅读(209)  评论(0编辑  收藏  举报