Memory Access Dataflow

Loading...
Thumbnail Image

Authors

Sankaralingam, Karthikeyan
Kim, Sung Jin
Ho, Chen-Han

Advisors

License

DOI

Type

Technical Report

Journal Title

Journal ISSN

Volume Title

Publisher

Grantor

Abstract

Specialization and accelerators are an effective way to address the slowdown of Dennard scaling. For a family of accelerators like DySER, NPU, CE, and SSE acceleration that rely on a high performance processor to interface with memory using a decoupled access/execute paradigm, the power/energy benefits of acceleration are curtailed by the host processor?s power consumption. We observe that the host processor is essentially performing three primitive tasks: i) computation to generate recurring address patterns/branches; ii) managing and triggering recurring events like arrival of value from cache, value from accelerator etc.; iii) actions to move information from one place to another; and iv) the above three are recurring and occur concurrently. Its overarching role is to orchestrate memory access dataflow. A conventional OOO processor is power-inefficient and over-provisioned for this. We observe that exposing these low level events, actions, and computation enables an efficient dataflow microarchitecture to build a memory access dataflow engine. We propose a new architecture/execution-model called memory access dataflow (MAD) that is built on these primitive tasks, exposes them in the MAD ISA, and an accompanying efficient microarchitecture.

Description

Related Material and Data

Citation

TR1802

Sponsorship

Endorsement

Review

Supplemented By

Referenced By