Memory Access Dataflow
| dc.contributor.author | Sankaralingam, Karthikeyan | |
| dc.contributor.author | Kim, Sung Jin | |
| dc.contributor.author | Ho, Chen-Han | |
| dc.date.accessioned | 2014-03-18T14:26:52Z | |
| dc.date.available | 2014-03-18T14:26:52Z | |
| dc.date.issued | 2014-03-07 | |
| dc.description.abstract | Specialization and accelerators are an effective way to address the slowdown of Dennard scaling. For a family of accelerators like DySER, NPU, CE, and SSE acceleration that rely on a high performance processor to interface with memory using a decoupled access/execute paradigm, the power/energy benefits of acceleration are curtailed by the host processor?s power consumption. We observe that the host processor is essentially performing three primitive tasks: i) computation to generate recurring address patterns/branches; ii) managing and triggering recurring events like arrival of value from cache, value from accelerator etc.; iii) actions to move information from one place to another; and iv) the above three are recurring and occur concurrently. Its overarching role is to orchestrate memory access dataflow. A conventional OOO processor is power-inefficient and over-provisioned for this. We observe that exposing these low level events, actions, and computation enables an efficient dataflow microarchitecture to build a memory access dataflow engine. We propose a new architecture/execution-model called memory access dataflow (MAD) that is built on these primitive tasks, exposes them in the MAD ISA, and an accompanying efficient microarchitecture. | en |
| dc.identifier.citation | TR1802 | en |
| dc.identifier.uri | http://digital.library.wisc.edu/1793/68516 | |
| dc.subject | Dataflow | en |
| dc.subject | Accelerators | en |
| dc.subject | Memory | en |
| dc.subject | computer architecture | en |
| dc.title | Memory Access Dataflow | en |
| dc.type | Technical Report | en |