Reading notes and references on CGRA

Jackcui NJU Loser

to achieve high compute efficiency, overhead of adaptability should be reduced, while still supporting:

  1. Single instruction steady state, e.g., single-cycle loops.
  2. Data transport reduction, e.g., explicit bypassing, local register files, and small memories close to function units.
  3. Application tailored exploitation of parallelism, e.g., VLIW with matching SIMD vector lanes.
  4. Programmability.

Definition of CGRA

  1. A spatial reconfiguration granularity at fixed functional unit level or above.
  2. A temporal reconfiguration granularity at region/loop-nest level or above.

Alt text
(1) the structure of the CGRA
(2) how it is controlled
(3) how it is integrated with a host processor (if any)
(4) the available tool support.

对于CGRA这个mapping似乎近似是一个Subgraph matching问题?(可能比这个还强)应该是一个NP-hard问题。也就是说目前来说只能寄希望于非确定性算法?
这篇应该就是提出了一个这么一种方法。
问题是因为确实没有机器学习的基础,所以看不太懂RL和MCTS算法部分的表达式。。。
Tsinghua report on CGRA
Zhihu article 1
Zhihu article 2

  • Post title:Reading notes and references on CGRA
  • Post author:Jackcui
  • Create time:2023-08-28 08:28:37
  • Post link:https://jackcuii.github.io/2023/08/28/cgra/
  • Copyright Notice:All articles in this blog are licensed under BY-NC-SA unless stating additionally.
 Comments