找回密碼
 To register

QQ登錄

只需一步,快速開(kāi)始

掃一掃,訪問(wèn)微社區(qū)

打印 上一主題 下一主題

Titlebook: Data Orchestration in Deep Learning Accelerators; Tushar Krishna,Hyoukjun Kwon,Ananda Samajdar Book 2020 Springer Nature Switzerland AG 20

[復(fù)制鏈接]
查看: 7944|回復(fù): 39
樓主
發(fā)表于 2025-3-21 17:19:43 | 只看該作者 |倒序?yàn)g覽 |閱讀模式
書(shū)目名稱Data Orchestration in Deep Learning Accelerators
編輯Tushar Krishna,Hyoukjun Kwon,Ananda Samajdar
視頻videohttp://file.papertrans.cn/263/262986/262986.mp4
叢書(shū)名稱Synthesis Lectures on Computer Architecture
圖書(shū)封面Titlebook: Data Orchestration in Deep Learning Accelerators;  Tushar Krishna,Hyoukjun Kwon,Ananda Samajdar Book 2020 Springer Nature Switzerland AG 20
描述This Synthesis Lecture focuses on techniques for efficient data orchestration within DNN accelerators. The End of Moore‘s Law, coupled with the increasing growth in deep learning and other AI applications has led to the emergence of custom Deep Neural Network (DNN) accelerators for energy-efficient inference on edge devices. Modern DNNs have millions of hyper parameters and involve billions of computations; this necessitates extensive data movement from memory to on-chip processing engines. It is well known that the cost of data movement today surpasses the cost of the actual computation; therefore, DNN accelerators require careful orchestration of data across on-chip compute, network, and memory elements to minimize the number of accesses to external DRAM. The book covers DNN dataflows, data reuse, buffer hierarchies, networks-on-chip, and automated design-space exploration. It concludes with data orchestration challenges with compressed and sparse DNNs and future trends. The target audience is students, engineers, and researchers interested in designing high-performance and low-energy accelerators for DNN inference.
出版日期Book 2020
版次1
doihttps://doi.org/10.1007/978-3-031-01767-4
isbn_softcover978-3-031-00639-5
isbn_ebook978-3-031-01767-4Series ISSN 1935-3235 Series E-ISSN 1935-3243
issn_series 1935-3235
copyrightSpringer Nature Switzerland AG 2020
The information of publication is updating

書(shū)目名稱Data Orchestration in Deep Learning Accelerators影響因子(影響力)




書(shū)目名稱Data Orchestration in Deep Learning Accelerators影響因子(影響力)學(xué)科排名




書(shū)目名稱Data Orchestration in Deep Learning Accelerators網(wǎng)絡(luò)公開(kāi)度




書(shū)目名稱Data Orchestration in Deep Learning Accelerators網(wǎng)絡(luò)公開(kāi)度學(xué)科排名




書(shū)目名稱Data Orchestration in Deep Learning Accelerators被引頻次




書(shū)目名稱Data Orchestration in Deep Learning Accelerators被引頻次學(xué)科排名




書(shū)目名稱Data Orchestration in Deep Learning Accelerators年度引用




書(shū)目名稱Data Orchestration in Deep Learning Accelerators年度引用學(xué)科排名




書(shū)目名稱Data Orchestration in Deep Learning Accelerators讀者反饋




書(shū)目名稱Data Orchestration in Deep Learning Accelerators讀者反饋學(xué)科排名




單選投票, 共有 0 人參與投票
 

0票 0%

Perfect with Aesthetics

 

0票 0%

Better Implies Difficulty

 

0票 0%

Good and Satisfactory

 

0票 0%

Adverse Performance

 

0票 0%

Disdainful Garbage

您所在的用戶組沒(méi)有投票權(quán)限
沙發(fā)
發(fā)表于 2025-3-21 21:55:53 | 只看該作者
Dataflow and Data Reuse,to billions of computations, we cannot fit all of the computations within an accelerator, which typically has hundreds to thousands of compute units. Therefore, we need to slice the problem into smaller chunks (i.e., computation tiles) and run them in a certain order (i.e., tile scheduling). Within
板凳
發(fā)表于 2025-3-22 01:21:47 | 只看該作者
Buffer Hierarchies,ic accelerators have constraints and goals that differ in key ways. It is important to understand in detail how these cause accelerator architects to make different hardware choices. In this chapter, we present a framework for understanding key options, and explore tradeoffs between design effort an
地板
發(fā)表于 2025-3-22 07:24:38 | 只看該作者
Networks-on-Chip, contain an array of hundreds of PEs. These accelerators aim to achieve high throughput by exploiting massive parallel computations over the PEs while keeping the cost-of-operation much lower than off-the-shelf components with the same compute budget. However, adding more compute elements in an acce
5#
發(fā)表于 2025-3-22 11:50:47 | 只看該作者
6#
發(fā)表于 2025-3-22 13:00:49 | 只看該作者
7#
發(fā)表于 2025-3-22 19:12:22 | 只看該作者
8#
發(fā)表于 2025-3-23 00:52:58 | 只看該作者
Buffer Hierarchies,ic accelerators have constraints and goals that differ in key ways. It is important to understand in detail how these cause accelerator architects to make different hardware choices. In this chapter, we present a framework for understanding key options, and explore tradeoffs between design effort and cross-project reuse.
9#
發(fā)表于 2025-3-23 02:15:42 | 只看該作者
Jason Gu,Rajeeb Dey,Nabanita Adhikaryrovide a brief background on Deep Neural Networks (DNNs), which are the underlying computational mechanisms within Deep Learning applications. Our objective is not to go into the theory behind the structure and accuracy of DNNs (which readers can find in any modern textbook on Machine Learning or De
10#
發(fā)表于 2025-3-23 06:55:08 | 只看該作者
and the Co-production of Men’s Healthto billions of computations, we cannot fit all of the computations within an accelerator, which typically has hundreds to thousands of compute units. Therefore, we need to slice the problem into smaller chunks (i.e., computation tiles) and run them in a certain order (i.e., tile scheduling). Within
 關(guān)于派博傳思  派博傳思旗下網(wǎng)站  友情鏈接
派博傳思介紹 公司地理位置 論文服務(wù)流程 影響因子官網(wǎng) 吾愛(ài)論文網(wǎng) 大講堂 北京大學(xué) Oxford Uni. Harvard Uni.
發(fā)展歷史沿革 期刊點(diǎn)評(píng) 投稿經(jīng)驗(yàn)總結(jié) SCIENCEGARD IMPACTFACTOR 派博系數(shù) 清華大學(xué) Yale Uni. Stanford Uni.
QQ|Archiver|手機(jī)版|小黑屋| 派博傳思國(guó)際 ( 京公網(wǎng)安備110108008328) GMT+8, 2025-10-13 12:02
Copyright © 2001-2015 派博傳思   京公網(wǎng)安備110108008328 版權(quán)所有 All rights reserved
快速回復(fù) 返回頂部 返回列表
崇州市| 保山市| 乌苏市| 灵台县| 望奎县| 蓬溪县| 永丰县| 汨罗市| 巴塘县| 阳朔县| 三原县| 泾阳县| 安义县| 泊头市| 扎兰屯市| 东宁县| 哈巴河县| 邵阳市| 锦州市| 惠水县| 襄樊市| 石家庄市| 剑川县| 基隆市| 邵武市| 横山县| 三河市| 宜都市| 台东县| 乌海市| 广宁县| 平安县| 延津县| 柞水县| 临夏县| 鄂托克旗| 阿拉善左旗| 涞源县| 明光市| 公主岭市| 宣城市|