語系:
繁體中文
English
說明(常見問題)
登入
回首頁
切換:
標籤
|
MARC模式
|
ISBD
Customized Computing and Machine Learning.
紀錄類型:
書目-語言資料,手稿 : Monograph/item
正題名/作者:
Customized Computing and Machine Learning./
作者:
Sohrabizadeh, Atefeh.
面頁冊數:
1 online resource (232 pages)
附註:
Source: Dissertations Abstracts International, Volume: 85-12, Section: B.
Contained By:
Dissertations Abstracts International85-12B.
標題:
Computer engineering. -
電子資源:
click for full text (PQDT)
ISBN:
9798382787336
Customized Computing and Machine Learning.
Sohrabizadeh, Atefeh.
Customized Computing and Machine Learning.
- 1 online resource (232 pages)
Source: Dissertations Abstracts International, Volume: 85-12, Section: B.
Thesis (Ph.D.)--University of California, Los Angeles, 2024.
Includes bibliographical references
Nowadays, abundant data across various domains necessitate high-performance computing capabilities. While we used to be able to answer this need by scaling the frequency, the breakdown of Dennard's scaling has rendered this approach obsolete. On the other hand, Domain-specific Accelerators (DSAs) have gained a growing interest since they can offer high performance while being energy efficient. This stems from several factors, such as, 1) they support utilizing special data types and operations, 2) they offer massive parallelism, 3) one can customize the memory access, 4) customizing the control/data path helps with amortizing the overhead of fixed instructions, and 5) one has the option of co-designing the algorithm with the hardware.Unfortunately, despite the huge speedups that DSAs can deliver compared to general-purpose processors, their programmability has not caught up. In the past few decades, High-Level Synthesis (HLS) tools were introduced to raise the abstraction level and free designers from delving into architecture details at the circuit level. While HLS can significantly reduce the efforts involved in the hardware architecture design, not every HLS code yields optimal performance, requiring designers to articulate the most suitable microarchitecture for the target application. This can affect the design turnaround times as there are more choices to explore at a higher level. Moreover, this limitation has confined the DSA community primarily to hardware designers, impeding widespread adoption. This dissertation endeavors to alleviate this problem by combining customized computing and machine learning. Consequently, this dissertation consists of two core parts: 1) customized computing tailored for machine learning applications, and 2) machine learning employed to automate the optimization process of customized computing. Our focus will be on FPGAs as their cost-effective nature and rapid prototyping capabilities make them especially suitable for our research. The large amounts of data available in data centers have motivated researchers to develop machine learning algorithms for processing them. Given that a significant portion of data stored in these centers exists in the form of images or graphs, our attention is directed towards two prominent algorithms designed for such tasks: Convolutional Neural Network (CNN) and Graph Convolutional Network (GCN). In the first part of the dissertation, we develop architecture templates for accelerating these applications. This approach facilitates a reduction in the development cycle, allowing the instantiation of module templates with customizable parameters based on the specific target application.In the second part of the dissertation, we move our focus to general applications and work on automating their optimization steps including design space exploration and performance/area modeling. Therefore, we structure our problem in a way that can be fed into the learning algorithms. We develop a highly efficient bottleneck optimizer to explore the search space. We also explore different learning algorithms including multi-layer perceptron, graph neural networks, attention networks, jumping knowledge networks, etc., aiming to create a performance predictor that is both highly accurate and robust. Our studies show that we can optimize the microarchitecture of general applications quickly using our automated tools. This can open new doors to those without hardware knowledge to try customized computing which in turn helps to broaden the FPGA community and further improve its technology.
Electronic reproduction.
Ann Arbor, Mich. :
ProQuest,
2024
Mode of access: World Wide Web
ISBN: 9798382787336Subjects--Topical Terms:
569006
Computer engineering.
Subjects--Index Terms:
Customized computingIndex Terms--Genre/Form:
554714
Electronic books.
Customized Computing and Machine Learning.
LDR
:04962ntm a22003857 4500
001
1146561
005
20240812064645.5
006
m o d
007
cr bn ---uuuuu
008
250605s2024 xx obm 000 0 eng d
020
$a
9798382787336
035
$a
(MiAaPQ)AAI31301917
035
$a
AAI31301917
040
$a
MiAaPQ
$b
eng
$c
MiAaPQ
$d
NTU
100
1
$a
Sohrabizadeh, Atefeh.
$3
1471968
245
1 0
$a
Customized Computing and Machine Learning.
264
0
$c
2024
300
$a
1 online resource (232 pages)
336
$a
text
$b
txt
$2
rdacontent
337
$a
computer
$b
c
$2
rdamedia
338
$a
online resource
$b
cr
$2
rdacarrier
500
$a
Source: Dissertations Abstracts International, Volume: 85-12, Section: B.
500
$a
Advisor: Cong, Jinsheng Jason.
502
$a
Thesis (Ph.D.)--University of California, Los Angeles, 2024.
504
$a
Includes bibliographical references
520
$a
Nowadays, abundant data across various domains necessitate high-performance computing capabilities. While we used to be able to answer this need by scaling the frequency, the breakdown of Dennard's scaling has rendered this approach obsolete. On the other hand, Domain-specific Accelerators (DSAs) have gained a growing interest since they can offer high performance while being energy efficient. This stems from several factors, such as, 1) they support utilizing special data types and operations, 2) they offer massive parallelism, 3) one can customize the memory access, 4) customizing the control/data path helps with amortizing the overhead of fixed instructions, and 5) one has the option of co-designing the algorithm with the hardware.Unfortunately, despite the huge speedups that DSAs can deliver compared to general-purpose processors, their programmability has not caught up. In the past few decades, High-Level Synthesis (HLS) tools were introduced to raise the abstraction level and free designers from delving into architecture details at the circuit level. While HLS can significantly reduce the efforts involved in the hardware architecture design, not every HLS code yields optimal performance, requiring designers to articulate the most suitable microarchitecture for the target application. This can affect the design turnaround times as there are more choices to explore at a higher level. Moreover, this limitation has confined the DSA community primarily to hardware designers, impeding widespread adoption. This dissertation endeavors to alleviate this problem by combining customized computing and machine learning. Consequently, this dissertation consists of two core parts: 1) customized computing tailored for machine learning applications, and 2) machine learning employed to automate the optimization process of customized computing. Our focus will be on FPGAs as their cost-effective nature and rapid prototyping capabilities make them especially suitable for our research. The large amounts of data available in data centers have motivated researchers to develop machine learning algorithms for processing them. Given that a significant portion of data stored in these centers exists in the form of images or graphs, our attention is directed towards two prominent algorithms designed for such tasks: Convolutional Neural Network (CNN) and Graph Convolutional Network (GCN). In the first part of the dissertation, we develop architecture templates for accelerating these applications. This approach facilitates a reduction in the development cycle, allowing the instantiation of module templates with customizable parameters based on the specific target application.In the second part of the dissertation, we move our focus to general applications and work on automating their optimization steps including design space exploration and performance/area modeling. Therefore, we structure our problem in a way that can be fed into the learning algorithms. We develop a highly efficient bottleneck optimizer to explore the search space. We also explore different learning algorithms including multi-layer perceptron, graph neural networks, attention networks, jumping knowledge networks, etc., aiming to create a performance predictor that is both highly accurate and robust. Our studies show that we can optimize the microarchitecture of general applications quickly using our automated tools. This can open new doors to those without hardware knowledge to try customized computing which in turn helps to broaden the FPGA community and further improve its technology.
533
$a
Electronic reproduction.
$b
Ann Arbor, Mich. :
$c
ProQuest,
$d
2024
538
$a
Mode of access: World Wide Web
650
4
$a
Computer engineering.
$3
569006
650
4
$a
Computer science.
$3
573171
653
$a
Customized computing
653
$a
Electronic Design Automation
653
$a
Graph neural networks
653
$a
High-level synthesis
653
$a
Machine learning
655
7
$a
Electronic books.
$2
local
$3
554714
690
$a
0984
690
$a
0464
710
2
$a
University of California, Los Angeles.
$b
Computer Science 0201.
$3
1182286
710
2
$a
ProQuest Information and Learning Co.
$3
1178819
773
0
$t
Dissertations Abstracts International
$g
85-12B.
856
4 0
$u
http://pqdd.sinica.edu.tw/twdaoapp/servlet/advanced?query=31301917
$z
click for full text (PQDT)
筆 0 讀者評論
多媒體
評論
新增評論
分享你的心得
Export
取書館別
處理中
...
變更密碼[密碼必須為2種組合(英文和數字)及長度為10碼以上]
登入