語系:
繁體中文
English
說明(常見問題)
登入
回首頁
切換:
標籤
|
MARC模式
|
ISBD
Adaptive Distributed Caching for Sca...
~
Carnegie Mellon University.
Adaptive Distributed Caching for Scalable Machine Learning Services.
紀錄類型:
書目-語言資料,手稿 : Monograph/item
正題名/作者:
Adaptive Distributed Caching for Scalable Machine Learning Services./
作者:
Drolia, Utsav.
面頁冊數:
1 online resource (103 pages)
附註:
Source: Dissertation Abstracts International, Volume: 79-01(E), Section: B.
Contained By:
Dissertation Abstracts International79-01B(E).
標題:
Computer engineering. -
電子資源:
click for full text (PQDT)
ISBN:
9780355198430
Adaptive Distributed Caching for Scalable Machine Learning Services.
Drolia, Utsav.
Adaptive Distributed Caching for Scalable Machine Learning Services.
- 1 online resource (103 pages)
Source: Dissertation Abstracts International, Volume: 79-01(E), Section: B.
Thesis (Ph.D.)
Includes bibliographical references
Applications for Internet-enabled devices use machine learning to process captured data to make intelligent decisions or provide information to users. Typically, the computation to process the data is executed in cloud-based backends. The devices are used for sensing data, offloading it to the cloud, receiving responses and acting upon them. However, this approach leads to high end-to-end latency due to communication over the Internet.
Electronic reproduction.
Ann Arbor, Mich. :
ProQuest,
2018
Mode of access: World Wide Web
ISBN: 9780355198430Subjects--Topical Terms:
569006
Computer engineering.
Index Terms--Genre/Form:
554714
Electronic books.
Adaptive Distributed Caching for Scalable Machine Learning Services.
LDR
:03618ntm a2200385Ki 4500
001
910829
005
20180517112611.5
006
m o u
007
cr mn||||a|a||
008
190606s2017 xx obm 000 0 eng d
020
$a
9780355198430
035
$a
(MiAaPQ)AAI10621640
035
$a
(MiAaPQ)cmu:10146
035
$a
AAI10621640
040
$a
MiAaPQ
$b
eng
$c
MiAaPQ
099
$a
TUL
$f
hyy
$c
available through World Wide Web
100
1
$a
Drolia, Utsav.
$3
1182306
245
1 0
$a
Adaptive Distributed Caching for Scalable Machine Learning Services.
264
0
$c
2017
300
$a
1 online resource (103 pages)
336
$a
text
$b
txt
$2
rdacontent
337
$a
computer
$b
c
$2
rdamedia
338
$a
online resource
$b
cr
$2
rdacarrier
500
$a
Source: Dissertation Abstracts International, Volume: 79-01(E), Section: B.
500
$a
Adviser: Priya Narasimhan.
502
$a
Thesis (Ph.D.)
$c
Carnegie Mellon University
$d
2017.
504
$a
Includes bibliographical references
520
$a
Applications for Internet-enabled devices use machine learning to process captured data to make intelligent decisions or provide information to users. Typically, the computation to process the data is executed in cloud-based backends. The devices are used for sensing data, offloading it to the cloud, receiving responses and acting upon them. However, this approach leads to high end-to-end latency due to communication over the Internet.
520
$a
This dissertation proposes reducing this response time by minimizing offloading, and pushing computation close to the source of the data, i.e. to edge servers and devices themselves. To adapt to the resource constrained environment at the edge, it presents an approach that leverages spatiotemporal locality to push subparts of the model to the edge. This approach is embodied in a distributed caching framework, Cachier.
520
$a
Cachier is built upon a novel caching model for recognition, and is distributed across edge servers and devices. The analytical caching model for recognition provides a formulation for expected latency for recognition requests in Cachier. The formulation incorporates the effects of compute time and accuracy. It also incorporates network conditions, thus providing a method to compute expected response times under various conditions. This is utilized as a cost function by Cachier, at edge servers and devices. By analyzing requests at the edge server, Cachier caches relevant parts of the trained model at edge servers, which is used to respond to requests, minimizing the number of requests that go to the cloud. Then, Cachier uses context-aware prediction to prefetch parts of the trained model onto devices. The requests can then be processed on the devices, thus minimizing the number of offloaded requests. Finally, Cachier enables cooperation between nearby devices to allow exchanging prefetched data, reducing the dependence on remote servers even further.
520
$a
The efficacy of Cachier is evaluated by using it with an art recognition application. The application is driven using real world traces gathered at museums. By conducting a large-scale study with different control variables, we show that Cachier can lower latency, increase scalability and decrease infrastructure resource usage, while maintaining high accuracy.
533
$a
Electronic reproduction.
$b
Ann Arbor, Mich. :
$c
ProQuest,
$d
2018
538
$a
Mode of access: World Wide Web
650
4
$a
Computer engineering.
$3
569006
650
4
$a
Computer science.
$3
573171
655
7
$a
Electronic books.
$2
local
$3
554714
690
$a
0464
690
$a
0984
710
2
$a
ProQuest Information and Learning Co.
$3
1178819
710
2
$a
Carnegie Mellon University.
$b
Electrical and Computer Engineering.
$3
1182305
773
0
$t
Dissertation Abstracts International
$g
79-01B(E).
856
4 0
$u
http://pqdd.sinica.edu.tw/twdaoapp/servlet/advanced?query=10621640
$z
click for full text (PQDT)
筆 0 讀者評論
多媒體
評論
新增評論
分享你的心得
Export
取書館別
處理中
...
變更密碼[密碼必須為2種組合(英文和數字)及長度為10碼以上]
登入