語系:
繁體中文
English
說明(常見問題)
登入
回首頁
切換:
標籤
|
MARC模式
|
ISBD
Adaptive Preference Learning with Ba...
~
ProQuest Information and Learning Co.
Adaptive Preference Learning with Bandit Feedback : = Information Filtering, Dueling Bandits and Incentivizing Exploration.
紀錄類型:
書目-語言資料,手稿 : Monograph/item
正題名/作者:
Adaptive Preference Learning with Bandit Feedback :/
其他題名:
Information Filtering, Dueling Bandits and Incentivizing Exploration.
作者:
Chen, Bangrui.
面頁冊數:
1 online resource (141 pages)
附註:
Source: Dissertation Abstracts International, Volume: 79-03(E), Section: B.
Contained By:
Dissertation Abstracts International79-03B(E).
標題:
Operations research. -
電子資源:
click for full text (PQDT)
ISBN:
9780355528138
Adaptive Preference Learning with Bandit Feedback : = Information Filtering, Dueling Bandits and Incentivizing Exploration.
Chen, Bangrui.
Adaptive Preference Learning with Bandit Feedback :
Information Filtering, Dueling Bandits and Incentivizing Exploration. - 1 online resource (141 pages)
Source: Dissertation Abstracts International, Volume: 79-03(E), Section: B.
Thesis (Ph.D.)--Cornell University, 2017.
Includes bibliographical references
In this thesis, we study adaptive preference learning, in which a machine learning system learns users' preferences from feedback while simultaneously using these learned preferences to help them find preferred items. We study three different types of user feedback in three application setting: cardinal feedback with application in information filtering systems, ordinal feedback with application in personalized content recommender systems, and attribute feedback with application in review aggregators. We connect these settings respectively to existing work on classical multi-armed bandits, dueling bandits, and incentivizing exploration. For each type of feedback and application setting, we provide an algorithm and a theoretical analysis bounding its regret. We demonstrate through numerical experiments that our algorithms outperform existing benchmarks.
Electronic reproduction.
Ann Arbor, Mich. :
ProQuest,
2018
Mode of access: World Wide Web
ISBN: 9780355528138Subjects--Topical Terms:
573517
Operations research.
Index Terms--Genre/Form:
554714
Electronic books.
Adaptive Preference Learning with Bandit Feedback : = Information Filtering, Dueling Bandits and Incentivizing Exploration.
LDR
:02184ntm a2200361Ki 4500
001
916834
005
20180928111502.5
006
m o u
007
cr mn||||a|a||
008
190606s2017 xx obm 000 0 eng d
020
$a
9780355528138
035
$a
(MiAaPQ)AAI10680541
035
$a
(MiAaPQ)cornellgrad:10605
035
$a
AAI10680541
040
$a
MiAaPQ
$b
eng
$c
MiAaPQ
$d
NTU
100
1
$a
Chen, Bangrui.
$3
1190684
245
1 0
$a
Adaptive Preference Learning with Bandit Feedback :
$b
Information Filtering, Dueling Bandits and Incentivizing Exploration.
264
0
$c
2017
300
$a
1 online resource (141 pages)
336
$a
text
$b
txt
$2
rdacontent
337
$a
computer
$b
c
$2
rdamedia
338
$a
online resource
$b
cr
$2
rdacarrier
500
$a
Source: Dissertation Abstracts International, Volume: 79-03(E), Section: B.
500
$a
Adviser: Peter I. Frazier.
502
$a
Thesis (Ph.D.)--Cornell University, 2017.
504
$a
Includes bibliographical references
520
$a
In this thesis, we study adaptive preference learning, in which a machine learning system learns users' preferences from feedback while simultaneously using these learned preferences to help them find preferred items. We study three different types of user feedback in three application setting: cardinal feedback with application in information filtering systems, ordinal feedback with application in personalized content recommender systems, and attribute feedback with application in review aggregators. We connect these settings respectively to existing work on classical multi-armed bandits, dueling bandits, and incentivizing exploration. For each type of feedback and application setting, we provide an algorithm and a theoretical analysis bounding its regret. We demonstrate through numerical experiments that our algorithms outperform existing benchmarks.
533
$a
Electronic reproduction.
$b
Ann Arbor, Mich. :
$c
ProQuest,
$d
2018
538
$a
Mode of access: World Wide Web
650
4
$a
Operations research.
$3
573517
650
4
$a
Computer science.
$3
573171
650
4
$a
Statistics.
$3
556824
650
4
$a
Artificial intelligence.
$3
559380
655
7
$a
Electronic books.
$2
local
$3
554714
690
$a
0796
690
$a
0984
690
$a
0463
690
$a
0800
710
2
$a
ProQuest Information and Learning Co.
$3
1178819
710
2
$a
Cornell University.
$b
Operations Research.
$3
1180996
773
0
$t
Dissertation Abstracts International
$g
79-03B(E).
856
4 0
$u
http://pqdd.sinica.edu.tw/twdaoapp/servlet/advanced?query=10680541
$z
click for full text (PQDT)
筆 0 讀者評論
多媒體
評論
新增評論
分享你的心得
Export
取書館別
處理中
...
變更密碼[密碼必須為2種組合(英文和數字)及長度為10碼以上]
登入