語系:
繁體中文
English
說明(常見問題)
登入
回首頁
切換:
標籤
|
MARC模式
|
ISBD
Discourse-Level Language Understandi...
~
University of Maryland, College Park.
Discourse-Level Language Understanding with Deep Learning.
紀錄類型:
書目-語言資料,手稿 : Monograph/item
正題名/作者:
Discourse-Level Language Understanding with Deep Learning./
作者:
Iyyer, Mohit.
面頁冊數:
1 online resource (145 pages)
附註:
Source: Dissertation Abstracts International, Volume: 79-02(E), Section: B.
標題:
Computer science. -
電子資源:
click for full text (PQDT)
ISBN:
9780355395303
Discourse-Level Language Understanding with Deep Learning.
Iyyer, Mohit.
Discourse-Level Language Understanding with Deep Learning.
- 1 online resource (145 pages)
Source: Dissertation Abstracts International, Volume: 79-02(E), Section: B.
Thesis (Ph.D.)--University of Maryland, College Park, 2017.
Includes bibliographical references
Designing computational models that can understand language at a human level is a foundational goal in the field of natural language processing (NLP). Given a sentence, machines are capable of translating it into many different languages, generating a corresponding syntactic parse tree, marking words that refer to people or places, and much more. These tasks are solved by statistical machine learning algorithms, which leverage patterns in large datasets to build predictive models. Many recent advances in NLP are due to deep learning models (parameterized as neural networks), which bypass user-specified features in favor of building representations of language directly from the text.
Electronic reproduction.
Ann Arbor, Mich. :
ProQuest,
2018
Mode of access: World Wide Web
ISBN: 9780355395303Subjects--Topical Terms:
573171
Computer science.
Index Terms--Genre/Form:
554714
Electronic books.
Discourse-Level Language Understanding with Deep Learning.
LDR
:03124ntm a2200337K 4500
001
912168
005
20180608102941.5
006
m o u
007
cr mn||||a|a||
008
190606s2017 xx obm 000 0 eng d
020
$a
9780355395303
035
$a
(MiAaPQ)AAI10616153
035
$a
(MiAaPQ)umd:18370
035
$a
AAI10616153
040
$a
MiAaPQ
$b
eng
$c
MiAaPQ
100
1
$a
Iyyer, Mohit.
$3
1184404
245
1 0
$a
Discourse-Level Language Understanding with Deep Learning.
264
0
$c
2017
300
$a
1 online resource (145 pages)
336
$a
text
$b
txt
$2
rdacontent
337
$a
computer
$b
c
$2
rdamedia
338
$a
online resource
$b
cr
$2
rdacarrier
500
$a
Source: Dissertation Abstracts International, Volume: 79-02(E), Section: B.
500
$a
Advisers: Jordan Boyd-Graber; Hal Daume.
502
$a
Thesis (Ph.D.)--University of Maryland, College Park, 2017.
504
$a
Includes bibliographical references
520
$a
Designing computational models that can understand language at a human level is a foundational goal in the field of natural language processing (NLP). Given a sentence, machines are capable of translating it into many different languages, generating a corresponding syntactic parse tree, marking words that refer to people or places, and much more. These tasks are solved by statistical machine learning algorithms, which leverage patterns in large datasets to build predictive models. Many recent advances in NLP are due to deep learning models (parameterized as neural networks), which bypass user-specified features in favor of building representations of language directly from the text.
520
$a
Despite many deep learning-fueled advances at the word and sentence level, however, computers still struggle to understand high-level discourse structure in language, or the way in which authors combine and order different units of text (e.g., sentences, paragraphs, chapters) to express a coherent message or narrative. Part of the reason is data-related, as there are no existing datasets for many contextual language-based problems, and some tasks are too complex to be framed as supervised learning problems; for the latter type, we must either resort to unsupervised learning or devise training objectives that simulate the supervised setting. Another reason is architectural: neural networks designed for sentence-level tasks require additional functionality, interpretability, and efficiency to operate at the discourse level. In this thesis, I design deep learning architectures for three NLP tasks that require integrating information across high-level linguistic context: question answering, fictional relationship understanding, and comic book narrative modeling. While these tasks are very different from each other on the surface, I show that similar neural network modules can be used in each case to form contextual representations.
533
$a
Electronic reproduction.
$b
Ann Arbor, Mich. :
$c
ProQuest,
$d
2018
538
$a
Mode of access: World Wide Web
650
4
$a
Computer science.
$3
573171
650
4
$a
Artificial intelligence.
$3
559380
655
7
$a
Electronic books.
$2
local
$3
554714
690
$a
0984
690
$a
0800
710
2
$a
ProQuest Information and Learning Co.
$3
1178819
710
2
$a
University of Maryland, College Park.
$b
Computer Science.
$3
1180862
856
4 0
$u
http://pqdd.sinica.edu.tw/twdaoapp/servlet/advanced?query=10616153
$z
click for full text (PQDT)
筆 0 讀者評論
多媒體
評論
新增評論
分享你的心得
Export
取書館別
處理中
...
變更密碼[密碼必須為2種組合(英文和數字)及長度為10碼以上]
登入