語系:
繁體中文
English
說明(常見問題)
登入
回首頁
切換:
標籤
|
MARC模式
|
ISBD
Towards Understanding the Role of Knowledge in Improving Transformer-Based Language Models.
紀錄類型:
書目-語言資料,手稿 : Monograph/item
正題名/作者:
Towards Understanding the Role of Knowledge in Improving Transformer-Based Language Models./
作者:
Pal, Kuntal Kumar.
面頁冊數:
1 online resource (283 pages)
附註:
Source: Dissertations Abstracts International, Volume: 85-02, Section: B.
Contained By:
Dissertations Abstracts International85-02B.
標題:
Computer science. -
電子資源:
click for full text (PQDT)
ISBN:
9798380109390
Towards Understanding the Role of Knowledge in Improving Transformer-Based Language Models.
Pal, Kuntal Kumar.
Towards Understanding the Role of Knowledge in Improving Transformer-Based Language Models.
- 1 online resource (283 pages)
Source: Dissertations Abstracts International, Volume: 85-02, Section: B.
Thesis (Ph.D.)--Arizona State University, 2023.
Includes bibliographical references
In natural language processing, language models have achieved remarkable success over the last few years. The Transformers are at the core of most of these models. Their success can be mainly attributed to an enormous amount of curated data they are trained on. Even though such language models are trained on massive curated data, they often need specific extracted knowledge to understand better and reason. This is because often relevant knowledge may be implicit or missing, which hampers machine reasoning. Apart from that, manual knowledge curation is time-consuming and erroneous. Hence, finding fast and effective methods to extract such knowledge from data is important for improving language models. This leads to finding ideal ways to utilize such knowledge by incorporating them into language models. Successful knowledge extraction and integration lead to an important question of knowledge evaluation of such models by developing tools or introducing challenging test suites to learn about their limitations and improve them further. So to improve the transformer-based models, understanding the role of knowledge becomes important.In the pursuit to improve language models with knowledge, in this dissertation I study three broad research directions spanning across the natural language, biomedical and cybersecurity domains: (1) Knowledge Extraction (KX) - How can transformer-based language models be leveraged to extract knowledge from data? (2) Knowledge Integration (KI) - How can such specific knowledge be used to improve such models? (3) Knowledge Evaluation (KE) - How can language models be evaluated for specific skills and understand their limitations?I propose methods to extract explicit textual, implicit structural, missing textual, and missing structural knowledge from natural language and binary programs using transformer-based language models. I develop ways to improve the language model's multi-step and commonsense reasoning abilities using external knowledge. Finally, I develop challenging datasets which assess their numerical reasoning skills in both in-domain and out-of-domain settings.
Electronic reproduction.
Ann Arbor, Mich. :
ProQuest,
2024
Mode of access: World Wide Web
ISBN: 9798380109390Subjects--Topical Terms:
573171
Computer science.
Subjects--Index Terms:
EvaluationIndex Terms--Genre/Form:
554714
Electronic books.
Towards Understanding the Role of Knowledge in Improving Transformer-Based Language Models.
LDR
:03506ntm a22003977 4500
001
1143620
005
20240517104614.5
006
m o d
007
cr mn ---uuuuu
008
250605s2023 xx obm 000 0 eng d
020
$a
9798380109390
035
$a
(MiAaPQ)AAI30531226
035
$a
AAI30531226
040
$a
MiAaPQ
$b
eng
$c
MiAaPQ
$d
NTU
100
1
$a
Pal, Kuntal Kumar.
$3
1468360
245
1 0
$a
Towards Understanding the Role of Knowledge in Improving Transformer-Based Language Models.
264
0
$c
2023
300
$a
1 online resource (283 pages)
336
$a
text
$b
txt
$2
rdacontent
337
$a
computer
$b
c
$2
rdamedia
338
$a
online resource
$b
cr
$2
rdacarrier
500
$a
Source: Dissertations Abstracts International, Volume: 85-02, Section: B.
500
$a
Advisor: Baral, Chitta.
502
$a
Thesis (Ph.D.)--Arizona State University, 2023.
504
$a
Includes bibliographical references
520
$a
In natural language processing, language models have achieved remarkable success over the last few years. The Transformers are at the core of most of these models. Their success can be mainly attributed to an enormous amount of curated data they are trained on. Even though such language models are trained on massive curated data, they often need specific extracted knowledge to understand better and reason. This is because often relevant knowledge may be implicit or missing, which hampers machine reasoning. Apart from that, manual knowledge curation is time-consuming and erroneous. Hence, finding fast and effective methods to extract such knowledge from data is important for improving language models. This leads to finding ideal ways to utilize such knowledge by incorporating them into language models. Successful knowledge extraction and integration lead to an important question of knowledge evaluation of such models by developing tools or introducing challenging test suites to learn about their limitations and improve them further. So to improve the transformer-based models, understanding the role of knowledge becomes important.In the pursuit to improve language models with knowledge, in this dissertation I study three broad research directions spanning across the natural language, biomedical and cybersecurity domains: (1) Knowledge Extraction (KX) - How can transformer-based language models be leveraged to extract knowledge from data? (2) Knowledge Integration (KI) - How can such specific knowledge be used to improve such models? (3) Knowledge Evaluation (KE) - How can language models be evaluated for specific skills and understand their limitations?I propose methods to extract explicit textual, implicit structural, missing textual, and missing structural knowledge from natural language and binary programs using transformer-based language models. I develop ways to improve the language model's multi-step and commonsense reasoning abilities using external knowledge. Finally, I develop challenging datasets which assess their numerical reasoning skills in both in-domain and out-of-domain settings.
533
$a
Electronic reproduction.
$b
Ann Arbor, Mich. :
$c
ProQuest,
$d
2024
538
$a
Mode of access: World Wide Web
650
4
$a
Computer science.
$3
573171
650
4
$a
Linguistics.
$3
557829
653
$a
Evaluation
653
$a
Knowledge extraction
653
$a
Language model
653
$a
Reasoning
653
$a
Transformers
655
7
$a
Electronic books.
$2
local
$3
554714
690
$a
0984
690
$a
0290
690
$a
0800
710
2
$a
ProQuest Information and Learning Co.
$3
1178819
710
2
$a
Arizona State University.
$b
Computer Science.
$3
845377
773
0
$t
Dissertations Abstracts International
$g
85-02B.
856
4 0
$u
http://pqdd.sinica.edu.tw/twdaoapp/servlet/advanced?query=30531226
$z
click for full text (PQDT)
筆 0 讀者評論
多媒體
評論
新增評論
分享你的心得
Export
取書館別
處理中
...
變更密碼[密碼必須為2種組合(英文和數字)及長度為10碼以上]
登入
第一次登入時,112年前入學、到職者,密碼請使用身分證號登入;112年後入學、到職者,密碼請使用身分證號"後六碼"登入,請注意帳號密碼有區分大小寫!
帳號(學號)
密碼
請在此電腦上記得個人資料
取消
忘記密碼? (請注意!您必須已在系統登記E-mail信箱方能使用。)