語系:
繁體中文
English
說明(常見問題)
登入
回首頁
切換:
標籤
|
MARC模式
|
ISBD
Context Based Multi-Image Visual Que...
~
ProQuest Information and Learning Co.
Context Based Multi-Image Visual Question Answering (VQA) in Deep Learning.
紀錄類型:
書目-語言資料,手稿 : Monograph/item
正題名/作者:
Context Based Multi-Image Visual Question Answering (VQA) in Deep Learning./
作者:
Peddinti, Sudhakar Reddy.
面頁冊數:
1 online resource (53 pages)
附註:
Source: Masters Abstracts International, Volume: 57-04.
標題:
Computer science. -
電子資源:
click for full text (PQDT)
ISBN:
9780355615746
Context Based Multi-Image Visual Question Answering (VQA) in Deep Learning.
Peddinti, Sudhakar Reddy.
Context Based Multi-Image Visual Question Answering (VQA) in Deep Learning.
- 1 online resource (53 pages)
Source: Masters Abstracts International, Volume: 57-04.
Thesis (M.S.)--University of Missouri - Kansas City, 2018.
Includes bibliographical references
Image question answering has gained huge popularity in recent years due to advancements in Deep Learning technologies and computer processing hardware which are able to achieve higher accuracies with faster processing capabilities. Processing image details over natural language information is one of the most challenging tasks in Artificial Intelligence. Most recently, there has been tremendous interest in both creating datasets and proposing deep neural network models for addressing the problem of learning both the images and text information through a question-answering task called Visual Question Answering (VQA). VQA gets us a level closer in terms of human computer interaction through AI. However, VQA is limited in terms of capturing attention only to a certain extent in image (attributes) instead of understanding the semantics of the context in images.
Electronic reproduction.
Ann Arbor, Mich. :
ProQuest,
2018
Mode of access: World Wide Web
ISBN: 9780355615746Subjects--Topical Terms:
573171
Computer science.
Index Terms--Genre/Form:
554714
Electronic books.
Context Based Multi-Image Visual Question Answering (VQA) in Deep Learning.
LDR
:03034ntm a2200349K 4500
001
912196
005
20180608102941.5
006
m o u
007
cr mn||||a|a||
008
190606s2018 xx obm 000 0 eng d
020
$a
9780355615746
035
$a
(MiAaPQ)AAI10743252
035
$a
(MiAaPQ)umkc:11229
035
$a
AAI10743252
040
$a
MiAaPQ
$b
eng
$c
MiAaPQ
100
1
$a
Peddinti, Sudhakar Reddy.
$3
1184446
245
1 0
$a
Context Based Multi-Image Visual Question Answering (VQA) in Deep Learning.
264
0
$c
2018
300
$a
1 online resource (53 pages)
336
$a
text
$b
txt
$2
rdacontent
337
$a
computer
$b
c
$2
rdamedia
338
$a
online resource
$b
cr
$2
rdacarrier
500
$a
Source: Masters Abstracts International, Volume: 57-04.
500
$a
Adviser: Yugyung Lee.
502
$a
Thesis (M.S.)--University of Missouri - Kansas City, 2018.
504
$a
Includes bibliographical references
520
$a
Image question answering has gained huge popularity in recent years due to advancements in Deep Learning technologies and computer processing hardware which are able to achieve higher accuracies with faster processing capabilities. Processing image details over natural language information is one of the most challenging tasks in Artificial Intelligence. Most recently, there has been tremendous interest in both creating datasets and proposing deep neural network models for addressing the problem of learning both the images and text information through a question-answering task called Visual Question Answering (VQA). VQA gets us a level closer in terms of human computer interaction through AI. However, VQA is limited in terms of capturing attention only to a certain extent in image (attributes) instead of understanding the semantics of the context in images.
520
$a
In this thesis, we propose a semantic framework known as Context VQA (CVQA) that aims to extend the existing VQA models in two aspects. First, we built a contextual model for defining the semantics of similar contexts from a multi-image set instead of a single image. In the CVQA framework, a two-stage model was proposed (1) to identify one or more images by mapping the semantic sense of the question to the contextual model built from similar contexts of the images; (2) for the select images, provide the appropriate answer for a given question based on the proposed contextual model. Second, CVQA is an enhancement of one of the VQA implementations (VGG-16), which is extended with a more complex model like ResNet-152, and we analyzed the performance of our CVQA framework on 3 datasets---DAQUAR, VQA version1, and VQA version2. From our experiments, we gained improvement in accuracy and runtime. We also present a CVQA application for context-based visual question answering.
533
$a
Electronic reproduction.
$b
Ann Arbor, Mich. :
$c
ProQuest,
$d
2018
538
$a
Mode of access: World Wide Web
650
4
$a
Computer science.
$3
573171
650
4
$a
Artificial intelligence.
$3
559380
650
4
$a
Information science.
$3
561178
655
7
$a
Electronic books.
$2
local
$3
554714
690
$a
0984
690
$a
0800
690
$a
0723
710
2
$a
ProQuest Information and Learning Co.
$3
1178819
710
2
$a
University of Missouri - Kansas City.
$b
Computer Science.
$3
1182463
856
4 0
$u
http://pqdd.sinica.edu.tw/twdaoapp/servlet/advanced?query=10743252
$z
click for full text (PQDT)
筆 0 讀者評論
多媒體
評論
新增評論
分享你的心得
Export
取書館別
處理中
...
變更密碼[密碼必須為2種組合(英文和數字)及長度為10碼以上]
登入