語系:
繁體中文
English
說明(常見問題)
登入
回首頁
切換:
標籤
|
MARC模式
|
ISBD
Spatial-Temporal Hierarchical Model ...
~
Nie, Xiaohan.
Spatial-Temporal Hierarchical Model for Joint Learning and Inference of Human Action and Pose.
紀錄類型:
書目-語言資料,手稿 : Monograph/item
正題名/作者:
Spatial-Temporal Hierarchical Model for Joint Learning and Inference of Human Action and Pose./
作者:
Nie, Xiaohan.
面頁冊數:
1 online resource (119 pages)
附註:
Source: Dissertation Abstracts International, Volume: 78-10(E), Section: B.
Contained By:
Dissertation Abstracts International78-10B(E).
標題:
Statistics. -
電子資源:
click for full text (PQDT)
ISBN:
9781369846980
Spatial-Temporal Hierarchical Model for Joint Learning and Inference of Human Action and Pose.
Nie, Xiaohan.
Spatial-Temporal Hierarchical Model for Joint Learning and Inference of Human Action and Pose.
- 1 online resource (119 pages)
Source: Dissertation Abstracts International, Volume: 78-10(E), Section: B.
Thesis (Ph.D.)--University of California, Los Angeles, 2017.
Includes bibliographical references
In the community of computer vision, human pose estimation and human action recognition are two classic and also of particular important tasks. They always serve as basic preprocessing steps for other high-level tasks such as group activity analysis, visual search and human identification and they are also widely used as key components in many real applications such as intelligent surveillance system and human-computer interaction based system. The two tasks are closely related for understanding human motion, most methods, however, learn separate models and combine them sequentially.
Electronic reproduction.
Ann Arbor, Mich. :
ProQuest,
2018
Mode of access: World Wide Web
ISBN: 9781369846980Subjects--Topical Terms:
556824
Statistics.
Index Terms--Genre/Form:
554714
Electronic books.
Spatial-Temporal Hierarchical Model for Joint Learning and Inference of Human Action and Pose.
LDR
:04581ntm a2200409Ki 4500
001
918799
005
20181106104111.5
006
m o u
007
cr mn||||a|a||
008
190606s2017 xx obm 000 0 eng d
020
$a
9781369846980
035
$a
(MiAaPQ)AAI10286374
035
$a
(MiAaPQ)ucla:15475
035
$a
AAI10286374
040
$a
MiAaPQ
$b
eng
$c
MiAaPQ
$d
NTU
100
1
$a
Nie, Xiaohan.
$3
1193221
245
1 0
$a
Spatial-Temporal Hierarchical Model for Joint Learning and Inference of Human Action and Pose.
264
0
$c
2017
300
$a
1 online resource (119 pages)
336
$a
text
$b
txt
$2
rdacontent
337
$a
computer
$b
c
$2
rdamedia
338
$a
online resource
$b
cr
$2
rdacarrier
500
$a
Source: Dissertation Abstracts International, Volume: 78-10(E), Section: B.
500
$a
Adviser: Song-Chun Zhu.
502
$a
Thesis (Ph.D.)--University of California, Los Angeles, 2017.
504
$a
Includes bibliographical references
520
$a
In the community of computer vision, human pose estimation and human action recognition are two classic and also of particular important tasks. They always serve as basic preprocessing steps for other high-level tasks such as group activity analysis, visual search and human identification and they are also widely used as key components in many real applications such as intelligent surveillance system and human-computer interaction based system. The two tasks are closely related for understanding human motion, most methods, however, learn separate models and combine them sequentially.
520
$a
In this dissertation, we build systems for pursuing a unified framework to integrate training and inference of human pose estimation and action recognition in a spatial-temporal And-Or Graph (ST-AOG) representation. Particularly, we study different ways to achieve this goal:
520
$a
(1) A two-level And-Or Tree structure is utilized for representing action as animated pose template (APT). Each action is a sequence of moving pose templates with transition probabilities. Each Pose template consists of a shape template represented by an And-node capturing part appearance, and a motion template represented by an Or-node capturing part motions. The transitions between moving pose templates are governed in a Hidden Markov Model. The part locations, pose types and action labels are estimated together in inference.
520
$a
(2) In order to tackle actions from unknown and unseen views we present a multi-view spatial-temporal And-Or Graph (MST-AOG) for cross-view action recognition. As a compositional model, the MST-AOG compactly represents the hierarchical combinatorial structures of cross-view actions by explicitly modeling the geometry, appearance and motion variations. The model training takes advantage of the 3D human skeleton data obtained from Kinect cameras to avoid annotating video frames. The efficient inference enables action recognition from novel views. A new Multi-view Action3D dataset has been created and released.
520
$a
(3) To further represent part, pose and action jointly and improve performance, we represent action at three scales by a ST-AOG model. Each action is decomposed into poses which are further divided into mid-level spatial-temporal parts (ST-parts) and then parts. The hierarchical model structure captures the geometric and appearance variations of pose at each frame. The lateral connections between ST-parts at adjacent frames capture the action-specific motions. The model parameters at three scales are learned discriminatively and dynamic programming is utilized for efficient inference. The experiments demonstrate the large benet of joint modeling of the two tasks.
520
$a
(4) The last but not the least, we study a novel framework for full-body 3D human pose estimation which is a essential task for human attention recognition, robot-based human action prediction and interaction. We build a two-level hierarchy of Long Short-Term Memory (LSTM) network with tree-structure to predict the depth on 2D human joints and then reconstruct the 3D pose. Our two-level model utilizes two cues for depth prediction: 1) the global features from 2D skeleton. 2) the local features from image patches of body parts.
533
$a
Electronic reproduction.
$b
Ann Arbor, Mich. :
$c
ProQuest,
$d
2018
538
$a
Mode of access: World Wide Web
650
4
$a
Statistics.
$3
556824
650
4
$a
Computer science.
$3
573171
650
4
$a
Artificial intelligence.
$3
559380
655
7
$a
Electronic books.
$2
local
$3
554714
690
$a
0463
690
$a
0984
690
$a
0800
710
2
$a
ProQuest Information and Learning Co.
$3
1178819
710
2
$a
University of California, Los Angeles.
$b
Statistics 0891.
$3
1183048
773
0
$t
Dissertation Abstracts International
$g
78-10B(E).
856
4 0
$u
http://pqdd.sinica.edu.tw/twdaoapp/servlet/advanced?query=10286374
$z
click for full text (PQDT)
筆 0 讀者評論
多媒體
評論
新增評論
分享你的心得
Export
取書館別
處理中
...
變更密碼[密碼必須為2種組合(英文和數字)及長度為10碼以上]
登入