語系:
繁體中文
English
說明(常見問題)
登入
回首頁
切換:
標籤
|
MARC模式
|
ISBD
Low-rank graphical models and Bayesi...
~
Columbia University.
Low-rank graphical models and Bayesian inference in the statistical analysis of noisy neural data.
紀錄類型:
書目-語言資料,手稿 : Monograph/item
正題名/作者:
Low-rank graphical models and Bayesian inference in the statistical analysis of noisy neural data./
作者:
Smith, Carl.
面頁冊數:
1 online resource (136 pages)
附註:
Source: Dissertation Abstracts International, Volume: 75-02(E), Section: B.
標題:
Statistics. -
電子資源:
click for full text (PQDT)
ISBN:
9781303469688
Low-rank graphical models and Bayesian inference in the statistical analysis of noisy neural data.
Smith, Carl.
Low-rank graphical models and Bayesian inference in the statistical analysis of noisy neural data.
- 1 online resource (136 pages)
Source: Dissertation Abstracts International, Volume: 75-02(E), Section: B.
Thesis (Ph.D.)--Columbia University, 2013.
Includes bibliographical references
We develop new methods of Bayesian inference, largely in the context of analysis of neuroscience data. The work is broken into several parts. In the first part, we introduce a novel class of joint probability distributions in which exact inference is tractable. Previously it has been difficult to find general constructions for models in which efficient exact inference is possible, outside of certain classical cases. We identify a class of such models that are tractable owing to a certain "low-rank" structure in the potentials that couple neighboring variables. In the second part we develop methods to quantify and measure information loss in analysis of neuronal spike train data due to two types of noise, making use of the ideas developed in the first part. Information about neuronal identity or temporal resolution may be lost during spike detection and sorting, or precision of spike times may be corrupted by various effects. We quantify the information lost due to these effects for the relatively simple but sufficiently broad class of Markovian model neurons. We find that decoders that model the probability distribution of spike-neuron assignments significantly outperform decoders that use only the most likely spike assignments. We also apply the ideas of the low-rank models from the first section to defining a class of prior distributions over the space of stimuli (or other covariate) which, by conjugacy, preserve the tractability of inference. In the third part, we treat Bayesian methods for the estimation of sparse signals, with application to the locating of synapses in a dendritic tree. We develop a compartmentalized model of the dendritic tree. Building on previous work that applied and generalized ideas of least angle regression to obtain a fast Bayesian solution to the resulting estimation problem, we describe two other approaches to the same problem, one employing a horseshoe prior and the other using various spike-and-slab priors. In the last part, we revisit the low-rank models of the first section and apply them to the problem of inferring orientation selectivity maps from noisy observations of orientation preference. The relevant low-rank model exploits the self-conjugacy of the von Mises distribution on the circle. Because the orientation map model is loopy, we cannot do exact inference on the low-rank model by the forward backward algorithm, but block-wise Gibbs sampling by the forward backward algorithm speeds mixing. We explore another von Mises coupling potential Gibbs sampler that proves to effectively smooth noisily observed orientation maps.
Electronic reproduction.
Ann Arbor, Mich. :
ProQuest,
2018
Mode of access: World Wide Web
ISBN: 9781303469688Subjects--Topical Terms:
556824
Statistics.
Index Terms--Genre/Form:
554714
Electronic books.
Low-rank graphical models and Bayesian inference in the statistical analysis of noisy neural data.
LDR
:03784ntm a2200325K 4500
001
913718
005
20180622095236.5
006
m o u
007
cr mn||||a|a||
008
190606s2013 xx obm 000 0 eng d
020
$a
9781303469688
035
$a
(MiAaPQ)AAI3598394
035
$a
(MiAaPQ)columbia:11623
035
$a
AAI3598394
040
$a
MiAaPQ
$b
eng
$c
MiAaPQ
100
1
$a
Smith, Carl.
$3
1058886
245
1 0
$a
Low-rank graphical models and Bayesian inference in the statistical analysis of noisy neural data.
264
0
$c
2013
300
$a
1 online resource (136 pages)
336
$a
text
$b
txt
$2
rdacontent
337
$a
computer
$b
c
$2
rdamedia
338
$a
online resource
$b
cr
$2
rdacarrier
500
$a
Source: Dissertation Abstracts International, Volume: 75-02(E), Section: B.
500
$a
Adviser: Liam Paninski.
502
$a
Thesis (Ph.D.)--Columbia University, 2013.
504
$a
Includes bibliographical references
520
$a
We develop new methods of Bayesian inference, largely in the context of analysis of neuroscience data. The work is broken into several parts. In the first part, we introduce a novel class of joint probability distributions in which exact inference is tractable. Previously it has been difficult to find general constructions for models in which efficient exact inference is possible, outside of certain classical cases. We identify a class of such models that are tractable owing to a certain "low-rank" structure in the potentials that couple neighboring variables. In the second part we develop methods to quantify and measure information loss in analysis of neuronal spike train data due to two types of noise, making use of the ideas developed in the first part. Information about neuronal identity or temporal resolution may be lost during spike detection and sorting, or precision of spike times may be corrupted by various effects. We quantify the information lost due to these effects for the relatively simple but sufficiently broad class of Markovian model neurons. We find that decoders that model the probability distribution of spike-neuron assignments significantly outperform decoders that use only the most likely spike assignments. We also apply the ideas of the low-rank models from the first section to defining a class of prior distributions over the space of stimuli (or other covariate) which, by conjugacy, preserve the tractability of inference. In the third part, we treat Bayesian methods for the estimation of sparse signals, with application to the locating of synapses in a dendritic tree. We develop a compartmentalized model of the dendritic tree. Building on previous work that applied and generalized ideas of least angle regression to obtain a fast Bayesian solution to the resulting estimation problem, we describe two other approaches to the same problem, one employing a horseshoe prior and the other using various spike-and-slab priors. In the last part, we revisit the low-rank models of the first section and apply them to the problem of inferring orientation selectivity maps from noisy observations of orientation preference. The relevant low-rank model exploits the self-conjugacy of the von Mises distribution on the circle. Because the orientation map model is loopy, we cannot do exact inference on the low-rank model by the forward backward algorithm, but block-wise Gibbs sampling by the forward backward algorithm speeds mixing. We explore another von Mises coupling potential Gibbs sampler that proves to effectively smooth noisily observed orientation maps.
533
$a
Electronic reproduction.
$b
Ann Arbor, Mich. :
$c
ProQuest,
$d
2018
538
$a
Mode of access: World Wide Web
650
4
$a
Statistics.
$3
556824
650
4
$a
Neurosciences.
$3
593561
655
7
$a
Electronic books.
$2
local
$3
554714
690
$a
0463
690
$a
0317
710
2
$a
ProQuest Information and Learning Co.
$3
1178819
710
2
$a
Columbia University.
$b
Chemical Physics.
$3
1186670
856
4 0
$u
http://pqdd.sinica.edu.tw/twdaoapp/servlet/advanced?query=3598394
$z
click for full text (PQDT)
筆 0 讀者評論
多媒體
評論
新增評論
分享你的心得
Export
取書館別
處理中
...
變更密碼[密碼必須為2種組合(英文和數字)及長度為10碼以上]
登入