首页> 外文学位 >Continuous-Time Models of Arrival Times and Optimization Methods for Variable Selection
【24h】

Continuous-Time Models of Arrival Times and Optimization Methods for Variable Selection

机译:到达时间连续时间模型和变量选择的优化方法

获取原文
获取原文并翻译 | 示例

摘要

This thesis naturally divides itself into two sections. The first two chapters concern the development of Bayesian semi-parametric models for arrival times. Chapter 2 considers Bayesian inference for a Gaussian process modulated temporal inhomogeneous Poisson point process, made challenging by an intractable likelihood. The intractable likelihood is circumvented by two novel data augmentation strategies which result in Gaussian measurements of the Gaussian process, connecting the model with a larger literature on modelling time-dependent functions from Bayesian non-parametric regression to time series. A scalable state-space representation of the Matern Gaussian process in 1 dimension is used to provide access to linear time filtering algorithms for performing inference. An MCMC algorithm based on Gibbs sampling with slice-sampling steps is provided and illustrated on simulated and real datasets. The MCMC algorithm exhibits excellent mixing and scalability.;Chapter 3 builds on the previous model to detect specific signals in temporal point patterns arising in neuroscience. The firing of a neuron over time in response to an external stimulus generates a temporal point pattern or "spike train". Of special interest is how neurons encode information from dual simultaneous external stimuli. Among many hypotheses is the presence multiplexing - interleaving periods of firing as it would for each individual stimulus in isolation. Statistical models are developed to quantify evidence for a variety of experimental hypotheses. Each experimental hypothesis translates to a particular form of intensity function for the dual stimuli trials. The dual stimuli intensity is modelled as a dynamic superposition of single stimulus intensities, defined by a time-dependent weight function that is modelled non-parametrically as a transformed Gaussian process. Experiments on simulated data demonstrate that the model is able to learn the weight function very well, but other model parameters which have meaningful physical interpretations less well.;Chapters 4 and 5 concern mathematical optimization and theoretical properties of Bayesian models for variable selection. Such optimizations are challenging due to non-convexity, non-smoothness and discontinuity of the objective. Chapter 4 presents advances in continuous optimization algorithms based on relating mathematical and statistical approaches defined in connection with several iterative algorithms for penalized linear regression. I demonstrate the equivalence of parameter mappings using EM under several data augmentation strategies - location-mixture representations, orthogonal data augmentation and LQ design matrix decompositions. I show that these model-based approaches are equivalent to algorithmic derivation via proximal gradient methods. This provides new perspectives on model-based and algorithmic approaches, connects across several research themes in optimization and statistics, and provides access, beyond EM, to relevant theory from the proximal gradient and convex analysis literatures.;Chapter 5 presents a modern and technologically up-to-date approach to discrete optimization for variable selection models through their formulation as mixed integer programming models. Mixed integer quadratic and quadratically constrained programs are developed for the point-mass-Laplace and g-prior. Combined with warm-starts and optimality-based bounds tightening procedures provided by the heuristics of the previous chapter, the MIQP model developed for the point-mass-Laplace prior converges to global optimality in a matter of seconds for moderately sized real datasets. The obtained estimator is demonstrated to possess superior predictive performance over that obtained by cross-validated lasso in a number of real datasets. The MIQCP model for the g-prior struggles to match the performance of the former and highlights the fact that the performance of the mixed integer solver depends critically on the ability of the prior to rapidly concentrate posterior mass on good models.
机译:本论文自然将自己分为两个部分。前两章介绍了到达时间的贝叶斯半参数模型的发展。第2章将贝叶斯推论考虑为高斯过程调制的时间不均匀泊松点过程,该过程由于难以解决的可能性而具有挑战性。两种新颖的数据增强策略规避了这种棘手的可能性,这些策略导致对高斯过程进行高斯测量,将模型与有关从贝叶斯非参数回归到时间序列的时间相关函数建模的大量文献相联系。一维Matern高斯过程的可伸缩状态空间表示用于提供对线性时间滤波算法的访问以执行推理。提供了基于Gibbs采样和切片采样步骤的MCMC算法,并在模拟和真实数据集上进行了说明。 MCMC算法具有出色的混合和可伸缩性。;第3章在先前模型的基础上,检测神经科学中出现的时间点模式中的特定信号。响应于外部刺激,神经元随着时间的推移而发射会产生时间点模式或“峰值训练”。特别令人感兴趣的是神经元如何编码来自双重同时外部刺激的信息。在许多假设中是存在多路复用-触发的交错周期,就像隔离每个单独刺激时一样。开发统计模型以量化各种实验假设的证据。每个实验假设都转化为双重刺激试验的一种特定形式的强度函数。双重刺激强度被建模为单个刺激强度的动态叠加,由时间相关的权重函数定义,该函数随时间变化的权重被非参数地建模为变换的高斯过程。仿真数据实验表明,该模型能够很好地学习权重函数,而其他具有有意义的物理解释的模型参数则不太好。第4章和第5章涉及贝叶斯模型用于变量选择的数学优化和理论性质。由于物镜的非凸性,非光滑性和不连续性,因此此类优化具有挑战性。第4章介绍了基于相关数学和统计方法的连续优化算法的进展,这些方法是结合几种用于惩罚线性回归的迭代算法定义的。我演示了在几种数据增强策略下使用EM进行参数映射的等效方法-位置混合表示,正交数据增强和LQ设计矩阵分解。我证明了这些基于模型的方法等效于通过近端梯度法的算法推导。这为基于模型和算法的方法提供了新的观点,跨越了优化和统计方面的多个研究主题,并提供了除EM之外的近端梯度和凸分析文献中的相关理论的访问权限。第5章介绍了现代技术。通过将变量选择模型表示为混合整数规划模型,对变量选择模型进行了最新的离散优化方法。针对点质量-拉普拉斯和g-prior开发了混合整数二次和二次约束程​​序。结合上一章的启发式方法提供的热启动和基于最优性的边界加紧程序,针对中等质量的实际数据集,为点质量-拉普拉斯先验开发的MIQP模型在几秒钟内收敛到全局最优性。事实证明,与许多实际数据集中通过交叉验证的套索所获得的估计值相比,所获得的估计值具有更高的预测性能。用于g优先级的MIQCP模型难以与前者的性能相匹配,并强调了一个事实,即混合整数求解器的性能关键取决于先验质量在快速集中后验质量的能力。

著录项

  • 作者

    Lindon, Michael.;

  • 作者单位

    Duke University.;

  • 授予单位 Duke University.;
  • 学科 Statistics.
  • 学位 Ph.D.
  • 年度 2018
  • 页码 168 p.
  • 总页数 168
  • 原文格式 PDF
  • 正文语种 eng
  • 中图分类
  • 关键词

  • 入库时间 2022-08-17 11:37:07

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号