...
首页> 外文期刊>Journal of proteome research >Machine Learning Strategy That Leverages Large Data sets to Boost Statistical Power in Small-Scale Experiments
【24h】

Machine Learning Strategy That Leverages Large Data sets to Boost Statistical Power in Small-Scale Experiments

机译:机器学习策略利用大型数据集来提高小规模实验中的统计功率

获取原文
获取原文并翻译 | 示例
           

摘要

y Machine learning methods have proven invaluable for increasing the sensitivity of peptide detection in proteomics experiments. Most modern tools, such as Percolator and PeptideProphet, use semisupervised algorithms to learn models directly from the data sets that they analyze. Although these methods are effective for many proteomics experiments, we suspected that they may be suboptimal for experiments of smaller scale. In this work, we found that the power and consistency of Percolator results were reduced as the size of the experiment was decreased. As an alternative, we propose a different operating mode for Percolator: learn a model with Percolator from a large data set and use the learned model to evaluate the small-scale experiment. We call this a "static modeling" approach, in contrast to Percolator's usual "dynamic model" that is trained anew for each data set. We applied this static modeling approach to two settings: small, gel-based experiments and single-cell proteomics. In both cases, static models increased the yield of detected peptides and eliminated the model-induced variability of the standard dynamic approach. These results suggest that static models are a powerful tool for bringing the full benefits of Percolator and other semisupervised algorithms to small-scale experiments.
机译:y机床学习方法已经证明,用于提高蛋白质组学实验中肽检测的敏感性的无价值。大多数现代工具,如渗滤器和肽前容,使用半体验算法直接从他们分析的数据集中学习模型。虽然这些方法对许多蛋白质组学实验有效,但我们怀疑它们可能是较小规模的实验的次优。在这项工作中,我们发现,随着实验的大小降低,渗透结果的功率和一致性降低。作为替代方案,我们提出了一种不同的运算模式来实现渗滤器:从大数据集中使用渗滤器学习模型,并使用学习模型来评估小规模实验。我们称之为“静态建模”方法,与培训每次数据集培训的培训的常规“动态模型”相反。我们将这种静态建模方法应用于两个设置:小型,凝胶的实验和单细胞蛋白质组学。在这两种情况下,静态模型增加了检测到的肽的产量,并消除了标准动态方法的模型诱导的可变性。这些结果表明,静态模型是一种强大的工具,用于将渗滤器和其他半体验算法带到小规模实验的强大工具。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号