首页> 外文会议>Annual Conference on Learning Theory(COLT 2007); 20070613-15; San Diego,CA(US) >Nonlinear Estimators and Tail Bounds for Dimension Reduction in l_1 Using Cauchy Random Projections
【24h】

Nonlinear Estimators and Tail Bounds for Dimension Reduction in l_1 Using Cauchy Random Projections

机译:使用柯西随机投影的l_1中的降维的非线性估计器和尾部边界

获取原文
获取原文并翻译 | 示例

摘要

For dimension reduction in l_1, one can multiply a data matrix A ∈ R~(n×D) by R ∈ R~(D×k) (k << D) whose entries are I.I.d. samples of Cauchy. The impossibility result says one can not recover the pairwise l_1 distances in A from B = AR ∈ R~(n×k), using linear estimators. However, nonlinear estimators are still useful for certain applications in data stream computations, information retrieval, learning, and data mining.We propose three types of nonlinear estimators: the bias-corrected sample median estimator, the bias-corrected geometric mean estimator, and the bias-corrected maximum likelihood estimator. We derive tail bounds for the geometric mean estimator and establish that k = O (log n/∈~2) suffices with the constants explicitly given. Asymptotically (as k → ∞), both the sample median estimator and the geometric mean estimator are about 80% efficient compared to the maximum likelihood estimator (MLE). We analyze the moments of the MLE and propose approximating the distribution of the MLE by an inverse Gaussian.
机译:为了降低l_1中的维数,可以将数据矩阵A∈R〜(n×D)乘以R∈R〜(D×k)(k << D),其条目为I.I.d。柯西的样本。不可能的结果说,使用线性估计器无法从B = AR∈R〜(n×k)恢复A中成对的l_1距离。但是,非线性估计器仍可用于数据流计算,信息检索,学习和数据挖掘中的某些应用。我们提出了三种类型的非线性估计器:偏差校正样本中位数估计器,偏差校正几何均值估计器和偏差校正的最大似然估计。我们导出几何均值估计器的尾边界,并确定k = O(log n /∈〜2)足以满足明确给出的常数。渐近地(如k→∞),与最大似然估计器(MLE)相比,样本中值估计器和几何均值估计器的效率约为80%。我们分析了MLE的矩,并提出了通过反高斯近似MLE的分布。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号