首页> 外文期刊>Electronic Colloquium on Computational Complexity >Time-Space Lower Bounds for Two-Pass Learning
【24h】

Time-Space Lower Bounds for Two-Pass Learning

机译:两遍学习的时空下界

获取原文
           

摘要

A line of recent works showed that for a large class of learning problems, any learning algorithm requires either super-linear memory size or a super-polynomial number of samples [Raz16,KRT17,Raz17,MM18,BOGY18,GRT18]. For example, any algorithm for learning parities of size n requires either a memory of size ( n 2 ) or an exponential number of samples [Raz16].All these works modeled the learner as a one-pass branching program, allowing only one pass over the stream of samples. In this work, we prove the first memory-samples lower bounds (with a super-linear lower bound on the memory size and super-polynomial lower bound on the number of samples) when the learner is allowed two passes over the stream of samples. For example, we prove that any two-pass algorithm for learning parities of size n requires either a memory of size ( n 1 5 ) or at least 2 ( n ) samples.More generally, a matrix M : A X ? 1 1 corresponds to the following learning problem: An unknown element x X is chosen uniformly at random. A learner tries to learn x from a stream of samples, ( a 1 b 1 ) ( a 2 b 2 ) , where for every i , a i A is chosen uniformly at random and b i = M ( a i x ) .Assume that k l r are such that any submatrix of M of at least 2 ? k A rows and at least 2 ? l X columns, has a bias of at most 2 ? r . We show that any two-pass learning algorithm for the learning problem corresponding to M requires either a memory of size at least k min k l , or at least 2 ( min k l r ) samples.
机译:一系列最新研究表明,对于一大类学习问题,任何学习算法都需要超线性内存大小或超多项式样本数[Raz16,KRT17,Raz17,MM18,BOGY18,GRT18]。例如,任何用于学习大小为n的奇偶校验的算法都需要一个大小为(n 2)的内存或指数形式的样本[Raz16]。所有这些工作都将学习者建模为一个单程分支程序,只允许一个遍历样本流。在这项工作中,我们证明了当学习者被允许两次通过样本流时,第一个内存样本下界(内存大小具有超线性下界,样本数量具有超多项式下界)。例如,我们证明了用于学习大小为n的奇偶校验的任何两遍算法都需要大小为(n 1 5)的内存或至少2(n)个样本。 1 1对应于以下学习问题:未知元素x X随机均匀地选择。学习者尝试从样本流(a 1 b 1)(a 2 b 2)中学习x,其中对于每个i,ai A是随机随机选择的且bi = M(aix)。假设klr如此M的任何子矩阵至少为2? k行,至少2个? l X列,偏差最多为2? 。我们表明,用于与M对应的学习问题的任何两遍学习算法都需要一个大小至少为k min k l或至少2个(min k l r)样本的内存。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号