Numerous control and learning problems face the situation where sequences of high-dimensional highly dependent data are available, but no or little feedback is provided to the learner. In such situations it may be useful to find a concise representation of the input signal, that would preserve as much as possible of the relevant information. In this work we are interested in the problems where the relevant information is in the time-series dependence. Thus, the problem can be formalized as follows. Given a series of observations X_0, . . ., X_n coming from a large (high-dimensional) space X, find a representation function f mapping X to a finite space Y such that the series f(X_0), . . ., f(X_n) preserve as much information as possible about the original time-series dependence in X_0, . . ., X_n. For stationary time series, the function f can be selected as the one maximizing the time-series information I_∞(f) = h_0(f(X)) ? h_∞(f(X)) where h_0(f(X)) is the Shannon entropy of f(X_0) and h_∞(f(X)) is the entropy rate of the time series f(X_0), . . ., f(X_n), . . . . In this paper we study the functional I_∞(f) from the learning-theoretic point of view. Specifically, we provide some uniform approximation results, and study the behaviour of I_∞(f) in the problem of optimal control.
展开▼