Computer Science/Discrete Mathematics Seminar I

A time-space lower bound for a large class of learning problems

We prove a general time-space lower bound that applies for a large class of learning problems and shows that for every problem in that class, any learning algorithm requires either a memory of quadratic size or an exponential number of samples. As a special case, this gives a new proof for the time-space lower bound for parity learning [R16].Our result is stated in terms of the norm of the matrix that corresponds to the learning problem. Let $X$, $A$ be two finite sets. Let $M: A \times X \rightarrow \{-1,1\}$ be a matrix. The matrix $M$ corresponds to the following learning problem: An unknown element $x \in X$ was chosen uniformly at random. A learner tries to learn $x$ from a stream of samples, $(a_1, b_1), (a_2, b_2)\ldots$, where for every $i$, $a_i \in A$ is chosen uniformly at random and $b_i = M(a_i,x)$.Let $\sigma$ be the largest singular value of $M$ and note that always $\sigma \leq |A|^{1/2} \cdot |X|^{1/2}$. We show that if $\sigma \leq |A|^{1/2} \cdot |X|^{1/2 - \epsilon}$, then any learning algorithm for the corresponding learning problem requires either a memory of size quadratic in $\epsilon n$ or number of samples exponential in $\epsilon n$, where $n = \log_2 |X|$.

Date & Time

April 03, 2017 | 11:15am – 12:15pm

Affiliation

Princeton University