TY - GEN
T1 - A time-space lower bound for a large class of learning problems
AU - Raz, Ran
N1 - Publisher Copyright:
© 2017 IEEE.
PY - 2017/11/10
Y1 - 2017/11/10
N2 - We prove a general time-space lower bound that applies for a large class of learning problems and shows that for every problem in that class, any learning algorithm requires either a memory of quadratic size or an exponential number of samples. As a special case, this gives a new proof for the time-space lower bound for parity learning [R16]. Our result is stated in terms of the norm of the matrix that corresponds to the learning problem. Let X, A be two finite sets. Let M: A X \rightarrow \{-1,1\} be a matrix. The matrix M corresponds to the following learning problem: An unknown element x X was chosen uniformly at random. A learner tries to learn x from a stream of samples, (a-1, b-1), (a-2, b-2)..., where for every i, a-i A is chosen uniformly at random and b-i = M(a-i,x). Let \sigma be the largest singular value of M and note that always \sigma ≤ |A|^{1/2} |X|^{1/2}. We show that if \sigma ≤ |A|^{1/2} |X|^{1/2 - ∈, then any learning algorithm for the corresponding learning problem requires either a memory of size quadratic in ∈ n or number of samples exponential in ∈ n, where n = \log-2 |X|.As a special case, this gives a new proof for the memorysamples lower bound for parity learning [14].
AB - We prove a general time-space lower bound that applies for a large class of learning problems and shows that for every problem in that class, any learning algorithm requires either a memory of quadratic size or an exponential number of samples. As a special case, this gives a new proof for the time-space lower bound for parity learning [R16]. Our result is stated in terms of the norm of the matrix that corresponds to the learning problem. Let X, A be two finite sets. Let M: A X \rightarrow \{-1,1\} be a matrix. The matrix M corresponds to the following learning problem: An unknown element x X was chosen uniformly at random. A learner tries to learn x from a stream of samples, (a-1, b-1), (a-2, b-2)..., where for every i, a-i A is chosen uniformly at random and b-i = M(a-i,x). Let \sigma be the largest singular value of M and note that always \sigma ≤ |A|^{1/2} |X|^{1/2}. We show that if \sigma ≤ |A|^{1/2} |X|^{1/2 - ∈, then any learning algorithm for the corresponding learning problem requires either a memory of size quadratic in ∈ n or number of samples exponential in ∈ n, where n = \log-2 |X|.As a special case, this gives a new proof for the memorysamples lower bound for parity learning [14].
UR - http://www.scopus.com/inward/record.url?scp=85041097497&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85041097497&partnerID=8YFLogxK
U2 - 10.1109/FOCS.2017.73
DO - 10.1109/FOCS.2017.73
M3 - Conference contribution
AN - SCOPUS:85041097497
T3 - Annual Symposium on Foundations of Computer Science - Proceedings
SP - 732
EP - 742
BT - Proceedings - 58th Annual IEEE Symposium on Foundations of Computer Science, FOCS 2017
PB - IEEE Computer Society
T2 - 58th Annual IEEE Symposium on Foundations of Computer Science, FOCS 2017
Y2 - 15 October 2017 through 17 October 2017
ER -