TY - GEN
T1 - Memory-sample lower bounds for learning parity with noise
AU - Garg, Sumegha
AU - Kothari, Pravesh K.
AU - Liu, Pengda
AU - Raz, Ran
N1 - Publisher Copyright:
© Sumegha Garg, Pravesh K. Kothari, Pengda Liu, and Ran Raz; licensed under Creative Commons License CC-BY 4.0
PY - 2021/9/1
Y1 - 2021/9/1
N2 - In this work, we show, for the well-studied problem of learning parity under noise, where a learner tries to learn x = (x1,..., xn) ∈ {0, 1}n from a stream of random linear equations over F2 that are correct with probability 12 + ε and flipped with probability 12 − ε (0 < ε < 12 ), that any learning algorithm requires either a memory of size Ω(n2/ε) or an exponential number of samples. In fact, we study memory-sample lower bounds for a large class of learning problems, as characterized by [8], when the samples are noisy. A matrix M : A × X → {−1, 1} corresponds to the following learning problem with error parameter ε: an unknown element x ∈ X is chosen uniformly at random. A learner tries to learn x from a stream of samples, (a1, b1), (a2, b2)..., where for every i, ai ∈ A is chosen uniformly at random and bi = M(ai, x) with probability 1/2 + ε and bi = −M(ai, x) with probability 1/2 − ε (0 < ε < 12 ). Assume that k, ℓ, r are such that any submatrix of M of at least 2−k · |A| rows and at least 2−ℓ · |X| columns, has a bias of at most 2−r. We show that any learning algorithm for the learning problem corresponding to M, with error parameter ε, requires either a memory of size at least Ω (kε·ℓ), or at least 2Ω(r) samples. The result holds even if the learner has an exponentially small success probability (of 2−Ω(r)). In particular, this shows that for a large class of learning problems, same as those in [8], any learning algorithm requires either a memory of size at least Ω ( (log |X|)·(log |A|) ) or an exponential number of noisy samples. ε Our proof is based on adapting the arguments in [21, 8] to the noisy case.
AB - In this work, we show, for the well-studied problem of learning parity under noise, where a learner tries to learn x = (x1,..., xn) ∈ {0, 1}n from a stream of random linear equations over F2 that are correct with probability 12 + ε and flipped with probability 12 − ε (0 < ε < 12 ), that any learning algorithm requires either a memory of size Ω(n2/ε) or an exponential number of samples. In fact, we study memory-sample lower bounds for a large class of learning problems, as characterized by [8], when the samples are noisy. A matrix M : A × X → {−1, 1} corresponds to the following learning problem with error parameter ε: an unknown element x ∈ X is chosen uniformly at random. A learner tries to learn x from a stream of samples, (a1, b1), (a2, b2)..., where for every i, ai ∈ A is chosen uniformly at random and bi = M(ai, x) with probability 1/2 + ε and bi = −M(ai, x) with probability 1/2 − ε (0 < ε < 12 ). Assume that k, ℓ, r are such that any submatrix of M of at least 2−k · |A| rows and at least 2−ℓ · |X| columns, has a bias of at most 2−r. We show that any learning algorithm for the learning problem corresponding to M, with error parameter ε, requires either a memory of size at least Ω (kε·ℓ), or at least 2Ω(r) samples. The result holds even if the learner has an exponentially small success probability (of 2−Ω(r)). In particular, this shows that for a large class of learning problems, same as those in [8], any learning algorithm requires either a memory of size at least Ω ( (log |X|)·(log |A|) ) or an exponential number of noisy samples. ε Our proof is based on adapting the arguments in [21, 8] to the noisy case.
KW - Branching program
KW - Learning parity under noise
KW - Memory-sample tradeoffs
KW - Space lower bound
UR - http://www.scopus.com/inward/record.url?scp=85115620646&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85115620646&partnerID=8YFLogxK
U2 - 10.4230/LIPIcs-APPROX/RANDOM.2021.60
DO - 10.4230/LIPIcs-APPROX/RANDOM.2021.60
M3 - Conference contribution
AN - SCOPUS:85115620646
T3 - Leibniz International Proceedings in Informatics, LIPIcs
BT - Approximation, Randomization, and Combinatorial Optimization. Algorithms and Techniques, APPROX/RANDOM 2021
A2 - Wootters, Mary
A2 - Sanita, Laura
PB - Schloss Dagstuhl- Leibniz-Zentrum fur Informatik GmbH, Dagstuhl Publishing
T2 - 24th International Conference on Approximation Algorithms for Combinatorial Optimization Problems, APPROX 2021 and 25th International Conference on Randomization and Computation, RANDOM 2021
Y2 - 16 August 2021 through 18 August 2021
ER -