TY - JOUR
T1 - The complexity of properly learning simple concept classes
AU - Alekhnovich, Misha
AU - Braverman, Mark
AU - Feldman, Vitaly
AU - Klivans, Adam R.
AU - Pitassi, Toniann
N1 - Funding Information:
* Corresponding author. E-mail addresses: [email protected] (M. Alekhnovich), [email protected] (M. Braverman), [email protected] (V. Feldman), [email protected] (A.R. Klivans), [email protected] (T. Pitassi). 1 Supported by CCR grant N CCR-0324906. 2 Supported by an NSERC Postgraduate Scholarship. 3 Supported by NSF grant CCR-98-77049. 4 Part of this work done at Harvard University and supported by an NSF Mathematical Sciences Postdoctoral Research Fellowship. 5 Supported by NSERC and PREA research grants.
PY - 2008/2
Y1 - 2008/2
N2 - We consider the complexity of properly learning concept classes, i.e. when the learner must output a hypothesis of the same form as the unknown concept. We present the following new upper and lower bounds on well-known concept classes: •We show that unless NP = RP, there is no polynomial-time PAC learning algorithm for DNF formulas where the hypothesis is an OR-of-thresholds. Note that as special cases, we show that neither DNF nor OR-of-thresholds are properly learnable unless NP = RP. Previous hardness results have required strong restrictions on the size of the output DNF formula. We also prove that it is NP-hard to learn the intersection of ℓ ≥ 2 halfspaces by the intersection of k halfspaces for any constant k ≥ 0. Previous work held for the case when k = ℓ.•Assuming that NP ⊈ DTIME (2nε{lunate}) for a certain constant ε{lunate} < 1 we show that it is not possible to learn size s decision trees by size sk decision trees for any k ≥ 0. Previous hardness results for learning decision trees held for k ≤ 2.•We present the first non-trivial upper bounds on properly learning DNF formulas. More specifically, we show how to learn size s DNF by DNF in time 2over(O, ̃) (sqrt(n log s)). The hardness results for DNF formulas and intersections of halfspaces are obtained via specialized graph products for amplifying the hardness of approximating the chromatic number as well as applying recent work on the hardness of approximate hypergraph coloring. The hardness results for decision trees, as well as the new upper bounds, are obtained by developing a connection between automatizability in proof complexity and learnability, which may have other applications.
AB - We consider the complexity of properly learning concept classes, i.e. when the learner must output a hypothesis of the same form as the unknown concept. We present the following new upper and lower bounds on well-known concept classes: •We show that unless NP = RP, there is no polynomial-time PAC learning algorithm for DNF formulas where the hypothesis is an OR-of-thresholds. Note that as special cases, we show that neither DNF nor OR-of-thresholds are properly learnable unless NP = RP. Previous hardness results have required strong restrictions on the size of the output DNF formula. We also prove that it is NP-hard to learn the intersection of ℓ ≥ 2 halfspaces by the intersection of k halfspaces for any constant k ≥ 0. Previous work held for the case when k = ℓ.•Assuming that NP ⊈ DTIME (2nε{lunate}) for a certain constant ε{lunate} < 1 we show that it is not possible to learn size s decision trees by size sk decision trees for any k ≥ 0. Previous hardness results for learning decision trees held for k ≤ 2.•We present the first non-trivial upper bounds on properly learning DNF formulas. More specifically, we show how to learn size s DNF by DNF in time 2over(O, ̃) (sqrt(n log s)). The hardness results for DNF formulas and intersections of halfspaces are obtained via specialized graph products for amplifying the hardness of approximating the chromatic number as well as applying recent work on the hardness of approximate hypergraph coloring. The hardness results for decision trees, as well as the new upper bounds, are obtained by developing a connection between automatizability in proof complexity and learnability, which may have other applications.
KW - Automatizability
KW - DNF formulas
KW - Decision trees
KW - Hardness of learning
KW - Intersections of halfspaces
KW - PAC learning
KW - Proof complexity
KW - Proper learning
UR - http://www.scopus.com/inward/record.url?scp=35448997751&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=35448997751&partnerID=8YFLogxK
U2 - 10.1016/j.jcss.2007.04.011
DO - 10.1016/j.jcss.2007.04.011
M3 - Article
AN - SCOPUS:35448997751
SN - 0022-0000
VL - 74
SP - 16
EP - 34
JO - Journal of Computer and System Sciences
JF - Journal of Computer and System Sciences
IS - 1
ER -