
428 L Learning Automata
Corollary 15 Let C be the class of functions that can be
expressed in the following way: Let p
i;j
: ˙ ! K be arbi-
trary functions of a single variable (1 i `, 1 j n).
Let ` = O(log n) and g
i
: ˙
n
! K (1 i `)bedefined
by ˙
n
j=1
p
i;j
(z
j
). Finally, let f : ˙
n
! K be defined by
f =
Q
`
i=1
g
i
.Then,C is learnable in time poly(n; j˙j).
Corollary 16 Consider the class of decision trees of depth s,
where the query at each node v is a boolean function f
v
with r
max
t (as defined in Section "Key Results") such
that (t +1)
s
=poly(n).Then,thisclassislearnableintime
poly(n; j˙j).
The above class contains, for example,all the decision trees
of depth O(log n) that contain in each node a term or
aXORofasubsetofvariablesasdefinedin[15] (in this
case r
max
2).
Negative Results
In [4] some limitation of the learnability via the automa-
ton representation has been proved. One can show that the
main algorithm does not efficiently learn several impor-
tant classes of functions. More precisely, these classes con-
tain functions f that have no “small” automaton, i. e., by
Theorem 1, the corresponding Hankel matrix F is “large”
over every field
K.
Theorem 17 The following classes are not learnable in
time polynomial in n and the formula size using multiplic-
ity automata (over any field
K): DNF, Monotone DNF,
2-DNF, Read-once DNF, k-term DNF, for k = !(log n),
Satisfy-s DNF, for s = !(1), Read-j satisfy-s DNF, for
j = !(1) and s = ˝(log n).
Some of these classes are known to be learnable by
other methods, some are natural generalizations of
classes known to be learnable as automata (O(log n)-term
DNF [11,12,14], and satisfy-s DNF for s = O(1) (Corol-
lary 11)) or by other methods (read-j satisfy-s for js =
O(log n/loglogn)[10]), and the learnability of some of
the others is still an open problem.
Cross References
Learning Constant-Depth Circuits
Learning DNF Formulas
Recommended Reading
1. Angluin, D.: Learning regular sets from queries and counterex-
amples. Inf. Comput. 75, 87–106 (1987)
2. Angluin, D.: Queries and concept learning. Mach. Learn. 2(4),
319–342 (1988)
3. Beimel, A., Bergadano, F., Bshouty, N.H., Kushilevitz, E., Varric-
chio, S.: On the applications of multiplicity automata in learn-
ing. In: Proc. of the 37th Annu. IEEE Symp. on Foundations of
Computer Science, pp. 349–358, IEEE Comput. Soc. Press, Los
Alamitos (1996)
4. Beimel,A.,Bergadano,F.,Bshouty,N.H.,Kushilevitz,E.,Varric-
chio, S.: Learning Functions Represented as Multiplicity Au-
tomata. J. ACM 47, 506–530 (2000)
5. Beimel, A., Kushilevitz, E.: Learning boxes in high dimension.
In: Ben-David S. (ed.) 3rd European Conf. on Computational
Learning Theory (EuroCOLT ’97), Lecture Notes in Artificial In-
telligence, vol. 1208, pp. 3–15. Springer, Berlin (1997) Journal
version: Algorithmica 22, 76–90 (1998)
6. Bergadano, F., Catalano, D., Varricchio, S.: Learning sat-k-DNF
formulas from membership queries. In: Proc. of the 28th Annu.
ACM Symp. on the Theory of Computing, pp. 126–130. ACM
Press, New York (1996)
7. Bergadano, F., Varricchio, S.: Learning behaviors of automata
from multiplicity and equivalence queries. In: Proc. of 2nd
Italian Conf. on Algorithms and Complexity. Lecture Notes in
Computer Science, vol. 778, pp. 54–62. Springer, Berlin (1994).
Journalversion:SIAMJ.Comput.25(6), 1268–1280 (1996)
8. Bergadano, F., Varricchio, S.: Learning behaviors of automata
from shortest counterexamples. In: EuroCOLT ’95, Lecture
Notes in Artificial Intelligence, vol. 904, pp. 380–391. Springer,
Berlin (1996)
9. Bisht,L.,Bshouty,N.H.,Mazzawi,H.:OnOptimalLearningAlgo-
rithms for Multiplicity Automata. In: Proc. of 19th Annu. ACM
Conf. Comput. Learning Theory, Lecture Notes in Computer
Science. vol. 4005, pp. 184–198. Springer, Berlin (2006)
10. Blum,A.,Khardon,R.,Kushilevitz,E.,Pitt,L.,Roth,D.:Onlearn-
ing read-k-satisfy-j DNF. In: Proc. of 7th Annu. ACM Conf. on
Comput. Learning Theory, pp. 110–117. ACM Press, New York
(1994)
11. Bshouty, N.H.: Exact learning via the monotone theory. In: Proc.
of the 34th Annu. IEEE Symp. on Foundations of Computer
Science, pp. 302–311. IEEE Comput. Soc. Press, Los Alami-
tos (1993). Journal version: Inform. Comput. 123(1), 146–153
(1995)
12. Bshouty, N.H.: Simple learning algorithms using divide and
conquer. In: Proc. of 8th Annu. ACM Conf. on Comput. Learn-
ing Theory, pp. 447–453. ACM Press, New York (1995). Journal
version: Computational Complexity, 6 , 174–194 (1997)
13. Bshouty, N.H., Tamon, C., Wilson, D.K.: Learning Matrix Func-
tions over Rings. Algorithmica 22(1/2), 91–111 (1998)
14. Kushilevitz, E.: A simple algorithm for learning O(log n)-term
DNF. In: Proc. of 9th Annu. ACM Conf. on Comput. Learning
Theory, pp 266–269, ACM Press, New York (1996). Journal ver-
sion: Inform. Process. Lett. 61(6), 289–292 (1997)
15. Kushilevitz, E., Mansour, Y.: Learning decision trees using the
Fourier spectrum. SIAM J. Comput. 22(6), 1331–1348 (1993)
16. Melideo, G., Varricchio, S.: Learning unary output two-tape au-
tomata from multiplicity and equivalence queries. In: ALT ’98.
Lecture Notes in Computer Science, vol. 1501, pp. 87–102.
Springer, Berlin (1998)
17. Ohnishi, H., Seki, H., Kasami, T.: A polynomial time learning al-
gorithm for recognizable series. IEICE Transactions on Informa-
tion and Systems, E77-D(10)(5), 1077–1085 (1994)
18. Schapire, R.E., Sellie, L.M.: Learning sparse multivariate polyno-
mials over a field with queries and counterexamples. J. Com-
put. Syst. Sci. 52(2), 201–213 (1996)
19. Valiant, L.G.: A theory of the learnable. Commun. ACM 27(11),
1134–1142 (1984)