Professional Documents
Culture Documents
Colt Tutorial PDF
Colt Tutorial PDF
Vasant Honavar
Artificial Intelligence Research Laboratory
Department of Computer Science
honavar@cs.iastate.edu
Iowa State University, Ames, Iowa 50011
http://www.cs.iastate.edu/~honavar/aigroup.html
What are learning systems?
Algorithm
Initialize L={X1, ~X1, .... ~XN}
Predict according to match between an instance
and the conjunction of literals in L
Whenever a mistake is made on a positive
example, drop the offending literals from L
Eg: <0111, 1> will result in L = {~ X1, X2, X3, X4}
<1110, 1> will yield L = {X2, X3}
Mistake bound model
Proof of Theorem 1:
No literal in C is ever eliminated from L
Each mistake eliminates at least one literal from L
The first mistake eliminates N of the 2N literals
Conjunctive concepts can be learned with at most
(N+1) mistakes
Conclusion: Conjunctive concepts are easy to learn
in the mistake bound model
Optimal Mistake Bound Learning Algorithms
Examples Learner
Oracle
Concept
Samples
Instance Distribution
Probably Approximately Correct Learning
Consider:
An instance space X
A concept space C = { C: X {0,1}}
A hypothesis space H = { h : X {0 , 1 }}
An unknown, arbitrary, not necessarily
computable, stationary probability distribution D
over the instance space X
PAC Learning
Proof: omitted.
Occam algorithm is PAC for K-decision lists
1 1 1
m = O lg + d lg
Corollary: Acyclic, layered multi-layer networks of
s threshold logic units, each with r inputs, has VC
dimension 2 ( r + 1) s lg( es)
Using a Weak learner for PAC Learning
Question:
Can we relax the requirement of learning under
all probability distributions over the instance
space (including extremely pathological
distributions) by limiting the class of distributions
to a useful subset of all possible distributions?
What are the implications of doing so on the
learnability of concept classes that are PAC-hard?
What probability distributions are natural?
Learning from Simple Examples
1 Honavar, V. http://www.cs.iastate.edu/~honavar/cs673.s96.html
2 Kearns, M.J. & Vazirani, U.V. An Introduction to Computational
Learning Theory. Cambridge, MA: MIT Press. 1994.
3 Langley, P. Elements of Machine Learning. Palo Alto, CA: Morgan
Kaufmann. 1995.
4 Li, M. & Vitanyi, P. Kolmogorov Complexity and its Applications.
New York: Springer-Verlag. 1997.
5 Mitchell, T. Machine Learning. New York: McGraw Hill. 1997.
6 Natarajan, B.K. Machine Learning: A Theoretical Approach. Palo
Alto, CA: Morgan Kaufmann, 1992.