Computational learning theory: Difference between revisions

Jump to navigation Jump to search
No edit summary
Line 61: Line 61:
== External links ==
== External links ==
* [http://www.learningtheory.org Computational learning theory web site]
* [http://www.learningtheory.org Computational learning theory web site]
* [http://www.inference.phy.cam.ac.uk/mackay/itila/ On-line book: Information Theory, Inference, and Learning Algorithms], by [[David MacKay (scientist)]], gives a detailed account of the Bayesian approach to machine learning.
* [http://www.inference.phy.cam.ac.uk/mackay/itila/ On-line book: Information Theory, Inference, and Learning Algorithms], by David MacKay (scientist), gives a detailed account of the Bayesian approach to machine learning.
* [http://www.santafe.edu/~shalizi/reviews/kearns-vazirani/ Review of ''An Introduction to Computational Learning Theory'']
* [http://www.santafe.edu/~shalizi/reviews/kearns-vazirani/ Review of ''An Introduction to Computational Learning Theory'']
* [http://www.santafe.edu/~shalizi/reviews/vapnik-nature/ Review of ''The Nature of Statistical Learning Theory'']
* [http://www.santafe.edu/~shalizi/reviews/vapnik-nature/ Review of ''The Nature of Statistical Learning Theory'']

Revision as of 18:41, 10 June 2009

WikiDoc Resources for Computational learning theory

Articles

Most recent articles on Computational learning theory

Most cited articles on Computational learning theory

Review articles on Computational learning theory

Articles on Computational learning theory in N Eng J Med, Lancet, BMJ

Media

Powerpoint slides on Computational learning theory

Images of Computational learning theory

Photos of Computational learning theory

Podcasts & MP3s on Computational learning theory

Videos on Computational learning theory

Evidence Based Medicine

Cochrane Collaboration on Computational learning theory

Bandolier on Computational learning theory

TRIP on Computational learning theory

Clinical Trials

Ongoing Trials on Computational learning theory at Clinical Trials.gov

Trial results on Computational learning theory

Clinical Trials on Computational learning theory at Google

Guidelines / Policies / Govt

US National Guidelines Clearinghouse on Computational learning theory

NICE Guidance on Computational learning theory

NHS PRODIGY Guidance

FDA on Computational learning theory

CDC on Computational learning theory

Books

Books on Computational learning theory

News

Computational learning theory in the news

Be alerted to news on Computational learning theory

News trends on Computational learning theory

Commentary

Blogs on Computational learning theory

Definitions

Definitions of Computational learning theory

Patient Resources / Community

Patient resources on Computational learning theory

Discussion groups on Computational learning theory

Patient Handouts on Computational learning theory

Directions to Hospitals Treating Computational learning theory

Risk calculators and risk factors for Computational learning theory

Healthcare Provider Resources

Symptoms of Computational learning theory

Causes & Risk Factors for Computational learning theory

Diagnostic studies for Computational learning theory

Treatment of Computational learning theory

Continuing Medical Education (CME)

CME Programs on Computational learning theory

International

Computational learning theory en Espanol

Computational learning theory en Francais

Business

Computational learning theory in the Marketplace

Patents on Computational learning theory

Experimental / Informatics

List of terms related to Computational learning theory

Please Take Over This Page and Apply to be Editor-In-Chief for this topic: There can be one or more than one Editor-In-Chief. You may also apply to be an Associate Editor-In-Chief of one of the subtopics below. Please mail us [1] to indicate your interest in serving either as an Editor-In-Chief of the entire topic or as an Associate Editor-In-Chief for a subtopic. Please be sure to attach your CV and or biographical sketch.

In theoretical computer science, computational learning theory or computational learning problem is a mathematical field related to the analysis of machine learning algorithms. It is traditionally referred to as the grammatical inference problem.

Machine learning algorithms take a training set, form hypotheses or models, and make predictions about the future. Because the training set is finite and the future is uncertain, learning theory usually does not yield absolute guarantees of performance of the algorithms. Instead, probabilistic bounds on the performance of machine learning algorithms are quite common.

In addition to performance bounds, computational learning theorists study the time complexity and feasibility of learning. In computational learning theory, a computation is considered feasible if it can be done in polynomial time. There are two kinds of time complexity results:

  1. Positive results --- Showing that a certain class of functions is learnable in polynomial time.
  2. Negative results - Showing that certain classes cannot be learned in polynomial time.

Negative results are proven only by assumption. The assumptions that are common in negative results are:

  • Computational complexity - P ≠ NP
  • Cryptographic - One-way functions exist.

There are several different approaches to computational learning theory, which are often mathematically incompatible. This incompatibility arises from using different inference principles: principles which tell you how to generalize from limited data. The incompatibility also arises from differing definitions of probability (see frequency probability, Bayesian probability). The different approaches include:

Computational learning theory has led to practical algorithms. For example, PAC theory inspired boosting, VC theory led to support vector machines, and Bayesian inference led to belief networks (by Judea Pearl).

See also:

References

Surveys

  • Angluin, D. 1992. Computational learning theory: Survey and selected bibliography. In Proceedings of the Twenty-Fourth Annual ACM Symposium on Theory of Computing (May 1992), pp. 351--369. http://portal.acm.org/citation.cfm?id=129712.129746
  • D. Haussler. Probably approximately correct learning. In AAAI-90 Proceedings of the Eight National Conference on Artificial Intelligence, Boston, MA, pages 1101--1108. American Association for Artificial Intelligence, 1990. http://citeseer.ist.psu.edu/haussler90probably.html

VC dimension

  • V. Vapnik and A. Chervonenkis. On the uniform convergence of relative frequencies of events to their probabilities. Theory of Probability and its Applications, 16(2):264--280, 1971.

Feature selection

Inductive inference

  • E. M. Gold. Language identification in the limit. Information and Control, 10:447--474, 1967.

Optimal O notation learning

Negative results

Boosting

Occam's Razor

  • Blumer, A.; Ehrenfeucht, A.; Haussler, D.; Warmuth, M. K. "Occam's razor" Inf.Proc.Lett. 24, 377-380, 1987.
  • A. Blumer, A. Ehrenfeucht, D. Haussler, and M. K. Warmuth. Learnability and the Vapnik-Chervonenkis dimension. Journal of the ACM, 36(4):929--865, 1989.

Probably approximately correct learning

  • L. Valiant. A Theory of the Learnable. Communications of the ACM, 27(11):1134--1142, 1984.

Error tolerance

Equivalence

  • D.Haussler, M.Kearns, N.Littlestone and M. Warmuth, Equivalence of models for polynomial learnability, Proc. 1st ACM Workshop on Computational Learning Theory, (1988) 42-55.
  • L. Pitt and M. K. Warmuth: Prediction preserving reduction, Journal of Computer System and Science 41, 430--467, 1990.

A description of some of these publications is given at important publications in machine learning.

External links

Template:SIB

Template:WH Template:WS