Article thumbnail

A Bayesian Analysis of Algorithms for Learning Finite Functions

By James Cussens


We consider algorithms for learning functions f : X ! Y , where X and Y are finite, and there is assumed to be no noise in the data. Learning algorithms, Alg, are connected with \Gamma(Alg), the set of prior probability distributions for which they are optimal. A method for constructing \Gamma(Alg) from Alg is given and the relationship between the various \Gamma(Alg) is discussed. Improper algorithms are identified as those for which \Gamma(Alg) has zero volume. Improper algorithms are investigated using linear algebra and two examples of improper algorithms are given. This framework is then applied to the question of choosing between competing algorithms. "Leave-oneout " cross-validation is hence characterised as a crude method of ML-II prior selection. We conclude by examining how the mathematical results bear on practical problems and by discussing related work, as well as suggesting future work. 1 Introduction Given the plethora of different learning algorithms produced by the m..

Publisher: Morgan Kaufmann Publishers
Year: 1995
DOI identifier: 10.1016/b978-1-55860-377-6.50026-8
OAI identifier: oai:CiteSeerX.psu:
Provided by: CiteSeerX
Download PDF:
Sorry, we are unable to provide the full text but you may find it at the following location(s):
  • (external link)
  • (external link)
  • Suggested articles

    To submit an update or takedown request for this paper, please submit an Update/Correction/Removal Request.