Centre for Discrete and Applicable Mathematics, London School of Economics and Political Science
Abstract
This paper concerns learning binary-valued functions defined on IR, and investigates how a particular type of ‘regularity’ of hypotheses can be used to obtain better generalization error bounds. We derive error bounds that depend on the sample width (a notion similar to that of sample margin for real-valued functions). This motivates learning algorithms that seek to maximize sample width
Is data on this page outdated, violates copyrights or anything else? Report the problem now and we will take corresponding actions after reviewing your request.