๐”– Bobbio Scriptorium
โœฆ   LIBER   โœฆ

General Bounds on the Number of Examples Needed for Learning Probabilistic Concepts

โœ Scribed by Hans Ulrich Simon


Publisher
Elsevier Science
Year
1996
Tongue
English
Weight
631 KB
Volume
52
Category
Article
ISSN
0022-0000

No coin nor oath required. For personal study only.

โœฆ Synopsis


Given a p-concept class C, we define two important functions d C (#), d$ C (#) (related to the notion of #-shattering). We prove a lower bound of 0((d C (#)&1)ร‚(=# 2 )) on the number of examples required for learning C with an (=, #)-good model of probability. We prove similar lower bounds for some other learning models like learning with #-bounded absolute (or quadratic) difference or learning with a #-good decision rule. For the class ND of nondecreasing p-concepts on the real domain, d ND (#)=0(1ร‚#). It can be shown that the resulting lower bounds for learning ND (within the models in consideration) are tight to within a logarithmic factor. In order to get the ``almost-matching'' upper bounds, we introduce a new method for designing learning algorithms: dynamic partitioning of the domain by use of splitting trees. The paper also contains a discussion of the gaps between the general lower bounds and the corresponding general upper bounds. It can be shown that, under very mild conditions, these gaps are quite narrow.


๐Ÿ“œ SIMILAR VOLUMES


On Bounding the Number of Generators for
โœ Stephanie Fitchett ๐Ÿ“‚ Article ๐Ÿ“… 2001 ๐Ÿ› Elsevier Science ๐ŸŒ English โš– 145 KB

Let X be the surface obtained by blowing up general points p 1 p n of the projective plane over an algebraically closed ground field k, and let L be the pullback to X of a line on the plane. If C is a rational curve on X with C โ€ข L = d, then for every t there is a natural map C C t โŠ— X X L โ†’ C C t +