Two fundamental measures of the efficiency of a learning algorithm are its running time and the number of examples it requires (its sample complexity). In this paper we demonstrate that even for simple concept classes, an inherent tradeoff can exist between running time and sample complexity. We pre
On parallel attribute-efficient learning
โ Scribed by Peter Damaschke
- Publisher
- Elsevier Science
- Year
- 2003
- Tongue
- English
- Weight
- 209 KB
- Volume
- 67
- Category
- Article
- ISSN
- 0022-0000
No coin nor oath required. For personal study only.
โฆ Synopsis
This paper continues our earlier work on (non)adaptive attribute-efficient learning. We consider exact learning of Boolean functions of n variables by membership queries, assuming that at most r variables are relevant. The learner works in consecutive rounds, such that the set of simultaneous queries in every round may depend on all information gained so far. For deterministic learning of specific monotone functions we prove that any strategy that uses an optimal query number needs Yรฐrร rounds in the worst case. Furthermore, we make some progress regarding the constant factors in nearly query-optimal strategies. For example, we propose a strategy using roughly 2 rรพ1 รพ 2Z log 2 n queries in 3r rounds. In contrast to the limitations of deterministic strategies, there is a randomized strategy that learns monotone functions by 2 O รฐrร รพ Oรฐr log nร expected queries in Oรฐlog rร expected rounds. Actually, this result holds in more general function classes. The second part of the paper addresses the computational complexity of parallel learning of arbitrary Boolean functions with r relevant variables. We obtain several strategies which use a constant number of rounds, Oรฐ2 r polyรฐr log nรร queries, and only 2 Oรฐrร n polyรฐlog nร computations.
๐ SIMILAR VOLUMES
We consider the problem of attribute-efficient learning in query and mistake-bound models. Attribute-efficient algorithms make a number of queries or mistakes that is polynomial in the number of relevant variables in the target function, but only sublinear in the number of irrelevant variables. We c