Abstract
This paper is concerned with a sufficient condition under which a concept class is learnable in Gold’s classical model of identification in the limit from positive data. The standard principle of learning algorithms working under this model is called the MINL strategy, which is to conjecture a hypothesis representing a minimal concept among the ones consistent with the given positive data. The minimality of a concept is defined with respect to the set-inclusion relation – the strategy is semantics-based. On the other hand, refinement operators have been developed in the field of learning logic programs, where a learner constructs logic programs as hypotheses consistent with given logical formulae. Refinement operators have syntax-based definitions – they are defined based on inference rules in first-order logic. This paper investigates the relation between the MINL strategy and refinement operators in inductive inference. We first show that if a hypothesis space admits a refinement operator with certain properties, the concept class will be learnable by an algorithm based on the MINL strategy. We then present an additional condition that ensures the learnability of the class of unbounded finite unions of concepts. Furthermore, we show that under certain assumptions a learning algorithm runs in polynomial time.
Original language | English |
---|---|
Pages (from-to) | 181-203 |
Number of pages | 23 |
Journal | Annals of Mathematics and Artificial Intelligence |
Volume | 79 |
Issue number | 1-3 |
DOIs | |
Publication status | Published - 2017 Mar 1 |
Externally published | Yes |
Keywords
- Identification in the limit from positive data
- Refinement operator
- Unions of concepts
ASJC Scopus subject areas
- Artificial Intelligence
- Applied Mathematics