Characterizations of Monotonic and Dual Monotonic Language Learning

The present paper deals with monotonic and dual monotonic language learning from positive as well as from positive and negative examples. The three notions of monotonicity reflect different formalizations of the requirement that the learner has to produce better and better generalizations when fed m...

Full description

Saved in:
Bibliographic Details
Published inInformation and computation Vol. 120; no. 2; pp. 155 - 173
Main Authors Zeugmann, T., Lange, S., Kapur, S.
Format Journal Article
LanguageEnglish
Published San Diego, CA Elsevier Inc 01.08.1995
Elsevier
Subjects
Online AccessGet full text
ISSN0890-5401
1090-2651
DOI10.1006/inco.1995.1107

Cover

More Information
Summary:The present paper deals with monotonic and dual monotonic language learning from positive as well as from positive and negative examples. The three notions of monotonicity reflect different formalizations of the requirement that the learner has to produce better and better generalizations when fed more and more data on the concept to be learned. The three versions of dual monotonicity describe the concept that the inference device has to produce specializations that fit better and better to the target language. We characterize strong-monotonic, monotonic, weak-monotonic, dual strong-monotonic, dual monotonic, and monotonic & dual monotonic, as well as finite language learning from positive data in terms of recursively generable finite sets. These characterizations provide a unifying framework for learning from positive data under the various monotonicity constraints. Moreover, they yield additional insight into the problem of what a natural learning algorithm should look like.
ISSN:0890-5401
1090-2651
DOI:10.1006/inco.1995.1107