The present paper investigates identification of indexed families L Of recursively enumerable languages from good examples. We distinguish class-presenting learning from good examples (the good examples have to be generated with respect to a hypothesis space having the same range as L) and class-comprising learning from good examples (the good examples have to be selected with respect to a hypothesis space comprising the range of L). A learner is required to learn a target language on every finite superset of the good examples for it, If the learners first and only conjecture is correct then the underlying learning model is referred to as finite identification from good examples and if the learner makes a finite number of incorrect conjectures before always outputting a correct one, the model is referred to as limit identification from good examples. In the context of class-preserving learning, it is shown that the learning power of finite and limit identification from good text examples coincide. When class comprising learning From good text examples is concerned, limit identification is strictly more powerful than finite learning, Furthermore, if learning from good informant examples is considered, limit identification is superior to finite identification in the class presenting as well as in the class-comprising case. Finally, we relate the models of learning from good examples to one another as well as to the standard learning models in the context of Gold-style language learning. (C) 2001 Elsevier Science B,V. All rights reserved. [References: 21]
展开▼