Iterative learning from positive data and negative counterexamples

被引:0
|
作者
Jain, Sanjay [1 ]
Kinber, Efim
机构
[1] Natl Univ Singapore, Sch Comp, Singapore 117543, Singapore
[2] Univ Sacred Heart, Dept Comp Sci, Fairfield, CT 06432 USA
来源
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
A model for learning in the limit is defined where a (so-called iterative) learner gets all positive examples from the target language, tests every new conjecture with a teacher (oracle) if it is a subset of the target language (and if it is not, then it receives a negative counterexample), and uses only limited long-term memory (incorporated in conjectures). Three variants of this model are compared: when a learner receives least negative counterexamples, the ones whose size is bounded by the maximum size of input seen so far, and arbitrary ones. We also compare our learnability model with other relevant models of learnability in the limit, study how our model works for indexed classes of recursive languages, and show that learners in our model can work in non-U-shaped way - never abandoning the first right conjecture.
引用
收藏
页码:154 / 168
页数:15
相关论文
共 50 条