Vidal's libraryTitle: | Instance-Based Learning Algorithms |
Author: | David W. Aha, Dennis Kibler, and Marc K. Albert |
Journal: | Machine Learning |
Volume: | 6 |
Number: | 1 |
Pages: | 33--66 |
Year: | 1991 |
DOI: | 10.1023/A:1022689900470 |
Abstract: | Storing and using specific instances improves the performance of several supervised learning algorithms. These include algorithms that learn decision trees, classification rules, and distributed networks. However, no investigation has analyzed algorithms that use only specific instances to solve incremental learning tasks. In this paper, we describe a framework and methodology, called instance-based learning, that generates classification predictions using only specific instances. Instance-based learning algorithms do not maintain a set of abstractions derived from specific instances. This approach extends the nearest neighbor algorithm, which has large storage requirements. We describe how storage requirements can be significantly reduced with, at most, minor sacrifices in learning rate and classification accuracy. While the storage-reducing algorithm performs well on several realworld databases, its performance degrades rapidly with the level of attribute noise in training instances. Therefore, we extended it with a significance test to distinguish noisy instances. This extended algorithm's performance degrades gracefully with increasing noise levels and compares favorably with a noise-tolerant decision tree algorithm. |
Cited by 1026 - Google Scholar
@Article{aha91a,
author = {David W. Aha and Dennis Kibler and Marc K. Albert},
title = {Instance-Based Learning Algorithms},
googleid = {CKC64N7thR4J:scholar.google.com/},
journal = {Machine Learning},
year = 1991,
volume = 6,
number = 1,
pages = {33--66},
abstract = {Storing and using specific instances improves the
performance of several supervised learning
algorithms. These include algorithms that learn
decision trees, classification rules, and
distributed networks. However, no investigation has
analyzed algorithms that use only specific instances
to solve incremental learning tasks. In this paper,
we describe a framework and methodology, called
instance-based learning, that generates
classification predictions using only specific
instances. Instance-based learning algorithms do not
maintain a set of abstractions derived from specific
instances. This approach extends the nearest
neighbor algorithm, which has large storage
requirements. We describe how storage requirements
can be significantly reduced with, at most, minor
sacrifices in learning rate and classification
accuracy. While the storage-reducing algorithm
performs well on several realworld databases, its
performance degrades rapidly with the level of
attribute noise in training instances. Therefore, we
extended it with a significance test to distinguish
noisy instances. This extended algorithm's
performance degrades gracefully with increasing
noise levels and compares favorably with a
noise-tolerant decision tree algorithm.},
keywords = {ai learning},
doi = {10.1023/A:1022689900470},
cluster = {2199425534549205000}
}
Last modified: Wed Mar 9 10:13:47 EST 2011