CanAI2001.pdf (376.25 kB)
Stacking for Misclassification Cost Performance
conference contribution
posted on 2023-05-26, 07:42 authored by Cameron-Jones, RM, Charman-Williams, AThis paper investigates the application of the multiple classifier technique known as "stacking" [23], to the task of classifier learning for misclassification cost performance, by straightforwardly adapting a technique successfully developed by Ting and Witten [19, 20] for the task of classifier learning for accuracy performance. Experiments are reported comparing the performance of the stacked classifier with that of its component classifiers, and of other proposed cost-sensitive multiple classifier methods -- a variation of "bagging", and two "boosting" style methods. These experiments confirm that stacking is competitive with the other methods that have previously been proposed. Some further experiments examine the performance of stacking methods with different numbers of component classifiers, including the case of stacking a single classifier, and provide the first demonstration that stacking a single classifier can be beneficial for many data sets.
History
Publication status
- Published
Event title
Advances in Artificial Intelligence, 14th Biennial Conference of the Canadian AI Society for Computational Studies of IntelligenceEvent Venue
Ottawa, CanadaDate of Event (Start Date)
2001-06-07Date of Event (End Date)
2001-06-09Rights statement
The original publication is available at www.springerlink.comRepository Status
- Open
Usage metrics
Categories
No categories selectedKeywords
Licence
Exports
RefWorks
BibTeX
Ref. manager
Endnote
DataCite
NLM
DC