Abstract
With the growth of interest in data mining, there has been increasing interest in applying machine learning algorithms to real-world problems. This raises the question of how to evaluate the performance of machine learning algorithms. The standard procedure performs random sampling of predictive accuracy until a statistically significant difference arises between competing algorithms. That procedure fails to take into account the calibration of predictions. An alternative procedure uses an information reward measure (from I.J. Good) which is sensitive both to domain knowledge (predictive accuracy) and calibration. We analyze this measure, relating it to Kullback-Leibler distance. We also apply it to five well-known machine learning algorithms across a variety of problems, demonstrating some variations in their assessments using accuracy vs. information reward. We also look experimentally at information reward as a function of calibration and accuracy.
Original language | English |
---|---|
Title of host publication | Machine Learning: ECML 2001 |
Subtitle of host publication | 12th European Conference on Machine Learning Freiburg, Germany, September 5-7, 2001 Proceedings |
Editors | Luc De Raedt, Peter Flach |
Place of Publication | Berlin Germany |
Publisher | Springer |
Pages | 276-287 |
Number of pages | 12 |
ISBN (Print) | 3540425365 |
DOIs | |
Publication status | Published - 2001 |
Event | European Conference on Machine Learning 2001 - Freiburg, Germany Duration: 5 Jul 2001 → 7 Jul 2001 Conference number: 12th https://link.springer.com/book/10.1007/3-540-44795-4 (Proceedings) |
Publication series
Name | Lecture Notes in Artificial Intelligence |
---|---|
Publisher | Springer |
Volume | 2167 |
ISSN (Print) | 0302-9743 |
Conference
Conference | European Conference on Machine Learning 2001 |
---|---|
Abbreviated title | ECML 2001 |
Country/Territory | Germany |
City | Freiburg |
Period | 5/07/01 → 7/07/01 |
Internet address |
|