Abstract
We study predicate selection functions (also known as splitting rules) for structural decision trees and propose two improvements to existing schemes. The first is in classification learning, where we reconsider the use of accuracy as a predicate selection function and show that, on practical grounds, it is a better alternative to other commonly used functions. The second is in regression learning, where we consider the standard mean squared error measure and give a predicate pruning result for it.
Original language | English |
---|---|
Pages (from-to) | 264-278 |
Number of pages | 15 |
Journal | Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) |
Volume | 3625 |
DOIs | |
Publication status | Published - 2005 |
Event | 15th International Conference on Inductive Logic Programming, ILP 2005 - Bonn, Germany Duration: 10 Aug 2005 → 13 Aug 2005 |