Using the symmetrical Tau criterion for feature selection decision tree and neural network learning

The data collected for various domain purposes usually contains some features irrelevant tothe concept being learned. The presence of these features interferes with the learning mechanism and as a result the predicted models tend to be more complex and less accurate. It is important to employ an eff...

Full description

Bibliographic Details
Main Authors: Hadzic, Fedja, Dillon, Tharam S
Other Authors: Huan Liu, Robert Stine and Leonardo Auslender
Format: Conference Paper
Published: ACM 2006
Subjects:
Online Access:http://hdl.handle.net/20.500.11937/31352
id curtin-20.500.11937-31352
recordtype eprints
spelling curtin-20.500.11937-313522017-10-02T02:27:39Z Using the symmetrical Tau criterion for feature selection decision tree and neural network learning Hadzic, Fedja Dillon, Tharam S Huan Liu, Robert Stine and Leonardo Auslender feature selection network pruning rule simplification The data collected for various domain purposes usually contains some features irrelevant tothe concept being learned. The presence of these features interferes with the learning mechanism and as a result the predicted models tend to be more complex and less accurate. It is important to employ an effective feature selection strategy so that only the necessary and significant features will be used to learn the concept at hand. The Symmetrical Tau (t) [13] is a statistical-heuristic measure for the capability of an attribute in predicting the class of another attribute, and it has successfully been used as a feature selection criterion during decision tree construction. In this paper we aim to demonstrate some other ways of effectively using the t criterion to filter out the irrelevant features prior to learning (pre-pruning) and after the learning process (post-pruning). For the pre-pruning approach we perform two experiments, one where the irrelevant features are filtered out according to their t value, and one where we calculate the t criterion for Boolean combinations of features and use the highest t-valued combination. In the post-pruning approach we use the t criterion to prune a trained neural network and thereby obtain a more accurate and simple rule set. The experiments are performed on data characterized by continuous and categorical attributes and the effectiveness of the proposed techniques is demonstrated by comparing the derived knowledge models in terms of complexity and accuracy. 2006 Conference Paper http://hdl.handle.net/20.500.11937/31352 ACM fulltext
repository_type Digital Repository
institution_category Local University
institution Curtin University Malaysia
building Curtin Institutional Repository
collection Online Access
topic feature selection
network pruning
rule simplification
spellingShingle feature selection
network pruning
rule simplification
Hadzic, Fedja
Dillon, Tharam S
Using the symmetrical Tau criterion for feature selection decision tree and neural network learning
description The data collected for various domain purposes usually contains some features irrelevant tothe concept being learned. The presence of these features interferes with the learning mechanism and as a result the predicted models tend to be more complex and less accurate. It is important to employ an effective feature selection strategy so that only the necessary and significant features will be used to learn the concept at hand. The Symmetrical Tau (t) [13] is a statistical-heuristic measure for the capability of an attribute in predicting the class of another attribute, and it has successfully been used as a feature selection criterion during decision tree construction. In this paper we aim to demonstrate some other ways of effectively using the t criterion to filter out the irrelevant features prior to learning (pre-pruning) and after the learning process (post-pruning). For the pre-pruning approach we perform two experiments, one where the irrelevant features are filtered out according to their t value, and one where we calculate the t criterion for Boolean combinations of features and use the highest t-valued combination. In the post-pruning approach we use the t criterion to prune a trained neural network and thereby obtain a more accurate and simple rule set. The experiments are performed on data characterized by continuous and categorical attributes and the effectiveness of the proposed techniques is demonstrated by comparing the derived knowledge models in terms of complexity and accuracy.
author2 Huan Liu, Robert Stine and Leonardo Auslender
author_facet Huan Liu, Robert Stine and Leonardo Auslender
Hadzic, Fedja
Dillon, Tharam S
format Conference Paper
author Hadzic, Fedja
Dillon, Tharam S
author_sort Hadzic, Fedja
title Using the symmetrical Tau criterion for feature selection decision tree and neural network learning
title_short Using the symmetrical Tau criterion for feature selection decision tree and neural network learning
title_full Using the symmetrical Tau criterion for feature selection decision tree and neural network learning
title_fullStr Using the symmetrical Tau criterion for feature selection decision tree and neural network learning
title_full_unstemmed Using the symmetrical Tau criterion for feature selection decision tree and neural network learning
title_sort using the symmetrical tau criterion for feature selection decision tree and neural network learning
publisher ACM
publishDate 2006
url http://hdl.handle.net/20.500.11937/31352
first_indexed 2018-09-06T21:44:39Z
last_indexed 2018-09-06T21:44:39Z
_version_ 1610896144501571584