International Journal of Soft Computing

Year: 2016
Volume: 11
Issue: 4
Page No. 255 - 269

Improving Classification Performance by Using Feature Selection with Resampling

Authors : Raya Ismail, Sherihan Abuelenin and Ahmed Aboelfetouh

Abstract: Feature selection methods tend to identify the most relevant features for classification and can be categorized as either subset selection (wrapper) methods or ranking (filter) methods. The main purpose of this stusy is to prove that a feature selection preprocessing step could enhance classifiers performance by eliminating redundant features. The proposed method consists of three stages; the first refines sample space domain by resample filtering, the second minimizes feature space by applying subset evaluation algorithm and the third measures the goodness of the resulting set of features using different classifiers. Tow experiments carried out on the data sets from UCI repository. The proposed method is evaluated by measuring the accuracy, number of selected features, precision, recall, f-measure, ROC area, time to build model, error rate and relative absolute error. Tests are done on two main types of classifiers Na�ve Bayes and its variance NBTREE, NBNET and J48 with other tree classifiers Random Forest, BFTREE.

How to cite this article:

Raya Ismail, Sherihan Abuelenin and Ahmed Aboelfetouh, 2016. Improving Classification Performance by Using Feature Selection with Resampling. International Journal of Soft Computing, 11: 255-269.

Design and power by Medwell Web Development Team. © Medwell Publishing 2024 All Rights Reserved