Open Access   Article Go Back

Analysis of Different Classifiers’ Performance After Applying Three Different Feature Selection Methods

Kasturi Ghosh1 , Susmita Nandi2

Section:Research Paper, Product Type: Journal Paper
Volume-07 , Issue-01 , Page no. 1-11, Jan-2019

Online published on Jan 20, 2019

Copyright © Kasturi Ghosh, Susmita Nandi . This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

View this paper at   Google Scholar | DPI Digital Library

How to Cite this Paper

  • IEEE Citation
  • MLA Citation
  • APA Citation
  • BibTex Citation
  • RIS Citation

IEEE Style Citation: Kasturi Ghosh, Susmita Nandi, “Analysis of Different Classifiers’ Performance After Applying Three Different Feature Selection Methods,” International Journal of Computer Sciences and Engineering, Vol.07, Issue.01, pp.1-11, 2019.

MLA Style Citation: Kasturi Ghosh, Susmita Nandi "Analysis of Different Classifiers’ Performance After Applying Three Different Feature Selection Methods." International Journal of Computer Sciences and Engineering 07.01 (2019): 1-11.

APA Style Citation: Kasturi Ghosh, Susmita Nandi, (2019). Analysis of Different Classifiers’ Performance After Applying Three Different Feature Selection Methods. International Journal of Computer Sciences and Engineering, 07(01), 1-11.

BibTex Style Citation:
@article{Ghosh_2019,
author = {Kasturi Ghosh, Susmita Nandi},
title = {Analysis of Different Classifiers’ Performance After Applying Three Different Feature Selection Methods},
journal = {International Journal of Computer Sciences and Engineering},
issue_date = {1 2019},
volume = {07},
Issue = {01},
month = {1},
year = {2019},
issn = {2347-2693},
pages = {1-11},
url = {https://www.ijcseonline.org/full_spl_paper_view.php?paper_id=583},
publisher = {IJCSE, Indore, INDIA},
}

RIS Style Citation:
TY - JOUR
UR - https://www.ijcseonline.org/full_spl_paper_view.php?paper_id=583
TI - Analysis of Different Classifiers’ Performance After Applying Three Different Feature Selection Methods
T2 - International Journal of Computer Sciences and Engineering
AU - Kasturi Ghosh, Susmita Nandi
PY - 2019
DA - 2019/01/20
PB - IJCSE, Indore, INDIA
SP - 1-11
IS - 01
VL - 07
SN - 2347-2693
ER -

           

Abstract

Feature selection (FS) is an important aspect of data mining. Now a days availability of information with hundreds of variables leads to high dimensional, irrelevant and redundant data. Thus FS techniques must be applied on the datasets before classification or rule generation. It basically aims at reducing the number of attributes by removing irrelevant or redundant ones, while trying to reduce computation time and improve performance of classifiers. In this paper three different FS methods are used, Correlation Based, Information Gain Based and Rough set Based FS method. A statistical analysis of three different classifier`s performance is also done in order to provide a detailed view.

Key-Words / Index Term

Data Mining (DM), Feature Selection (FS), Rough Set, Degree of Dependency, Decision Tree (J48 algorithm), Naive Bayes Algorithm (NB), K-Nearest Neighbor Algorithm (KNN), Classification, Statistical Analysis

References

[1] Imran Fareed Nizami, Muhammad Majid, Hammad Afzal and Khawar Khurshi, “Impact of Feature Selection Algorithms on Blind Image Quality Assessment”, Arabian Journal for Science and Engineering, pp 1–14, August 2017.
[2] Abdullah S. Ghareb, Abdul Razak Hamdan and Azuraliza Abu Bakar, “Integrating Noun-Based Feature Ranking and Selection Methods with Arabic Text Associative Classification Approach”, Arabian Journal for Science and Engineering, Vol.39, Issue.11, pp 7807–7822, November 2014.
[3] Z. Pawlak, Rough sets, International Journal of Computer and Information Sciences, 11, 341-356, 1982
[4] Javad Rahimipour Anaraki, Kerman, Iran, Mahdi Eftekhari, “Rough Set Based Feature Selection: A Review”, 5th Conference on Information and Knowledge Technology, IEEE, 2013.
[5] G. K. Gupta, “Introduction to Data Mining with Case Studies”, Prentice Hall of India New Delhi, 2006.
[6] P-N. Tan, M. Steinbach, V. Kumar, “Introduction to Data Mining”, Addison Wesley Publishing, 2006.
[7] O.Maimon and L.Rokach, “Data Mining and Knowledge Discovery”, Springer Science and Business Media, 2005.
[8] X. Niuniu and L. Yuxun, “Review of Decision Trees”, IEEE, 2010.
[9] Payam Emami Khoonsari and AhmadReza Motie, “A Comparison of Efficiency and Robustness of ID3 and C4.5 Algorithms Using Dynamic Test and Training Data Sets”, International Journal of Machine Learning and Computing, Vol.2, Issue.5, October 2012.
[10] V. Garcia, C. Debreuve, “Fast k Nearest Neighbor Search using GPU”, IEEE, 2008.
[11] A. Ashari I. Paryudi and A Min Tjoa, “Performance Comparison between Naïve Bayes Decision Tree and k-Nearest Neighbor in Searching Alternative Design in an Energy Simulation Tool”, International Journal of Advanced Computer Science and Applications, Vol.4, Issue. 11, 2013.
[12] Dougherty, J., R. Kohavi and M. Sahami, “Supervised and unsupervised discretization of continuous features”, Proceeding of the 12th International Conference on Machine Learning, 1995.
[13] https:// archive.ics.uci.edu/ ml/ datasets/ Diabetic + Retinopathy + Debrecen + Data + Set
[14] https:// archive.ics.uci.edu/ ml/ datasets/ EEG + Eye + State
[15] https:// archive.ics.uci.edu/ ml/ datasets/ cardiotocography
[16] https://archive.ics.uci.edu/ ml/ datasets/ Thoracic + Surgery + Data
[17] PIDD Dataset, https:// archive.ics.uci.edu/ ml/ datasets/ pima + indians + diabetes
[18] https:// archive.ics.uci.edu/ ml/ datasets/ ILPD + (Indian + Liver + Patient + Dataset)
[19] https:// archive.ics.uci.edu/ ml/ datasets/ breast + cancer + wisconsin + (original)