Open Access   Article Go Back

A Comparative Study of Spam Detection in Social Networks Using Bayesian Classifier and Correlation Based Feature Subset Selection

Sanjeev Dhawan1 , Kulvinder Singh2 , Meena Devi3

Section:Review Paper, Product Type: Journal Paper
Volume-3 , Issue-8 , Page no. 97-100, Aug-2015

Online published on Aug 31, 2015

Copyright © Sanjeev Dhawan, Kulvinder Singh , Meena Devi . This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

View this paper at   Google Scholar | DPI Digital Library

How to Cite this Paper

  • IEEE Citation
  • MLA Citation
  • APA Citation
  • BibTex Citation
  • RIS Citation

IEEE Style Citation: Sanjeev Dhawan, Kulvinder Singh , Meena Devi, “A Comparative Study of Spam Detection in Social Networks Using Bayesian Classifier and Correlation Based Feature Subset Selection,” International Journal of Computer Sciences and Engineering, Vol.3, Issue.8, pp.97-100, 2015.

MLA Style Citation: Sanjeev Dhawan, Kulvinder Singh , Meena Devi "A Comparative Study of Spam Detection in Social Networks Using Bayesian Classifier and Correlation Based Feature Subset Selection." International Journal of Computer Sciences and Engineering 3.8 (2015): 97-100.

APA Style Citation: Sanjeev Dhawan, Kulvinder Singh , Meena Devi, (2015). A Comparative Study of Spam Detection in Social Networks Using Bayesian Classifier and Correlation Based Feature Subset Selection. International Journal of Computer Sciences and Engineering, 3(8), 97-100.

BibTex Style Citation:
@article{Dhawan_2015,
author = { Sanjeev Dhawan, Kulvinder Singh , Meena Devi},
title = {A Comparative Study of Spam Detection in Social Networks Using Bayesian Classifier and Correlation Based Feature Subset Selection},
journal = {International Journal of Computer Sciences and Engineering},
issue_date = {8 2015},
volume = {3},
Issue = {8},
month = {8},
year = {2015},
issn = {2347-2693},
pages = {97-100},
url = {https://www.ijcseonline.org/full_paper_view.php?paper_id=616},
publisher = {IJCSE, Indore, INDIA},
}

RIS Style Citation:
TY - JOUR
UR - https://www.ijcseonline.org/full_paper_view.php?paper_id=616
TI - A Comparative Study of Spam Detection in Social Networks Using Bayesian Classifier and Correlation Based Feature Subset Selection
T2 - International Journal of Computer Sciences and Engineering
AU - Sanjeev Dhawan, Kulvinder Singh , Meena Devi
PY - 2015
DA - 2015/08/31
PB - IJCSE, Indore, INDIA
SP - 97-100
IS - 8
VL - 3
SN - 2347-2693
ER -

VIEWS PDF XML
2535 2325 downloads 2350 downloads
  
  
           

Abstract

The article gives an overview of some of the most popular machine learning methods (Naïve Bayesian classifier, naïve Bayesian k-cross validation, naïve Bayesian info gain, Bayesian classification and Bayesian net with correlation based feature subset selection) and of their applicability to the problem of spam-filtering. Brief descriptions of the algorithms are presented, which are meant to be understandable by a reader not familiar with them before. Classification and clustering techniques in data mining are useful for a wide variety of real time applications dealing with large amount of data. Some of the application areas of data mining are text classification, medical diagnosis, intrusion detection systems etc. The Naive Bayesian Classifier technique is based on the Bayesian theorem and is particularly suited when the dimensionality of the inputs is high. Despite its simplicity, Naive Bayesian can often outperform more sophisticated classification methods. The approach is called “naïve” because it assumes the independence between the various attribute values. Naïve Bayesian classification can be viewed as both a descriptive and a predictive type of algorithm. The probabilities are descriptive are used to predict the class membership for a untrained data.

Key-Words / Index Term

Bayesian Classifier, Feature Subset Selection, Naïve Bayesian Classifier, Correlation Based FSS, Info Gain, K-cross validation, Spam, Non-Spam

References

[1] Rushdi Shams and Robert Mercer,” Classifying Spam Emails using Text and Readability Features,” IEEE 13th International Conference on Data Mining (ICDM), 2013, pp. 657-666.
[2] Chotirat “ANN” Ratana Mahatana and Dimitrios Gunppulos,” Feature Selection For the Naïve Bayesian Classifier Using Decision Trees,” Applied Artificial Intelligence, Volume-17, 2003, pp. 475-487.
[3] Mehdi Naseriparsa, Amir-Masoud Bidgoli, Touraj Varaee,”A Hybrid Feature Selection Method to Improve Performance of a Group of Classification Algorithms,” International Journal of Computer Applications (0975-8887), Volume 69, No-17, May 2013.
[4] Aakriti Aggarwal and Ankur Gupta, “Detection of DDoS Attack Using UCLA Dataset on Different Classifiers, International Journal of Computer Science and Engineering, Volume-03, Issue-08, August 2015, pp. 33-37.
[5] Ioannis Kanaris, Konstantinos Kanaris, Ioannis Houvardas, And Efstathios Stamatatos, “Words Vs. Character N-Grams For Anti-Spam Filtering,” International Journal on Artificial Intelligence Tools, 2006, pp.1-20.
[6] Mehdi Naseriparsa, Amir-Masoud Bidgoli and Touraj Varaee,” A Hybrid Feature Selection Method to Improve Performance of a Group of Classification Algorithms” International Journal of Computer Applications (0975 – 8887),Volume 69, Issue- 17,May 2013
[7] Sanjeev Dhawan and Meena Devi, “Spam Detection in Social Networks Using Correlation Based Feature Subset Selection,” International Journal of Computer Applications Technology and Research, Volume 4, Issue-8, August 2015, pp. 629-632.
[8] Dipali Bhosale and Roshani Ade,” Feature Selection based Classification using Naive Bayesian, J48 and Support Vector Machine,” International Journal of Computer Applications (0975 – 8887) Volume 99– No.16, August 2014.
[9] Anjana Kumari,” Study on Naive Bayesian Classifier and its relation to Information Gain,” International Journal on Recent and Innovation Trends in Computing and Communication, Volume: 2, Issue- 3, March 2014, pp.601 – 603.