Evaluation of Rule-Based Learning and Feature Selection Approaches For Classification

Authors Fatima Chiroma, Mihaela Cocea, Han Liu



PDF
Thumbnail PDF

File

OASIcs.ICCSW.2018.6.pdf
  • Filesize: 308 kB
  • 6 pages

Document Identifiers

Author Details

Fatima Chiroma
  • School of Computing, University of Portsmouth, United Kingdom
Mihaela Cocea
  • School of Computing, University of Portsmouth, United Kingdom
Han Liu
  • School of Computer Science and Informatics, Cardiff University, United Kingdom

Cite AsGet BibTex

Fatima Chiroma, Mihaela Cocea, and Han Liu. Evaluation of Rule-Based Learning and Feature Selection Approaches For Classification. In 2018 Imperial College Computing Student Workshop (ICCSW 2018). Open Access Series in Informatics (OASIcs), Volume 66, pp. 6:1-6:6, Schloss Dagstuhl – Leibniz-Zentrum für Informatik (2019)
https://doi.org/10.4230/OASIcs.ICCSW.2018.6

Abstract

Feature selection is typically employed before or in conjunction with classification algorithms to reduce the feature dimensionality and improve the classification performance, as well as reduce processing time. While particular approaches have been developed for feature selection, such as filter and wrapper approaches, some algorithms perform feature selection through their learning strategy. In this paper, we are investigating the effect of the implicit feature selection of the PRISM algorithm, which is rule-based, when compared with the wrapper feature selection approach employing four popular algorithms: decision trees, naïve bayes, k-nearest neighbors and support vector machine. Moreover, we investigate the performance of the algorithms on target classes, i.e. where the aim is to identify one or more phenomena and distinguish them from their absence (i.e. non-target classes), such as when identifying benign and malign cancer (two target classes) vs. non-cancer (the non-target class).

Subject Classification

ACM Subject Classification
  • Computing methodologies → Feature selection
Keywords
  • Feature Selection
  • Prism
  • Rule-based Learning
  • Wrapper Approach

Metrics

  • Access Statistics
  • Total Accesses (updated on a weekly basis)
    0
    PDF Downloads

References

  1. Neda Abdelhamid, Aladdin Ayesh, and Fadi Thabtah. Phishing detection based Associative Classification data mining. Expert Systems with Applications, 41(13):5948-5959, 2014. Google Scholar
  2. Shigeo Abe. Modified backward feature selection by cross validation. In ESANN, pages 163-168, 2005. Google Scholar
  3. Maher Aburrous, M Alamgir Hossain, Keshav Dahal, and Fadi Thabtah. Intelligent phishing detection system for e-banking using fuzzy data mining. Expert systems with applications, 37(12):7913-7921, 2010. Google Scholar
  4. Jesús Alcalá-Fdez, Alberto Fernández, Julián Luengo, Joaquín Derrac, Salvador García, Luciano Sánchez, and Francisco Herrera. Keel data-mining software tool: data set repository, integration of algorithms and experimental analysis framework. Journal of Multiple-Valued Logic &Soft Computing, 17, 2011. Google Scholar
  5. Jadzia Cendrowska. PRISM: An algorithm for inducing modular rules. International Journal of Man-Machine Studies, 27(4):349-370, 1987. URL: http://dx.doi.org/10.1016/S0020-7373(87)80003-2.
  6. Girish Chandrashekar and Ferat Sahin. A survey on feature selection methods. Computers &Electrical Engineering, 40(1):16-28, 2014. Google Scholar
  7. Chih-Chung Chang and Chih-Jen Lin. LIBSVM: a library for support vector machines. ACM transactions on intelligent systems and technology (TIST), 2(3):27, 2011. Google Scholar
  8. Fatima Chiroma, Han Liu, and Mihaela Cocea. Suicide Related Text Classification with Prism Algorithm. International Conference on Machine Learning and Cybernetics (ICMLC), pages 1-6, 2018. Google Scholar
  9. Dua Dheeru and Efi Karra Taniskidou. UCI machine learning repository, 2017. URL: http://archive.ics.uci.edu/ml.
  10. Pedro Domingos. A few useful things to know about machine learning. Communications of the ACM, 55(10):78-87, 2012. Google Scholar
  11. Brian Johnson, Ryutaro Tateishi, and Zhixiao Xie. Using geographically weighted variables for image classification. Remote sensing letters, 3(6):491-499, 2012. Google Scholar
  12. Vipin Kumar and Sonajharia Minz. Feature selection. SmartCR, 4(3):211-229, 2014. Google Scholar
  13. Han Liu, Mihaela Cocea, and Weili Ding. Multi-task learning for intelligent data processing in granular computing context. Granular Computing, 3(3):257-273, 2017. Google Scholar
  14. Mehdi Naseriparsa, Amir-Masoud Bidgoli, and Touraj Varaee. A hybrid feature selection method to improve performance of a group of classification algorithms. International Journal of Computer Applications, 69(17):28-35, 2013. Google Scholar
  15. Rattanawadee Panthong and Anongnart Srivihok. Wrapper feature subset selection for dimension reduction based on ensemble learning algorithm. Procedia Computer Science, 72:162-169, 2015. Google Scholar
  16. The Royal Society. Machine learning: the power and promise of computers that learn by example. Online, April 2017. URL: https://royalsociety.org/~/media/policy/projects/machine-learning/publications/machine-learning-report.pdf.
  17. Jozsef Suto, Stefan Oniga, and Petrica Pop Sitar. Comparison of wrapper and filter feature selection algorithms on human activity recognition. In Computers Communications and Control (ICCCC), 2016 6th International Conference on, pages 124-129. IEEE, 2016. Google Scholar
  18. Ian H Witten, Eibe Frank, Mark A Hall, and Christopher J Pal. Data Mining: Practical machine learning tools and techniques. Morgan Kaufmann, 2016. Google Scholar
Questions / Remarks / Feedback
X

Feedback for Dagstuhl Publishing


Thanks for your feedback!

Feedback submitted

Could not send message

Please try again later or send an E-mail