Searching for just a few words should be enough to get started. If you need to make more complex queries, use the tips below to guide you.
Issue title: Recent Advances in Language & Knowledge Engineering
Guest editors: David Pinto, Beatriz Beltrán and Vivek Singh
Article type: Research Article
Authors: Balouchzahi, Fazlourrahmana; * | Sidorov, Grigoria; * | Shashirekha, Hosahalli Lakshmaiahb; *
Affiliations: [a] Instituto Politécnico Nacional, Centro de Investigación en Computación, CDMX, Mexico | [b] Department of Computer Science, Mangalore University, Mangalore, India
Correspondence: [*] Correspondence to: Fazlourrahman Balouchzahi and Grigori Sidorov, Instituto Politécnico Nacional, Centro de Investigación en Computación, CDMX, Mexico. [email protected] (Fazlourrahman Balouchzahi); Email: [email protected] (Grigori Sidorov) and Hosahalli Lakshmaiah Shashirekha, Department of Computer Science, Mangalore University, Mangalore, India. E-mail: [email protected] (Hosahalli Lakshmaiah Shashirekha).
Abstract: Complex learning approaches along with complicated and expensive features are not always the best or the only solution for Natural Language Processing (NLP) tasks. Despite huge progress and advancements in learning approaches such as Deep Learning (DL) and Transfer Learning (TL), there are many NLP tasks such as Text Classification (TC), for which basic Machine Learning (ML) classifiers perform superior to DL or TL approaches. Added to this, an efficient feature engineering step can significantly improve the performance of ML based systems. To check the efficacy of ML based systems and feature engineering on TC, this paper explores char, character sequences, syllables, word n-grams as well as syntactic n-grams as features and SHapley Additive exPlanations (SHAP) values to select the important features from the collection of extracted features. Voting Classifiers (VC) with soft and hard voting of four ML classifiers, namely: Support Vector Machine (SVM) with Linear and Radial Basis Function (RBF) kernel, Logistic Regression (LR), and Random Forest (RF) was trained and evaluated on Fake News Spreaders Profiling (FNSP) shared task dataset in PAN 2020. This shared task consists of profiling fake news spreaders in English and Spanish languages. The proposed models exhibited an average accuracy of 0.785 for both languages in this shared task and outperformed the best models submitted to this task.
Keywords: Fake news, Learning approaches, N-grams, Feature engineering, SHAP values
DOI: 10.3233/JIFS-219233
Journal: Journal of Intelligent & Fuzzy Systems, vol. 42, no. 5, pp. 4437-4448, 2022
IOS Press, Inc.
6751 Tepper Drive
Clifton, VA 20124
USA
Tel: +1 703 830 6300
Fax: +1 703 830 2300
[email protected]
For editorial issues, like the status of your submitted paper or proposals, write to [email protected]
IOS Press
Nieuwe Hemweg 6B
1013 BG Amsterdam
The Netherlands
Tel: +31 20 688 3355
Fax: +31 20 687 0091
[email protected]
For editorial issues, permissions, book requests, submissions and proceedings, contact the Amsterdam office [email protected]
Inspirees International (China Office)
Ciyunsi Beili 207(CapitaLand), Bld 1, 7-901
100025, Beijing
China
Free service line: 400 661 8717
Fax: +86 10 8446 7947
[email protected]
For editorial issues, like the status of your submitted paper or proposals, write to [email protected]
如果您在出版方面需要帮助或有任何建, 件至: [email protected]