Comparative Performance of Neural Networks and Ensemble Methods for Command Classification in ALEXA Virtual Assistant
Abstract
Our study investigates the classification of commands for the ALEXA virtual assistant using various machine-learning models. The dataset includes 16,521 samples, and data preprocessing steps, such as vectorization and remove all stop words and punctuation, were applied before training. Decision Trees, Random Forest, Hist Gradient Boosting, AdaBoost, and Neural Networks are employed to classify textual commands into respective classes. The dataset consists of commands and their classes, transformed into feature vectors using the TF-IDF method. Our neural network architecture comprises three dense layers and two dropout layers, totaling 272,850 trainable parameters, and uses RMSprop for optimization and categorical cross-entropy as the loss function. Performance is evaluated utilizing metrics like accuracy, precision, recall, and F1 score. Results have shown that neural networks perform better in comparison to classical algorithms and outperform AdaBoost explicitly in all metrics. The comparative results between neural networks and AdaBoost in evaluation metrics are, respectively, as follows:(0.851695 / 0.620157), (0.857729 / 0.771549),(0.851695 / 0.62057) and (0.85236 / 0.639389). Therefore, deep learning will indeed provide many promises toward solving challenging NLP tasks in a virtual assistant system like Alexa. The findings provide enormous insight into effective methodologies regarding the classification of commands and further establish the relevance of neural networks within extending virtual assistant technology. Further research may consider discussing more recent neural network structures and exploring their scalability and generalizability across several domains and languages.
Full Text:
PDFReferences
C. Zhou, C. Sun, Z. Liu, and F. C. M. Lau, “A C-LSTM Neural Network for Text Classification,” Nov. 2015, [Online]. Available: http://arxiv.org/abs/1511.08630. https://doi.org/10.48550/arXiv.1511.08630
S. Lai, L. Xu, K. Liu, and J. Zhao, “Recurrent Convolutional Neural Networks for Text Classification,” 2015. [Online]. Available: www.aaai.org https://doi.org/10.1609/aaai.v29i1.9513
M. Granik and V. Mesyura, “Fake News Detection Using Naive Bayes Classifier,” in IEEE First Ukraine Conference on Electrical and Computer Engineering (UKRCON), 2017. DOI:10.1109/UKRCON.2017.8100379
J. Kim, S. Jang, S. Choi, and E. Park, “Text Classification using Capsules,” Aug. 2018, [Online]. Available: http://arxiv.org/abs/1808.03976
C. Qiao et al., “A NEW METHOD OF REGION EMBEDDING FOR TEXT CLASSIFICATION,” 2018. https://doi.org/10.1145/3404555.3404643
M. Z. Islam, J. Liu, J. Li, L. Liu, and W. Kang, “A semantics aware random forest for text classification,” in International Conference on Information and Knowledge Management, Proceedings, Association for Computing Machinery, Nov. 2019, 1061–1070. Doi: 10.1145/3357384.3357891.
J. Chen, Z. Gong, and W. Liu, “A nonparametric model for online topic discovery with word embeddings,” Inf Sci (N Y), 504, 32–47, Dec. 2019, Doi: 10.1016/j.ins.2019.07.048.
T. Yao, Z. Zhai, and B. Gao, “Text Classification Model Based on fast Text,” in IEEE International Conference on Artificial Intelligence and Information Systems (ICAIIS), 2020. DOI:10.1109/ICAIIS49377.2020.9194939
F. Gargiulo, S. Silvestri, M. Ciampi, and G. De Pietro, “Deep neural network for hierarchical extreme multi-label text classification,” Applied Soft Computing Journal, 79, 125–138, Jun. 2019, Doi: 10.1016/j.asoc.2019.03.041.
Z. Wang, P. Wang, L. Huang, X. Sun, and H. Wang, “Incorporating Hierarchy into Text Encoder: A Contrastive Learning Approach for Hierarchical Text Classification,” Mar. 2022, [Online]. Available: http://arxiv.org/abs/2203.03825
X. Sun et al., “Text Classification via Large Language Models,” 2023. https://doi.org/10.48550/arXiv.2305.08377
A. Navada, A. Nizam Ansari, S. Patil, and B. A.Sonkamble, “Overview of Use of Decision Tree algorithms in Machine Learning,” IEEE Control and System Graduate Research Colloquium, 2011. DOI:10.1109/ICSGRC.2011.5991826
Y. Y. Song and Y. Lu, “Decision tree methods: applications for classification and prediction,” Shanghai Arch Psychiatry, 27(2):130–135, Apr. 2015, Doi: 10.11919/j.issn.1002-0829.215044.
G. Biau and E. Scornet, “A Random Forest Guided Tour,” Nov. 2015, [Online]. Available: http://arxiv.org/abs/1511.05741
G. Louppe, “Understanding Random Forests: From Theory to Practice,” Jul. 2014, [Online]. Available: http://arxiv.org/abs/1407.7502
Y. J. Ong, Y. Zhou, N. Baracaldo, and H. Ludwig, “Adaptive Histogram-Based Gradient Boosted Trees for Federated Learning,” Dec. 2020, [Online]. Available: http://arxiv.org/abs/2012.06670
A. Guryanov, “Histogram-Based Algorithm for Building Gradient Boosting Ensembles of Piecewise Linear Decision Trees,” in 8th International Conference, AIST 2019 Kazan, Russia, July 17–19, Goos Gerhard and Hartmanis Juris, Eds., Kazan: Springer, Jul. 2019, 39–50. https://doi.org/10.1007/978-3-030-37334-4_4
J. Zhu, H. Zou, S. Rosset, and T. Hastie, “Multi-class AdaBoost *,” 2009. DOI:10.4310/SII. 2009.v2. n3. a8
R. E. Schapire, “Explaining AdaBoost,” 2013. https://doi.org/10.1007/978-3-642-41136-6_5
Y. Goldberg, “Neural Network Methods for Natural Language Processing,” Synthesis Lectures on Human Language Technologies, 10(1):1–311, 2017, Doi: 10.2200/S00762ED1V01Y201703HLT037.
Z. Zhang and M. R. Sabuncu, “Generalized Cross Entropy Loss for Training Deep Neural Networks with Noisy Labels,” 2018. https://doi.org/10.48550/arXiv.1805.07836
T. Kurbiel and S. Khaleghian, “Training of Deep Neural Networks based on Distance Measures using RMSProp,” Aug. 2017, [Online]. Available: http://arxiv.org/abs/1708.01911
R. Elshamy, O. Abu-Elnasr, M. Elhoseny, and S. Elmougy, “Improving the efficiency of RMSProp optimizer by utilizing Nestrove in deep learning,” Sci Rep, 13(1): 2023, Doi: 10.1038/s41598-023-35663-x.
DOI: https://doi.org/10.31449/inf.v49i2.7725

This work is licensed under a Creative Commons Attribution 3.0 License.