Context-Enriched Dynamic Graph Word Embeddings for Robust NLP Applications
Abstract
Understanding the contextual relationships between words is essential for effective natural language processing(NLP). Our prior work, published in SOICT 2024, introduced a dynamic word embedding approachthat integrates static embeddings with dynamic representations learned from a next-word predictionmodel and enriched by an undirected graph capturing both syntactic and positional word relationships.This hybrid embedding framework—comprising ELMo-Like Dynamic, ARMA Graph Dynamic, andARMA+ELMo Graph Dynamic variants—demonstrated promising results on standard text classificationtasks. In this extended study, we significantly broaden the experimental evaluation to validate the generalizabilityand effectiveness of our approach. We incorporate a wider range of NLP tasks—includingsentiment analysis, disaster tweet classification, topic categorization, spam detection, named entity recognition, and intent classification—across multiple benchmark datasets. Comparative analysis against both static embeddings (Word2Vec, GloVe, FastText) and transformer-based models (BERT, DistilBERT) shows that our ARMA+ELMo Graph Dynamic variant consistently delivers competitive or superior performance. Notably, our method achieves a classification accuracy of 93.2% on the AG News topic classification task and an F1-score of 94.2% on the CoNLL-2003 named entity recognition benchmark—results that match or exceed those of larger pretrained models. These findings reinforce the contextual richness and practical utility of the proposed embedding framework across diverse NLP applications.References
T. Mikolov, K. Chen, G. Corrado, and J. Dean, “Efficient estimation of word representations in vector space,” arXiv preprint arXiv:1301.3781, 2013.
J. Pennington, R. Socher, and C. D. Manning,
“Glove: Global vectors for word representation,” in Proceedings of the 2014 conference on empirical methods in natural language processing (EMNLP), 2014, pp. 1532–1543.
M. E. Peters, M. Neumann, M. Iyyer, M. Gard-
ner, C. Clark, K. Lee, and L. Zettlemoyer, “Deep
contextualized word representations,” in Proceed-
ings of NAACL-HLT, 2018, pp. 2227–2237.
ings of NAACL-HLT, 2018, pp. 2227–2237.
J. Devlin, M.-W. Chang, K. Lee, and
K. Toutanova, “Bert: Pre-training of deep
bidirectional transformers for language under-
standing,” in Proceedings of NAACL-HLT, 2019,
pp. 4171–4186.
R. E. Himes, H.-A. Tran, and T. X. Tran, “Lever-
aging dynamic graph word embedding for ef-
ficient contextual representations,” in Interna-
tional Symposium on Information and Communi-
cation Technology. Springer, 2024, pp. 243–254.
P. Bojanowski, E. Grave, A. Joulin, and
T. Mikolov, “Enriching word vectors with sub-
word information,” Transactions of the Associ-
ation for Computational Linguistics, vol. 5, pp.
–146, 2017.
O. Levy and Y. Goldberg, “Dependency-based
word embeddings,” in Proceedings of the 52nd
Annual Meeting of the Association for Computa-
tional Linguistics (Volume 2), 2014, pp. 302–308.
T. N. Kipf and M. Welling, “Semi-supervised
classification with graph convolutional networks,”
arXiv preprint arXiv:1609.02907, 2016.
F. M. Bianchi, D. Grattarola, L. Livi, and
C. Alippi, “Graph neural networks with convolu-
tional arma filters,” IEEE Transactions on Pat-
tern Analysis and Machine Intelligence, vol. 44,
no. 7, pp. 3496–3507, 2021.
D. Grattarola and C. Alippi, “Graph neural
networks in tensorflow and keras with spek-
tral,” IEEE Computational Intelligence Maga-
zine, vol. 16, no. 1, pp. 99–106, 2021.
E. Saravia, H.-C. T. Liu, Y.-H. Huang, J. Wu, and
Y.-S. Chen, “Carer: Contextualized affect repre-
sentations for emotion recognition,” in Proceed-
ings of the 2018 conference on empirical methods
in natural language processing, 2018, pp. 3687–
A. Howard, d. de Vries, P. Culliton, and
Y. Guo, “Natural language processing with
disaster tweets,” Kaggle competition, 2019,
https://www.kaggle.com/competitions/nlp-
getting-started.
X. Zhang, J. Zhao, and Y. LeCun, “Character-
level convolutional networks for text classifica-
tion,” in Advances in Neural Information Process-
ing Systems 28 (NeurIPS 2015), 2015, pp. 649–
T. A. Almeida, J. M. G. Hidalgo, and A. Ya-
makami, “Contributions to the study of sms spam
filtering: New collection and results,” in Proceed-
ings of the 11th ACM Symposium on Document
Engineering (DocEng ’11), 2011, pp. 259–262.
E. F. Tjong Kim Sang and F. De Meulder, “Intro-
duction to the conll-2003 shared task: Language-
independent named entity recognition,” in Pro-
ceedings of the Seventh Conference on Natural
Language Learning at HLT-NAACL 2003, 2003,
pp. 142–147.
A. Coucke, A. Saade, A. Ball, T. Bluche,
A. Caulier, D. Leroy, C. Doumouro, T. Gissel-
brecht, T. Lavril, M. Primet, and J. Dureau,
“Snips voice platform: An embedded spo-
ken language understanding system for private-
by-design voice interfaces,” arXiv preprint
arXiv:1805.10190, 2018.
V. Sanh, L. Debut, J. Chaumond, and
T. Wolf, “Distilbert, a distilled version of
bert: smaller, faster, cheaper and lighter,”
arXiv preprint arXiv:1910.01108, 2019. [Online].
Available: https://arxiv.org/abs/1910.01108
DOI:
https://doi.org/10.31449/inf.v49i3.10121Downloads
Published
How to Cite
Issue
Section
License
Authors retain copyright in their work. By submitting to and publishing with Informatica, authors grant the publisher (Slovene Society Informatika) the non-exclusive right to publish, reproduce, and distribute the article and to identify itself as the original publisher.
All articles are published under the Creative Commons Attribution license CC BY 3.0. Under this license, others may share and adapt the work for any purpose, provided appropriate credit is given and changes (if any) are indicated.
Authors may deposit and share the submitted version, accepted manuscript, and published version, provided the original publication in Informatica is properly cited.







