Knowledge Graph-Augmented GNN Encoder with Transformer Decoder for Cross-Lingual Neural Machine Translation: Modeling, Optimization, and Scalable Deployment
Abstract
In cross-language information interaction, translation accuracy and efficiency determine the reliability of multilingual services. This paper proposes a knowledge graph-augmented neural machine translation framework that integrates a 3-layer Graph Attention Network (GAT) encoder with a 6-layer Transformer decoder, fused with XLM-R contextual embeddings. The fusion mechanism injects entity and relation information into decoding, enhancing semantic alignment and long-sentence reasoning. A distributed parallel architecture with cache optimization supports scalable deployment. Experiments on WMT and OpenSubtitles datasets evaluate the model against Transformer and NMT baselines. Results show that compared with the vanilla Transformer and mBART baselines, the proposed system achieves an average BLEU improvement of 17.0%±0.6, a 26.8%±1.5 reduction in PPL, inference time shortened to 0.92s±0.04, and entity alignment error rate reduced to 3.4%±0.3. These results confirm that the performance gains are statistically significant, with confidence intervals reported for all metrics. These findings confirm that knowledge graph augmentation substantially improves translation quality, semantic consistency, real-time performance, and robustness under multilingual and complex contexts. The contributions of this work include: ①a GAT-based encoder for capturing cross-lingual dependencies; ② a fusion method with XLM-R for semantic enhancement; ③ a scalable optimization framework ensuring low-latency translation. This research provides a reproducible and deployable approach for intelligent multilingual interaction and demonstrates significant potential for cross-language applications.
Full Text:
PDFReferences
Zhao Y, Kang X, Zhang Y, Zhou Y, Zong C. Knowledge Graph Guided Neural Machine Translation with Dynamic Reinforce-Selected Triples [J]. ACM Trans. Asian Low-Resour. Lang. Inf. Process.2024,23(12):163:1-163:21.https://doi.org/10.1145/3696664
Srivastava N, Perevalov A, Kuchelev D, Moussallem D, Ngomo A-C N, Both A. Lingua Franca – Entity-Aware Machine Translation Approach for Question Answering over Knowledge Graphs [C]. Proceedings of the 12th Knowledge Capture Conference (K-CAP), ACM, 2023: 122–130.https://doi.org/10.1145/3587259.3627567
Zhao Y , Zhang J , Zhou Y ,et al.Knowledge Graphs Enhanced Neural Machine Translation[C]//Twenty-Ninth International Joint Conference on Artificial Intelligence and Seventeenth Pacific Rim International Conference on Artificial Intelligence {IJCAI-PRICAI-20.2020.https://doi.org/10.24963/ijcai.2020/555.
Li Z , Huang R , Zhang Y ,et al.Two flexible translation-based models for knowledge graph embedding[J].J. Intell. Fuzzy Syst. 2022,44:3093-3105.https://doi.org/10.3233/jifs-211553
Hu B , Ye Y , Zhong Y ,et al.TransMKR: Translation-based knowledge graph enhanced multi-task point-of-interest recommendation[J].Neurocomputing, 2022(14):474.https://dol.org/10.1016/j.neucom.2021.11.049
Yu M , Bai C , Yu J ,et al.Translation-Based Embeddings with Octonion for Knowledge Graph Completion[J].Applied Sciences,2022,12(8).2076-3417.https://doi.org/10.3390/app12083935
Wan B , Niu Y , Chen C ,et al.TransRFT: A Knowledge Representation Learning Model Based on a Relational Neighborhood and Flexible Translation[J].Applied Sciences-Basel, 2023,13(19):11.https://doi.org/10.3390/app131910864
Yueting X ,Yangfen C ,Wenjuan F , et al.Enhancing computer-aided translation system with BiLSTM and convolutional neural network using a knowledge graph approach[J].The Journal of Supercomputing,2023,80(5):5847-5869.https://doi.org/10.1007/s11227-023-05686-2
Zepeng L ,Rikui H ,Yufeng Z , et al.Two flexible translation-based models for knowledge graph embedding[J].Journal of Intelligent & Fuzzy Systems,2023,44(2):3093-3105.https:/doi.org/10.3233/jifs-211553
Zhang Y, Wang Z, Qiu H, Li D. Enhancing temporal knowledge graph representation with curriculum learning[J]. Electronics, 2023,13(17):3397.https://doi.org/10.3390/electronics13173397
Bojing H ,Yaqin Y ,Yingqiang Z , et al.TransMKR: Translation-based knowledge graph enhanced multi-task point-of-interest recommendation[J].Neurocomputing,2022,474107-114.https://dol.org/10.1016/j.neucom.2021.11.049
Hussein B ,Hussein H ,Hassan H , et al.TransModE: Translational Knowledge Graph Embedding Using Modular Arithmetic[J].Procedia Computer Science,2022,2071154-1163.https://doi.org/10.1016/j.procs.2022.09.171
Xiong W, Hoang T, Wang W Y. DeepPath: A Reinforcement Learning Method for Knowledge Graph Reasoning [J]. Proc. of EMNLP,2017,(September):564–573.https://doi.org/10.18653/v1/D17-1060
Li D. Causal Reinforcement Learning for Knowledge Graph Reasoning [J]. Applied Sciences,2024,14(6):2498.https://doi.org/10.3390/app14062498
Peng H , Wu Y .A Dynamic Convolutional Network-Based Model for Knowledge Graph Completion[J].Inf.2022,13:133.https://doi.org/10.3390/info13030133
Mohiuddin T, Koehn P, Chaudhary V, Cross J, Bhosale S, Joty S. Data selection curriculum for neural machine translation[J]. Findings of the Association for Computational Linguistics: EMNLP, 2022,(1):1569–1582.https://doi.org/10.48550/arXiv.2203.13867
Zhang X , Zhang C , Guo J ,et al.Graph attention network with dynamic representation of relations for knowledge graph completion[J].Expert Systems with Applications,2023,219:119616-.https://doi.org/10.1016/j.eswa.2023.119616
Liu Y, Zhou C, Zhang P, Cao Y, Liu Y, Li Z, Chen H. CL4KGE: A curriculum learning method for knowledge graph embedding[J]. arXiv preprint, 2024, arXiv:2408.14840.https://dol.org/10.48550/arXiv.2408.14840
Ngoc Q, Wang Q, Gao Z, Sun Q, Guan B, Floyd L. Reinforcement Learning Approach for Integrating Compressed Contexts into Knowledge Graphs [J]. arXiv preprint,2024,arXiv:2404.12587.https://dol.org/10.48550/arXiv.2404.12587
Conia S, Lee D, Li M, Minhas U F, Potdar S, Li Y. Towards cross-cultural machine translation with retrieval-augmented generation from multilingual knowledge graphs[J]. Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing, 2024, (1): 16343–16360.https://dol.org/10.48550/arXiv.2410.14057
DOI: https://doi.org/10.31449/inf.v49i26.11515
This work is licensed under a Creative Commons Attribution 3.0 License.








