Knowledge Graph-Augmented GNN Encoder with Transformer Decoder for Cross-Lingual Neural Machine Translation: Modeling, Optimization, and Scalable Deployment

Bo Wang

Abstract


In cross-language information interaction, translation accuracy and efficiency determine the reliability of multilingual services. This paper proposes a knowledge graph-augmented neural machine translation framework that integrates a 3-layer Graph Attention Network (GAT) encoder with a 6-layer Transformer decoder, fused with XLM-R contextual embeddings. The fusion mechanism injects entity and relation information into decoding, enhancing semantic alignment and long-sentence reasoning. A distributed parallel architecture with cache optimization supports scalable deployment. Experiments on WMT and OpenSubtitles datasets evaluate the model against Transformer and NMT baselines. Results show that compared with the vanilla Transformer and mBART baselines, the proposed system achieves an average BLEU improvement of 17.0%±0.6, a 26.8%±1.5 reduction in PPL, inference time shortened to 0.92s±0.04, and entity alignment error rate reduced to 3.4%±0.3. These results confirm that the performance gains are statistically significant, with confidence intervals reported for all metrics. These findings confirm that knowledge graph augmentation substantially improves translation quality, semantic consistency, real-time performance, and robustness under multilingual and complex contexts. The contributions of this work include: ①a GAT-based encoder for capturing cross-lingual dependencies; ② a fusion method with XLM-R for semantic enhancement; ③ a scalable optimization framework ensuring low-latency translation. This research provides a reproducible and deployable approach for intelligent multilingual interaction and demonstrates significant potential for cross-language applications.


Full Text:

PDF

References


Zhao Y, Kang X, Zhang Y, Zhou Y, Zong C. Knowledge Graph Guided Neural Machine Translation with Dynamic Reinforce-Selected Triples [J]. ACM Trans. Asian Low-Resour. Lang. Inf. Process.2024,23(12):163:1-163:21.https://doi.org/10.1145/3696664

Srivastava N, Perevalov A, Kuchelev D, Moussallem D, Ngomo A-C N, Both A. Lingua Franca – Entity-Aware Machine Translation Approach for Question Answering over Knowledge Graphs [C]. Proceedings of the 12th Knowledge Capture Conference (K-CAP), ACM, 2023: 122–130.https://doi.org/10.1145/3587259.3627567

Zhao Y , Zhang J , Zhou Y ,et al.Knowledge Graphs Enhanced Neural Machine Translation[C]//Twenty-Ninth International Joint Conference on Artificial Intelligence and Seventeenth Pacific Rim International Conference on Artificial Intelligence {IJCAI-PRICAI-20.2020.https://doi.org/10.24963/ijcai.2020/555.

Li Z , Huang R , Zhang Y ,et al.Two flexible translation-based models for knowledge graph embedding[J].J. Intell. Fuzzy Syst. 2022,44:3093-3105.https://doi.org/10.3233/jifs-211553

Hu B , Ye Y , Zhong Y ,et al.TransMKR: Translation-based knowledge graph enhanced multi-task point-of-interest recommendation[J].Neurocomputing, 2022(14):474.https://dol.org/10.1016/j.neucom.2021.11.049

Yu M , Bai C , Yu J ,et al.Translation-Based Embeddings with Octonion for Knowledge Graph Completion[J].Applied Sciences,2022,12(8).2076-3417.https://doi.org/10.3390/app12083935

Wan B , Niu Y , Chen C ,et al.TransRFT: A Knowledge Representation Learning Model Based on a Relational Neighborhood and Flexible Translation[J].Applied Sciences-Basel, 2023,13(19):11.https://doi.org/10.3390/app131910864

Yueting X ,Yangfen C ,Wenjuan F , et al.Enhancing computer-aided translation system with BiLSTM and convolutional neural network using a knowledge graph approach[J].The Journal of Supercomputing,2023,80(5):5847-5869.https://doi.org/10.1007/s11227-023-05686-2

Zepeng L ,Rikui H ,Yufeng Z , et al.Two flexible translation-based models for knowledge graph embedding[J].Journal of Intelligent & Fuzzy Systems,2023,44(2):3093-3105.https:/doi.org/10.3233/jifs-211553

Zhang Y, Wang Z, Qiu H, Li D. Enhancing temporal knowledge graph representation with curriculum learning[J]. Electronics, 2023,13(17):3397.https://doi.org/10.3390/electronics13173397

Bojing H ,Yaqin Y ,Yingqiang Z , et al.TransMKR: Translation-based knowledge graph enhanced multi-task point-of-interest recommendation[J].Neurocomputing,2022,474107-114.https://dol.org/10.1016/j.neucom.2021.11.049

Hussein B ,Hussein H ,Hassan H , et al.TransModE: Translational Knowledge Graph Embedding Using Modular Arithmetic[J].Procedia Computer Science,2022,2071154-1163.https://doi.org/10.1016/j.procs.2022.09.171

Xiong W, Hoang T, Wang W Y. DeepPath: A Reinforcement Learning Method for Knowledge Graph Reasoning [J]. Proc. of EMNLP,2017,(September):564–573.https://doi.org/10.18653/v1/D17-1060

Li D. Causal Reinforcement Learning for Knowledge Graph Reasoning [J]. Applied Sciences,2024,14(6):2498.https://doi.org/10.3390/app14062498

Peng H , Wu Y .A Dynamic Convolutional Network-Based Model for Knowledge Graph Completion[J].Inf.2022,13:133.https://doi.org/10.3390/info13030133

Mohiuddin T, Koehn P, Chaudhary V, Cross J, Bhosale S, Joty S. Data selection curriculum for neural machine translation[J]. Findings of the Association for Computational Linguistics: EMNLP, 2022,(1):1569–1582.https://doi.org/10.48550/arXiv.2203.13867

Zhang X , Zhang C , Guo J ,et al.Graph attention network with dynamic representation of relations for knowledge graph completion[J].Expert Systems with Applications,2023,219:119616-.https://doi.org/10.1016/j.eswa.2023.119616

Liu Y, Zhou C, Zhang P, Cao Y, Liu Y, Li Z, Chen H. CL4KGE: A curriculum learning method for knowledge graph embedding[J]. arXiv preprint, 2024, arXiv:2408.14840.https://dol.org/10.48550/arXiv.2408.14840

Ngoc Q, Wang Q, Gao Z, Sun Q, Guan B, Floyd L. Reinforcement Learning Approach for Integrating Compressed Contexts into Knowledge Graphs [J]. arXiv preprint,2024,arXiv:2404.12587.https://dol.org/10.48550/arXiv.2404.12587

Conia S, Lee D, Li M, Minhas U F, Potdar S, Li Y. Towards cross-cultural machine translation with retrieval-augmented generation from multilingual knowledge graphs[J]. Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing, 2024, (1): 16343–16360.https://dol.org/10.48550/arXiv.2410.14057




DOI: https://doi.org/10.31449/inf.v49i26.11515

Creative Commons License
This work is licensed under a Creative Commons Attribution 3.0 License.