Devising an approach to constructing a specialized dictionary to train chatbots with generative artificial intelligence

Authors

DOI:

https://doi.org/10.15587/1729-4061.2026.351414

Keywords:

large language model, subject-specific knowledge, terminology management, semantic consistency

Abstract

This paper investigates the process that builds a subject-specific training dictionary for training a chatbot involving generative artificial intelligence. The task addressed is to reproduce the structured vocabulary characteristic of the relevant subject area from subject-specific knowledge when interacting with a chatbot.

The result of this study is the construction of a model for the process that sequentially manages independent user requests. The model made it possible to estimate the mathematical expectation of the stage number at which the processing of the request by the chatbot is completed.

Based on the constructed mathematical model, linear and logical-probabilistic models for building a specialized dictionary have been proposed. The linear model searches for a combination of words by sequentially searching for terms. The result of this approach is the comparison of a keyword from the query with the corresponding term or word form from the dictionary. The logical-probabilistic model is based on the target cell – a probable word from the user's query. This is explained by the possibility of defining a word that agrees with the term of the XML dictionary and has maximum relevance to the user query.

A methodology and algorithm for building a specialized dictionary have been suggested. The tests made it possible to obtain average signature values for the response at an error of 0.004%, as well as ensure the stability of the results. In practice, this could be used under the conditions of forming a probability distribution of possible word combinations for generating a response.

The proposed approach could be used in practical tasks of chatbots' domain adaptation, in particular at project support portals and in scientific libraries, as well as for improving intelligent dialog systems focused on the formation of refined user queries

Author Biographies

Olha Kryazhych, Institute of Telecommunications and Global Information Space of the National Academy of Sciences of Ukraine

PhD, Senior Researcher, Associate Professor

Viacheslav Riznyk, Hryhorii Skovoroda University in Pereiaslav

Doctor of Pedagogical Sciences, Associate Professor, Professor

Department of Digital Methods of Teaching

Vasyl Vasenko, Hryhorii Skovoroda University in Pereiaslav

PhD, Associate Professor, Head of Department

Department of Theory and Techniques of Technology Education and Computer Graphics

Vasyl Yakuba, Hryhorii Skovoroda University in Pereiaslav

PhD, Associate Professor

Department of Digital Methods of Teaching

Kateryna Iushchenko, Institute of Telecommunications and Global Information Space of the National Academy of Sciences of Ukraine

Doctor of Philosophy (PhD), Junior Researcher

Oleksii Kuprin, Institute of Telecommunications and Global Information Space of the National Academy of Sciences of Ukraine

Doctor of Philosophy (PhD), Junior Researcher

Oleksandr Tsyrul, Institute of Telecommunications and Global Information Space of the National Academy of Sciences of Ukraine

PhD Student

References

  1. Yang, C., Zhao, R., Liu, Y., Jiang, L. (2025). Survey of specialized large language model. arXiv. https://arxiv.org/abs/2508.19667
  2. Adavala, K. M., Adavala, O. (2025). Domain-specific knowledge and context in large language models: challenges, concerns, and solutions. IAES International Journal of Artificial Intelligence (IJ-AI), 14 (4), 2568. https://doi.org/10.11591/ijai.v14.i4.pp2568-2578
  3. Zhu, Y., Yuan, H., Wang, S., Liu, J., Liu, W., Deng, C. et al. (2025). Large Language Models for Information Retrieval: A Survey. ACM Transactions on Information Systems, 44 (1), 1–54. https://doi.org/10.1145/3748304
  4. Ai, Q., Bai, T., Cao, Z., Chang, Y., Chen, J., Chen, Z. et al. (2023). Information Retrieval meets Large Language Models: A strategic report from Chinese IR community. AI Open, 4, 80–90. https://doi.org/10.1016/j.aiopen.2023.08.001
  5. Sharma, K., Kumar, P., Li, Y. (2025). OG-RAG: Ontology-grounded retrieval-augmented generation for large language models. Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing, 32950–32969. https://doi.org/10.18653/v1/2025.emnlp-main.1674
  6. Manda, P. (2025). Large Language Models in Bio-Ontology Research: A Review. Bioengineering, 12 (11), 1260. https://doi.org/10.3390/bioengineering12111260
  7. Barron, R. C., Grantcharov, V., Wanna, S., Eren, M. E., Bhattarai, M., Solovyev, N. et al. (2024). Domain-Specific Retrieval-Augmented Generation Using Vector Stores, Knowledge Graphs, and Tensor Factorization. 2024 International Conference on Machine Learning and Applications (ICMLA), 1669–1676. https://doi.org/10.1109/icmla61862.2024.00258
  8. Fareedi, A. A., Ismail, M., Ahmed, S., Gagnon, S., Ghazawneh, A., Arooj, Z., Nazir, H. (2025). Enriching Human–AI Collaboration: The Ontological Service Framework Leveraging Large Language Models for Value Creation in Conversational AI. Knowledge, 6 (1), 2. https://doi.org/10.3390/knowledge6010002
  9. Mukanova, A., Nazyrova, A., Zulkhazhav, A., Lamasheva, Z., Dauletkaliyeva, A. (2025). Development of an Intelligent Information Retrieval System Based on Ontology, Linguistic Algorithms and Large Language Models. Applied Sciences, 15 (22), 12271. https://doi.org/10.3390/app152212271
  10. Ahmad, J. M., Liu, Y., Kim, J.-D., Yao, X., Larmande, P., Xia, J. (2025). A curation system of rice trait ontology with reliable interoperation by LLM and PubAnnotation. Genomics & Informatics, 23 (1). https://doi.org/10.1186/s44342-025-00058-z
  11. Chen, L.-C., Pardeshi, M. S., Liao, Y.-X., Pai, K.-C. (2025). Application of retrieval-augmented generation for interactive industrial knowledge management via a large language model. Computer Standards & Interfaces, 94, 103995. https://doi.org/10.1016/j.csi.2025.103995
  12. Wen, J., Liu, D., Xie, Y., Ren, Y., Wang, J., Xia, Y., Zhu, P. (2025). AcuGPT-Agent: An LLM-powered intelligent system for acupuncture-based infertility treatment. Neurocomputing, 652, 131116. https://doi.org/10.1016/j.neucom.2025.131116
  13. Rodríguez-Muñoz-de-Baena, I., Coronado-Vaca, M., Vaquero-Lafuente, E. (2025). Fine-tuning transformer models for M&A target prediction in the U.S. ENERGY sector. Cogent Business & Management, 12 (1). https://doi.org/10.1080/23311975.2025.2487219
  14. Byrd, C., Kingsbury, C., Niell, B., Funaro, K., Bhatt, A., Weinfurtner, R. J., Ataya, D. (2025). Appropriateness of acute breast symptom recommendations provided by ChatGPT. Clinical Imaging, 125, 110549. https://doi.org/10.1016/j.clinimag.2025.110549
  15. Brown, E. D. L., Ward, M., Maity, A., Mittler, M. A., Larry Lo, S.-F., D’Amico, R. S. (2024). Enhancing Diagnostic Support for Chiari Malformation and Syringomyelia: A Comparative Study of Contextualized ChatGPT Models. World Neurosurgery, 189, e86–e107. https://doi.org/10.1016/j.wneu.2024.05.172
  16. Ni, W., Shen, Q., Liu, T., Zeng, Q., Xu, L. (2023). Generating textual emergency plans for unconventional emergencies – A natural language processing approach. Safety Science, 160, 106047. https://doi.org/10.1016/j.ssci.2022.106047
  17. Ganzinger, M., Kunz, N., Fuchs, P., Lyu, C. K., Loos, M., Dugas, M., Pausch, T. M. (2025). Automated generation of discharge summaries: leveraging large language models with clinical data. Scientific Reports, 15 (1). https://doi.org/10.1038/s41598-025-01618-7
  18. Xu, Y., Wang, T., Yuan, Y., Huang, Z., Chen, X., Zhang, B. et al. (2025). LLM-Enhanced Framework for Building Domain-Specific Lexicon for Urban Power Grid Design. Applied Sciences, 15 (8), 4134. https://doi.org/10.3390/app15084134
  19. Keng-Jung, P., Chin-Hung, K., Cheng-Yen, W., Peng, J.-W., Huang, C.-Y., Chen, J.-C. (2021). Analyze the subordination structure between domain-specific vocabulary and meaning with the Word2Vec training process. 2021 IEEE International Conference on Consumer Electronics-Taiwan (ICCE-TW), 1–2. https://doi.org/10.1109/icce-tw52618.2021.9602966
  20. Xu, K., Feng, Y., Li, Q., Dong, Z., Wei, J. (2025). Survey on terminology extraction from texts. Journal of Big Data, 12 (1). https://doi.org/10.1186/s40537-025-01077-x
  21. Lu, R.-S., Lin, C.-C., Tsao, H.-Y. (2024). Empowering Large Language Models to Leverage Domain-Specific Knowledge in E-Learning. Applied Sciences, 14 (12), 5264. https://doi.org/10.3390/app14125264
  22. Kryazhych, O., Ivanov, I., Iushchenko, K., Kupri, O., Vasenko, O., Riznyk, V., Ryzhkov, O. (2025). Devising an approach to preventing information chaos in chat bots using generative artificial intelligence. Eastern-European Journal of Enterprise Technologies, 2 (2 (134)), 84–95. https://doi.org/10.15587/1729-4061.2025.324957
Devising an approach to constructing a specialized dictionary to train chatbots with generative artificial intelligence

Downloads

Published

2026-02-27

How to Cite

Kryazhych, O., Riznyk, V., Vasenko, V., Yakuba, V., Iushchenko, K., Kuprin, O., & Tsyrul, O. (2026). Devising an approach to constructing a specialized dictionary to train chatbots with generative artificial intelligence. Eastern-European Journal of Enterprise Technologies, 1(2 (139), 58–67. https://doi.org/10.15587/1729-4061.2026.351414