Effectiveness of the use of algorithms and methods of artificial technologies for sign language recognition for people with disabilities

Authors

DOI:

https://doi.org/10.15587/1729-4061.2022.262509

Keywords:

neural network model, convolutional neural network, LSTM module, convolution, sign language

Abstract

According to WHO, the number of people with disabilities in the world has exceeded 1 billion. At the same time, 80 percent of all people with disabilities live in developing countries. In this regard, the demand for the use of applications for people with disabilities is growing every day. The paper deals with neural network methods like MediaPipe Holistic and the LSTM module for determining the sign language of people with disabilities. MediaPipe has demonstrated unprecedented low latency and high tracking accuracy in real-world scenarios thanks to built-in monitoring solutions. Therefore, MediaPipe Holistic was used in this work, which combines pose, hand, and face control with detailed levels.

The main purpose of this paper is to show the effectiveness of the HAR algorithm for recognizing human actions, based on the architecture of in-depth learning for classifying actions into seven different classes. 

The main problem of this paper is the high level of recognition of the sign language of people with disabilities when implementing their work in cross-platform applications, web applications, social networks that facilitate the daily life of people with disabilities and interact with society. To solve this problem, an algorithm was used that combines the architecture of a convolutional neural network (CNN) and long short-term memory (LSTM) to study spatial and temporal capabilities from three-dimensional skeletal data taken only from a Microsoft Kinect camera. This combination allows you to take advantage of LSTM when modeling temporal data and CNN when modeling spatial data.

The results obtained based on calculations carried out by adding a new layer to the existing model showed higher accuracy than calculations carried out on the existing model.

Author Biographies

Aigulim Bayegizova, L. N. Gumilyov Eurasian National University

Candidate of Physical and Mathematical Sciences, Assistant Professor

Department of Radio Engineering, Electronics and Telecommunications

Aisulu Ismailova, S. Seifullin Kazakh Agrotechnical University

PhD, Associate Professor

Department of Information Systems

Ulzada Aitimova, S. Seifullin Kazakh Agrotechnical University

Candidate of Physical and Mathematical Sciences, acting Associate Professor

Department of Information Systems

Ayagoz Mukhanova, L. N. Gumilyov Eurasian National University

PhD, Associate Professor

Department of Information Systems

Zhanar Beldeubayeva, S. Seifullin Kazakh Agrotechnical University

PhD, Senior Lecturer

Department of Information Systems

Aliya Ainagulova, S. Seifullin Kazakh Agrotechnical University

Candidate of Technical Sciences, Senior Lecturer

Department of Information Systems

Akgul Naizagarayeva, S. Seifullin Kazakh Agrotechnical University

Master of Engineering

Department of Information Systems

References

  1. Rastgoo, R., Kiani, K., Escalera, S. (2021). Sign Language Recognition: A Deep Survey. Expert Systems with Applications, 164, 113794. doi: https://doi.org/10.1016/j.eswa.2020.113794
  2. Chuikov, A. V., Vulfin, A. M. (2017). Gesture recognition system. Vestnik UGATU, 21 (3 (77)), 113–122. Available at: https://cyberleninka.ru/article/n/sistema-raspoznavaniya-zhestov-na-osnove-neyrosetevyh-tehnologiy
  3. Wang, M., Lyu, X.-Q., Li, Y.-J., Zhang, F.-L. (2020). VR content creation and exploration with deep learning: A survey. Computational Visual Media, 6 (1), 3–28. doi: https://doi.org/10.1007/s41095-020-0162-z
  4. Murlin, A. G., Piotrovskiy, D. L., Rudenko, E. A., Yanaeva, M. V. (2014). Algorithms and methods for detection and recognition of hand gestures on video in real time. Politematicheskiy setevoy elektronniy nauchnyy zhurnal Kubanskogo gosudarstvennogo agrarnogo universiteta, 97, 626–635. Available at: https://www.elibrary.ru/item.asp?id=21527334
  5. Rautaray, S. S., Agrawal, A. (2012). Vision based hand gesture recognition for human computer interaction: a survey. Artificial Intelligence Review, 43 (1), 1–54. doi: https://doi.org/10.1007/s10462-012-9356-9
  6. Bae, S. H., Choi, I. K., Kim, N. S. (2016). Acoustic Scene Classification Using Parallel Combination of LSTM and CNN. Detection and Classification of Acoustic Scenes and Events. Available at: https://dcase.community/documents/workshop2016/proceedings/Bae-DCASE2016workshop.pdf
  7. Lee, D.-H., Hong, K.-S. (2010). A Hand gesture recognition system based on difference image entropy. In 2010 6th International Conference on Advanced Information Management and Service (IMS), 410–413.
  8. Chen, Y., Luo, B., Chen, Y.-L., Liang, G., Wu, X. (2015). A real-time dynamic hand gesture recognition system using kinect sensor. 2015 IEEE International Conference on Robotics and Biomimetics (ROBIO). doi: https://doi.org/10.1109/robio.2015.7419071
  9. Korchenko, A., Tereykovskiy, I., Karpinskiy, N., Tynymbaev, S. (2016). Neyrosetevye modeli, metody i sredstva otsenki parametrov bezopasnosti internet-orientirovannykh informatsionnykh sistem. Kyiv: TOV "Nash Format".
  10. Top 10 Deep Learning Algorithms You Should Know in 2022. Available at: https://www.simplilearn.com/tutorials/deep-learning-tutorial/deep-learning-algorithm
  11. Liu, N., Lovell, B. C. (2003). Gesture classification using hidden markov models and viterbi path counting. In VII-th Digital image computing: techniques and Applications. Available at: https://www.researchgate.net/publication/37616560_Gesture_Classification_Using_Hidden_Markov_Models_and_Viterbi_Path_Counting
  12. Phan, N. H., Bui, T. T. T., Spitsyn, V. G. (2013). Real-time hand gesture recognition base on Viola-Jones method, algorithm CAMShift, wavelet transform and principal component analysis. Upravlenie, vychislitel'naya tekhnika i informatika, 2 (23), 102–111. Available at: https://cyberleninka.ru/article/n/raspoznavanie-zhestov-na-videoposledovatelnosti-v-rezhime-realnogo-vremeni-na-osnove-primeneniya-metoda-violy-dzhonsa-algoritma
  13. Tkhang, N. T., Spitsyn, V. G. (2012). Algoritmicheskoe i programmnoe obespechenie dlya raspoznavaniya formy ruki v real'nom vremeni s ispol'zovaniem surfcdeskriptorov i neyronnoy seti. Izvestiya Tomskogo politekhnicheskogo universiteta. Inzhiniring georesursov, 320 (5), 48–54.

Downloads

Published

2022-08-31

How to Cite

Bayegizova, A., Murzabekova, G., Ismailova, A., Aitimova, U., Mukhanova, A., Beldeubayeva, Z., Ainagulova, A., & Naizagarayeva, A. (2022). Effectiveness of the use of algorithms and methods of artificial technologies for sign language recognition for people with disabilities. Eastern-European Journal of Enterprise Technologies, 4(2(118), 25–31. https://doi.org/10.15587/1729-4061.2022.262509