299
views
0
recommends
+1 Recommend
1 collections
    0
    shares

      King Salman Center for Disability Research is pleased to invite you to submit your scientific research to the Journal of Disability Research. JDR contributes to the Center's strategy to maximize the impact of the field, by supporting and publishing scientific research on disability and related issues, which positively affect the level of services, rehabilitation, and care for individuals with disabilities.
      JDR is an Open Access scientific journal that takes the lead in covering disability research in all areas of health and society at the regional and international level.

      scite_
      0
      0
      0
      0
      Smart Citations
      0
      0
      0
      0
      Citing PublicationsSupportingMentioningContrasting
      View Citations

      See how this article has been cited at scite.ai

      scite shows how a scientific paper has been cited by providing the context of the citation, a classification describing whether it supports, mentions, or contrasts the cited claim, and a label indicating in which section the citation was made.

       
      • Record: found
      • Abstract: found
      • Article: found
      Is Open Access

      Enhanced Bald Eagle Search Optimizer with Transfer Learning-based Sign Language Recognition for Hearing-impaired Persons

      Published
      research-article
      Bookmark

            Abstract

            Sign language recognition and classification for hearing-impaired people is a vital application of computer vision (CV) and machine learning (ML) approaches. It contains developing structures that take sign language gestures carried out by individuals and transform them into textual or auditory output for transmission aspects. It is critical to realize that establishing a robust and correct sign language recognition and classification method is a difficult task because of several challenges like differences in signing styles, occlusions, lighting conditions, and individual variances in hand movements and shapes. Thus, it needs a group of CV approaches, ML systems, and a varied and representative database for training and testing. In this study, we propose an Enhanced Bald Eagle Search Optimizer with Transfer Learning Sign Language Recognition (EBESO-TLSLR) technique for hearing-impaired persons. The presented EBESO-TLSLR technique aims to offer effective communication among hearing-impaired persons and normal persons using deep learning models. In the EBESO-TLSLR technique, the SqueezeNet model is used for feature map generation. For recognition of sign language classes, the long short-term memory (LSTM) method can be used. Finally, the EBESO approach is exploited for the optimal hyperparameter election of the LSTM method. The simulation results of the EBESO-TLSLR method are validated on the sign language dataset. The simulation outcomes illustrate the superior results of the EBESO-TLSLR technique in terms of different measures.

            Main article text

            INTRODUCTION

            Communication among dumb and deaf people is done through visual and textual expressions. Gestural interaction is in the scope of secure and confidential interaction ( Rastgoo et al., 2020). Facial parts and hands are hugely powerful in expressing the opinions of humans in confidential interaction. Ordinary people must assume few syntactic meanings for expressions done by people of dumb and deaf communities ( Das et al., 2023). There exist different ways of expression or communication; however, the main mode of human interaction is speech, and if it can be slowed down, individuals need to utilize a tactile–kinesthetic mode of interaction in its place ( Mannan et al., 2022). As per the National Statistical Office survey report, the percentage of people with these disabilities in India has been 2.2% since December 2018. One of the highest adaptations for individuals with hearing and speech impairments is sign language. It is termed as a visual language ( Hameed et al., 2022). It has five fundamental variables: movement, hand shape, place, orientation, and elements like eyebrow movements and mouth shape. The research was conducted on voice generation with smart gloves that can offer a voice-to-sign language movement ( Elakkiya, 2021). Nevertheless, people who have no idea of sign language typically reject or undervalue public damage due to the lack of proper interactions between them ( Li et al., 2020). Therefore, this study devises a mechanism designed to remove the interaction gap and present all individuals with an equal and fair chance. It takes videos of the individual making hand gestures, passing, and processing them to the presented method that forecasts words ( Sharma and Kumar, 2021). The scheme then generated meaningful sentences of those words that can be transformed into the language chosen by the correspondent.

            Long ago, numerous exciting studies existed for the detection of dynamic hand gestures for sign languages ( Pandey et al., 2023). The detection remains an inspiring problem despite efforts made in the domain in the past few years. The necessity for understanding multi-modal data like movement and hand gestures in the event of American Sign Language (ASL), whereas it creates a problem vaguer, is acute ( Aly and Aly, 2020). Additionally, a massive amount of words in sign language with comparable gestures having a few instances for all words makes the issue harder. Occasionally, the same signs from various viewpoints or various signers have various appearances ( Lee et al., 2021).

            In this study, we propose an Enhanced Bald Eagle Search Optimizer with Transfer Learning Sign Language Recognition (EBESO-TLSLR) technique for hearing-impaired persons. The presented EBESO-TLSLR technique aims to offer effective communication among hearing-impaired persons and normal people using deep learning (DL) models. In the EBESO-TLSLR technique, the SqueezeNet model is used for feature map generation. For recognition of sign language classes, the long short-term memory (LSTM) method can be used. Finally, the EBESO method is exploited for the optimal hyperparameter election of the LSTM approach. The simulation results of the EBESO-TLSLR method are validated on the sign language dataset.

            RELATED STUDIES

            Alnfiai (2023) presents an SSODL-ASLR method, expanded as shark smell optimization with DL-related automated sign language recognition, for speaking- and hearing-impaired people. The presented method focuses on the classification and recognition of sign language presented by deaf and dumb persons. In the initial phase, Mask RCNN approach can be used for sign language detection. Then, the SSO system with soft margin SVM method is employed for classifying sign languages. In Latif et al. (2021), the authors proposed an AI-related Arabic Sign Language (ArSL) Translator. This system can seize the image of the sign language effectuated by an individual who is deaf and offers real-time translation of hand gestures. New structures are devised to be mined from the imageries to be input to the four methods: regression, RF, RT, and bagging classifier.

            In Latif et al. (2020), an ArSL detection structure with the design of the deep CNN (DCNN) was proposed. The goal is to aid persons with hearing difficulties to interact with normal persons. The presented structure identifies the sign of the Arabic alphabet grounded on real-time user inputs. In Bilgin and Mutludoğan (2019), the detection of sign language characters using a structure that has been trained via images of letters in ASL is intended. Capsule networks, which can be projected in the recent period, were utilized for testing and training processes and likened with LeNet which can be the first successful and presently with a method of DL. Hossain et al. (2020) devised a method to identify Bangla sign language (BSL) gestures with CNN. A large amount of openly available sign language data have been utilized to find BSL.

            In Selvanambi et al. (2023), different methods to forecast what a user can be irritated to convey over hand gestures with ASL. A chapter emphasizes the presented approach or tool that is leveraged for practices of ASL by regular persons to practice and learn sign language. Marzouk et al. (2022) intended an ASODCAE-SLR method, expanded as atom search optimization include deep convolution AE-enabled sign language detection, for hearing and speaking disabled persons. The presented approach targets to support the interaction of speaking and hearing disabled people through the SLR procedure. As well, the devised method uses CapsNet feature extractors to generate a set of feature vectors.

            MATERIALS AND METHODS

            In this study, we have aimed to develop an automated sign language recognition using the EBESO-TLSLR technique for hearing-impaired persons. The presented EBESO-TLSLR technique is intended to offer effective communication among hearing-impaired persons and normal persons using the DL models. In the EBESO-TLSLR technique, a three-stage process is derived namely SqueezeNet feature extraction, LSTM-based sign recognition, and EBESO-based parameter tuning. Figure 1 depicts the workflow of the EBESO-TLSLR approach.

            Figure 1:

            Workflow of the EBESO-TLSLR approach. Abbreviation: EBESO-TLSLR, Enhanced Bald Eagle Search Optimizer with Transfer Learning Sign Language Recognition.

            Feature extraction process

            In the EBESO-TLSLR technique, the SqueezeNet model is used for feature map generation. SqueezeNet model can be a DCNN, which has a small amount of parameters, compressed structure, and attained higher precision than ImageNet and AlexNet with a similar amount of parameters ( Thanki, 2023). The major advantages are easier to deploy on a cloud platform, can be used and customized on hardware with restricted memory, and has fewer communication channels for training. This original SqueezeNet consists of 14 layers namely 3 max-pooling layers, 8 fire layers, 2 conventional convolution layers, 1 softmax, and a global avg-pooling layer. In this model, the convolution layer is exploited for extracting features from the input color retinal imageries. This sequential method comprises of a convolution layer, ith standalone Conv layer, and eight fire components. The amount of filters per fire component is increased progressively. The max pooling having the stride of 2 is implemented afterwards layers Conv layer 2, Conv layer 1, fire 3, and fire 7. The fire model is a squeeze Conv layer which comprises a 1×1 size fed into the expanding layer that has a mix of 1×1 and 3×3 convolutional filters. There are three tunable hyperparameters namely s1×1, e1×1, and e3×3. The e1×1 and e3×3 are expanded layers with filter sizes of 1×1 and 3×3, correspondingly, while s1×1 is a squeeze layer with a filter size of 1×1. Here, a fire module with hyperparameter s1×1 is used to limit the amount of input channels to the 3×3 size. After extracting features from the input retinal images, the flattened layer is used to convert features into a 1D0 layer.

            Recognition process using the optimal LSTM model

            To recognize various sign language classes, the LSTM approach was used. LSTM is a special kind of RNN used in the domain of DL. The main factor of the LSTM method is to present a storage unit for cyclic data communication, which records every past data up to the present moment ( Wang et al., 2022). Thus, in comparison to the short-term memory of conventional RNN, LSTM with long-term memory abilities: a gating model (including forget, input, and output gates) with a value between [0, 1] can be used for controlling the communication path of the internal data in the method.

            Where tanh denotes the activation function. c t− 1 and h t− 1 denote the output of the memory unit and hidden layer (HL) at t–1 time. c t , x t , it, f ta t h t , and o t characterize storage unit, input, forget, candidate state, input gate, the output of HL, and output gate at t time, correspondingly. δ shows the logistic sigmoid function. Figure 2 represents the infrastructure of LSTM.

            Figure 2:

            LSTM architecture. Abbreviation: LSTM, long short-term memory.

            The weights between HL and the output, forget, memory unit, and input gates, are represented as W ho , W hf , W hc , and W hi . W xo , W xf, W xi , and W xc correspondingly indicate the weight connection of the output gate, forget gate, input gate, and storage unit. ⊙ shows the product of vector components. b f , b i , and b o indicate the bias. The LSTM loop structure controls the data flow by controlling the degree of closing and opening of forget, input, and output gates. Steps 1 to 6 are given as follows:

            Step1: The forget gate f t takes input x t of the existing layer and the output h t −1 of HL at the prior moment as input, and the resultant output of forget gate can be multiplied with c t −1 for controlling what amount of data should be forgotten in the internal state c t −1 at the prior moment.

            (1) ft=σ(Wxfxt+Whfht1+bf)

            Step2: The input gate stores the present input data and the resultant output is utilized as the data to be updated.

            (2) it=σ(Wxixt+Whiht1+bi)

            Step3: The output gate 0 fj controls internal state c fj at the present moment, for controlling what amount of data should be outputted to the external state h fj .

            (3) ot=σ(Wxoxt+Whoht1+bo)

            Step4: The output gate o t is multiplied by the state of the storage unit processed by the tanh function. The output h fj of HL is evaluated using Eq. (4):

            (4) ht=ottanh(ct)

            Step5: The memory unit c fj records the prior data up to the present moment that is evaluated as follows:

            (5) ct=ftct1+itat

            Step6: The candidate state can be formulated as follows:

            (6) at=tanh(Wxcxt+Whcht1+bc)

            Finally, the EBESO algorithm is used for optimal hyperparameter tuning. Bald eagles often feast on protein-rich food. It targets fish ( Alsaidan et al., 2022). Bald eagle chooses a space between the deep water and the land surface. The search step begins after the hunter attains the specified region. As well, bald eagles have outstanding eyesight that allows them to mark the fish inside the water from a higher distance in the air. Once the hunter specifies its targets, it begins to slowly descend to grab the fish and to catch its victim. The mathematical representation and equation of the BESO algorithm label the three phases of hunting by the bald eagle.

            The first stage (selection)

            The bald eagle defines the region where it could capture the fish. This phase is shown in mathematical method as follows:

            (7) Pnew,i=Pbest+α×r(PmeanPi)

            In Equation (7), α denotes the location change control parameter ϵ [1.5, 2]. r ϵ [ O, 1] at random. Another exploration region is chosen that is nearby to the formerly chosen one. P best is an existing exploration space defined by the bald eagles. It can be selected based on the previous location defined. P mean demonstrates that the eagle benefits from the data of the prior position. This initial phase significantly improves the candidate solution based on the mean location and the better location.

            The second stage (search)

            In this stage, the hunter searches for the victim. The search can be performed in the previously defined exploration region. Next, the eagle travels in different directions in the spiral region so that the searching process can be speeded up:

            (8) Pi,new=Pi+y(i)×(PiPi+1)+x(i)×(PiPmean)x(i)=xr(i)max(|xr|),y(i)=yr(i)max(|xr|)xr(i)=r(i)×sin(θ(i)),yr(i)=r(i)×cos(θ(i))θ(i)=a×π×randr(i)=θ(i)+R × rand

            In Eq. (8), a denotes the random variables ϵ [5, 10]. ‘a’ and ‘R’ parameters control the variation in spiral shape. Likewise, the variable R ϵ [0.5, 2] controlled the determination of the exploration cycle number.

            In this stage, the location moves toward the center. Once the ‘a’ and ‘R’ change, the BESO algorithm diversifies to seek more accurate solutions and to prevent trapping in the local solution.

            The third stage (swooping)

            In the hunting process, swooping is the third and last phase, where the eagles move from the better position toward the targeted prey.

            (9) Pinew=rand×Pbest+x1(i)×(Pic1×Pmean)+y1(i)×(Pic2×Pbest)x1(i)=xr(i)max(|xr|),y1(i)=yr(i)max(|yr|)xr(i)=r(i)×sinh[θ(i)], yr(i)=r(i)×cosh[θ(i)]θ(i)=a×π×randr(i)=θ(i)

            From the expression, the two random parameters, c1 and c2, change between 1 and 2 to further intensify the motion of the hunting eagle toward the better position. The mean solution might push the presented method to diverse or intense toward the optimum solution.

            Finally, the optimum solution is accomplished by a minimal amount of iterations. All the stages are affected by intensification and diversification factors. They are significant for the continual upgrade to the candidate solution and lastly, to obtain the optimum one.

            In the EBESO algorithm, the levy function can be included for improving the outcome of the presented BES method. This can be performed by altering the search agent-utilizing factor “LF”. This factor is arithmetically evaluated using the following expression:

            (10) LF(γ)=0.01×u×σ|v|1γ,σ=(Γ(1+γ)×sin(πγ2)Γ(1+γ2)×γ×2(γ12))1γ

            Where v and u denote random values within [0, 1].

            In the third step, the Levy function is inserted, the Swoop step. The new population can be evaluated using Equation (11).

            (11) Pi,new={Pbest+P1×CF×Pi,newr1<0.5Pi,new+P×CF×stepsize2lr10.5

            Where r 1 denotes a random integer within [0, 1]. P1 and CF are constants. The stepsize l is evaluated using Equation (12):

            (12) stepsize2l=LF*(PbestLF*Pi,new)

            RESULTS AND DISCUSSION

            The sign language detection and classification outcomes of the EBESO-TLSLR method are demonstrated.

            In Table 1 and Figure 3, the overall sign language recognition results of the EBESO-TLSLR technique are illustrated. The experimental values suggest that the EBESO-TLSLR technique reaches improved recognition rates. For instance, with 70% of TRP, the EBESO-TLSLR technique obtains average accu y , prec n , reca l , and F score of 99.42, 99.30, 99.37, and 99.39%, respectively. On the other hand, with 30% of TSP, the EBESO-TLSLR method attains average accu y , prec n , reca l , and F score of 98.46, 98.55, 98.54, and 98.41%, correspondingly.

            Figure 3:

            Average outcome of the EBESO-TLSLR approach on 70:30 of TRP/TSP. Abbreviation: EBESO-TLSLR, Enhanced Bald Eagle Search Optimizer with Transfer Learning Sign Language Recognition.

            Table 1:

            Sign language recognition outcome of the EBESO-TLSLR system on 70:30 of TRP/TSP.

            Training phase (70%)
            Sign Accu y Prec n Reca l F score Sign Accu y Prec n Reca l F score
             A99.0399.2699.7299.15 P99.6799.2299.6899.28
             B99.5399.6799.2699.40 Q99.4499.0599.3199.20
             C99.6099.5999.7299.51 R99.4199.2599.6499.51
             D99.3199.4599.0099.65 S99.7199.3799.0999.09
             E99.5699.0899.7499.23 T99.1499.5799.0799.23
             F99.3199.0599.2799.30 U99.3499.4599.2499.77
             G99.0099.0699.5099.53 V99.6599.4399.2299.62
             H99.2899.1099.1099.79 W99.5799.3099.7199.07
             I99.7199.2699.6299.79 X99.3899.7399.7999.40
             J99.5799.2499.0399.10 Y99.7499.6899.0799.54
             K99.2499.1399.1699.13 Z99.2899.7299.0499.79
             L99.5499.7299.3699.55 Space99.3099.5599.0899.59
             M99.4299.4899.0199.28 Nothing99.5399.2599.4899.64
             N99.6599.0699.2799.06 Delete99.4499.3499.3099.28
             O99.5199.3199.7399.42 Average99.4299.3099.3799.39
            Testing phase (30%)
            Sign Accu y Prec n Reca l F score Sign Accu y Prec n Reca l F score
             A98.4398.4998.4998.76 P98.5498.7998.1998.68
             B98.6698.4698.6198.29 Q98.7298.1898.1398.39
             C98.1498.3898.7198.51 R98.8598.3598.0898.51
             D98.5898.6898.5898.24 S98.7098.3298.9198.31
             E98.6298.5698.1398.48 T98.5198.4598.5298.65
             F98.0798.2698.0498.52 U98.9698.5698.8598.68
             G98.2998.5998.2198.15 V98.0198.8098.6098.65
             H98.2598.4498.7998.54 W98.5798.6198.5398.40
             I98.4698.1598.7198.87 X98.6098.0398.1298.59
             J98.1498.6198.4298.66 Y98.6398.8898.1898.09
             K98.7598.9798.0898.05 Z98.4398.8098.7998.67
             L98.9398.8798.9498.79 Space99.0098.4498.0298.20
             M98.5398.9098.8298.05 Nothing98.5698.5198.8998.61
             N98.6098.7098.9598.14 Delete98.7598.8198.1998.60
             O98.4498.2198.5798.17 Average98.4698.5598.5498.41

            Abbreviation: EBESO-TLSLR, Enhanced Bald Eagle Search Optimizer with Transfer Learning Sign Language Recognition.

            Figure 4 inspects the accuracy of the EBESO-TLSLR method in the training and validation of the test database. The result specified that the EBESO-TLSLR technique reach higher accuracy values over greater epochs. Also, the higher validation accuracy over training accuracy portrayed that the EBESO-TLSLR technique learns productively on the test database.

            Figure 4:

            Accuracy curve of the EBESO-TLSLR approach. Abbreviation: EBESO-TLSLR, Enhanced Bald Eagle Search Optimizer with Transfer Learning Sign Language Recognition.

            The loss analysis of the EBESO-TLSLR technique in the training and validation is given on the test database in Figure 5. The result points out that the EBESO-TLSLR method reaches closer values of training and validation loss. The EBESO-TLSLR method learns productively on a test database.

            Figure 5:

            Loss curve of the EBESO-TLSLR approach. Abbreviation: EBESO-TLSLR, Enhanced Bald Eagle Search Optimizer with Transfer Learning Sign Language Recognition.

            In Table 2, a detailed comparison study of the EBESO-TLSLR method is reported with recent approaches ( Alrowais et al., 2022).

            Table 2:

            Comparative outcome of the EBESO-TLSLR approach with other methods.

            MethodsAccuracy (%)Computation time (minutes)
            KNN algorithm96.2516.54
            SVM model98.1014.34
            ANN model98.1115.41
            CNN model99.0911.21
            ODTL-SLRC99.0306.44
            EBESO-TLSLR99.4203.01

            Abbreviation: EBESO-TLSLR, Enhanced Bald Eagle Search Optimizer with Transfer Learning Sign Language Recognition.

            Figure 6 illustrates the accu y investigation of the EBESO-TLSLR method with existing techniques. The result demonstrates that the EBESO-TLSLR technique reaches improved performance. Based on accu y , the EBESO-TLSLR technique obtains accu y of 99.42% while the existing KNN, SVM, ANN, CNN, and ODTL-SLRC techniques attained accu y of 96.25, 98.10, 98.11, 99.09, and 99.03%, correspondingly.

            Figure 6:

            Accu y outcome of the EBESO-TLSLR approach with other approaches. Abbreviation: EBESO-TLSLR, Enhanced Bald Eagle Search Optimizer with Transfer Learning Sign Language Recognition.

            Figure 7 illustrates the CT investigation of the EBESO-TLSLR method with existing approaches. The result highlighted that the EBESO-TLSLR method reaches improved performance. Based on CT, the EBESO-TLSLR approach gains CT of 3.01 m while the existing KNN, SVM, ANN, CNN, and ODTL-SLRC algorithms attained CT of 16.54, 14.34, 15.41, 11.21, and 6.44 m correspondingly. These outcomes reassured the better performance of the EBESO-TLSLR method over other approaches.

            Figure 7:

            CT outcome of the EBESO-TLSLR approach with other methods. Abbreviation: EBESO-TLSLR, Enhanced Bald Eagle Search Optimizer with Transfer Learning Sign Language Recognition.

            CONCLUSION

            In this study, we have aimed to develop automated sign language recognition using the EBESO-TLSLR technique for hearing-impaired persons. The presented EBESO-TLSLR technique is intended to offer effective communication among hearing-impaired persons and normal persons using the DL models. In the EBESO-TLSLR technique, the SqueezeNet model is used for feature map generation. For recognition of sign language classes, the LSTM method is used. Finally, the EBESO method is exploited for the optimal hyperparameter election of the LSTM model. The simulation results of the EBESO-TLSLR approach are validated on the sign language dataset. The simulation outcomes illustrate the superior results of the EBESO-TLSLR technique with a maximum accuracy of 96.25%. In future, the detection rate of the EBESO-TLSLR method can be boosted by the feature fusion process.

            REFERENCES

            1. Alnfiai MM. 2023. Deep learning-based sign language recognition for hearing and speaking impaired people. Intell. Autom. Soft Comput. Vol. 36(2):1653–1669

            2. Alrowais F, Alotaibi SS, Dhahbi S, Marzouk R, Mohamed A, Hilal AM. 2022. Sign language recognition and classification model to enhance quality of disabled people. Comput. Mater. Continua. Vol. 73(2):3419–3432

            3. Alsaidan I, Shaheen MA, Hasanien HM, Alaraj M, Alnafisah AS. 2022. A PEMFC model optimization using the enhanced bald eagle algorithm. Ain. Shams. Eng. J. Vol. 13(6):101749

            4. Aly S, Aly W. 2020. DeepArSLR: a novel signer-independent deep learning framework for isolated Arabic sign language gestures recognition. IEEE Access. Vol. 8:83199–83212

            5. Bilgin M, Mutludoğan K. 2019. American sign language character recognition with capsule networks2019 3rd International Symposium on Multidisciplinary Studies and Innovative Technologies (ISMSIT); IEEE. Ankara, Turkey. 11-13 October 2019; p. 1–6

            6. Das S, Imtiaz MS, Neom NH, Siddique N, Wang H. 2023. A hybrid approach for Bangla sign language recognition using deep transfer learning model with random forest classifier. Expert Syst. Appl. Vol. 213:118914

            7. Elakkiya R. 2021. RETRACTED ARTICLE: Machine learning based sign language recognition: a review and its research frontier. J. Ambient Intell. Hum. Comput. Vol. 12(7):7205–7224

            8. Hameed H, Usman M, Khan MZ, Hussain A, Abbas H, Imran MA, et al.. 2022. Privacy-preserving British sign language recognition using deep learning2022 44th Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC); IEEE. Glasgow, Scotland, UK. July 11-15; p. 4316–4319

            9. Hossain S, Sarma D, Mittra T, Alam MN, Saha I, Johora FT. 2020. Bengali hand sign gestures recognition using convolutional neural network2020 Second International Conference on Inventive Research in Computing Applications (ICIRCA); IEEE. Coimbatore, India. July 15-17; p. 636–641

            10. Latif G, Mohammad N, AlKhalaf R, AlKhalaf R, Alghazo J, Khan M. 2020. An automatic Arabic sign language recognition system based on deep CNN: an assistive system for the deaf and hard of hearing. Int. J. Comput. Digital Syst. Vol. 9(4):715–724

            11. Latif G, Alghazo J, Mohammad N, Alghazo R. 2021. Communicating with the deaf and hard of hearing through automatic Arabic sign language translator. J. Phys. Vol. 1962(1):012055

            12. Lee CK, Ng KK, Chen CH, Lau HC, Chung SY, Tsoi T. 2021. American sign language recognition and training method with recurrent neural network. Expert Syst. Appl. Vol. 167:114403

            13. Li D, Rodriguez C, Yu X, Li H. 2020. Word-level deep sign language recognition from video: a new large-scale dataset and methods comparisonProceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision; Snowmass, CO, USA. March 1-5; p. 1459–1469

            14. Mannan A, Abbasi A, Javed AR, Ahsan A, Gadekallu TR, Xin Q. 2022. Hypertuned deep convolutional neural network for sign language recognition. Comput. Intell. Neurosci. Vol. 2022:1450822

            15. Marzouk R, Alrowais F, Al-Wesabi FN, Hilal AM. 2022. Atom search optimization with deep learning enabled Arabic sign language recognition for speaking and earing disability persons. Healthcare. Vol. 10(9):1606

            16. Pandey A, Chauhan A, Gupta A. 2023. Voice based sign language detection for dumb people communication using machine learning. J. Pharm. Negat. Results. 22–30

            17. Rastgoo R, Kiani K, Escalera S. 2020. Hand sign language recognition using multi-view hand skeleton. Expert Syst. Appl. Vol. 150:113336

            18. Selvanambi R, Karuppiah M, Islabudeen M. 2023. Mobile application-based sign language detector for deaf peopleDesigning and Developing Innovative Mobile Applications. IGI Global. p. 329–350

            19. Sharma S, Kumar K. 2021. ASL-3DCNN: American sign language recognition technique using 3-D convolutional neural networks. Multimedia Tools Appl. Vol. 80(17):26319–26331

            20. Thanki R. 2023. A deep neural network and machine learning approach for retinal fundus image classification. Healthcare Anal. Vol. 3:100140

            21. Wang J, Zhang D, Zhou Y. 2022. Ensemble deep learning for automated classification of power quality disturbances signals. Electr. Power Syst. Res. Vol. 213:108695

            Author and article information

            Journal
            jdr
            Journal of Disability Research
            King Salman Centre for Disability Research (Riyadh, Saudi Arabia )
            20 October 2023
            : 2
            : 3
            : 86-93
            Affiliations
            [1 ] Department of Computer Science, College of Science and Arts at Mahayil, King Khalid University, Abha, Saudi Arabia ( https://ror.org/052kwzs30)
            [2 ] King Salman Center for Disability Research, Riyadh, Saudi Arabia ( https://ror.org/01ht2b307)
            [3 ] Department of Information Systems, College of Business Administration in Hawtat Bani Tamim, Prince Sattam Bin Abdulaziz University, Al Hulwah, Saudi Arabia ( https://ror.org/04jt46d36)
            [4 ] Department of Computer and Self Development, Preparatory Year Deanship, Prince Sattam Bin Abdulaziz University, AlKharj, Saudi Arabia ( https://ror.org/04jt46d36)
            Author notes
            Author information
            https://orcid.org/0000-0002-5128-7684
            Article
            10.57197/JDR-2023-0039
            dd3fb1e4-0341-4b1d-ac07-73c58c51f8ef
            Copyright © 2023 The Authors.

            This is an open access article distributed under the terms of the Creative Commons Attribution License (CC BY) 4.0, which permits unrestricted use, distribution and reproduction in any medium, provided the original author and source are credited.

            History
            : 23 May 2023
            : 22 September 2023
            : 25 September 2023
            Page count
            Figures: 7, Tables: 2, References: 21, Pages: 8
            Funding
            Funded by: funder-id http://dx.doi.org/10.13039/501100019345, King Salman Center for Disability Research;
            Categories

            Computer science
            sign language recognition,hearing-impaired people,transfer learning,bald eagle search optimizer,deep learning

            Comments

            Comment on this article