Processing math: 100%
364
views
0
recommends
+1 Recommend
1 collections
    0
    shares

      King Salman Center for Disability Research is pleased to invite you to submit your scientific research to the Journal of Disability Research. JDR contributes to the Center's strategy to maximize the impact of the field, by supporting and publishing scientific research on disability and related issues, which positively affect the level of services, rehabilitation, and care for individuals with disabilities.
      JDR is an Open Access scientific journal that takes the lead in covering disability research in all areas of health and society at the regional and international level.

      scite_
      0
      0
      0
      0
      Smart Citations
      0
      0
      0
      0
      Citing PublicationsSupportingMentioningContrasting
      View Citations

      See how this article has been cited at scite.ai

      scite shows how a scientific paper has been cited by providing the context of the citation, a classification describing whether it supports, mentions, or contrasts the cited claim, and a label indicating in which section the citation was made.

       
      • Record: found
      • Abstract: found
      • Article: found
      Is Open Access

      Emotion Analysis Using Improved Cat Swarm Optimization with Machine Learning for Speech-impaired People

      Published
      research-article
      Bookmark

            Abstract

            Emotion analysis is an effective method for improving interaction and understanding for speech-impaired people. We can provide the best interaction and support emotional well-being by analyzing the emotion conveyed through speech. Using deep learning (DL) or machine learning algorithms for training an emotion classification method. This might include training classifiers namely random forests, deep neural networks, or support vector machines. It is noteworthy that emotion analysis could be effective; however, it is crucial to consider individual differences and context while interpreting emotion. Furthermore, ensuring data protection and privacy and obtaining consent are vital features to consider while working with sensitive speech data. Therefore, this study presents an emotion analysis approach using improved cat swarm optimization with machine learning (EA-ICSOML) technique. The EA-ICSOML technique applies the concepts of computer vision and DL to identify various types of emotions. For feature vector generation, the ShuffleNet model is used in this work. To adjust the hyperparameters compared to the ShuffleNet system, the ICSO algorithm is used. Finally, the recognition and classification of emotions are performed using the Transient Chaotic Neural Network approach. The performance validation of the EA-ICSOML technique is validated on facial emotion databases. The simulation result inferred the improved emotion recognition results of the EA-ICSOML approach compared to other recent models in terms of different evaluation measures.

            Main article text

            INTRODUCTION

            Emotion can be described as a mental state linked with the nervous system; that is, what an individual feels inside as the effects of the environment. The emotions of a person can be identified in many ways ( Nandwani and Verma, 2021). Some can be examined by body gestures, tonal properties, and facial expressions. The classification or computing of emotions from facial or speech expressions formed a significant part of human information processing ( Ahire and Borse, 2022). In the intellectual learning environment, emotion detection of learners’ images during class hours using computer and deep learning (DL) methods enables prompt monitoring of the emotional and psychological states of learners. Emotion detection using facial expression images needs high-quality cameras for capturing facial images, resulting in high application costs ( Zad et al., 2021). Hence, the speech-related human emotion detection approach has slowly become the principal approach to studying human–computer emotion detection. In expression and communication, the speech of humans does not have semantic data but implies rich data like the emotions of speakers ( Sailunaz et al., 2018). Thus, the study of emotion detection related to image and human speech using computer and intellectual methods of DL is of great significance ( Vasantharajan et al., 2022). Automated emotion detection is a significant research study that solves two subjects: artificial intelligence and human emotion recognition. The emotional state of a person can be gained from verbal and nonverbal data captured using the different sensors, for instance from facial changes, physiological signals, and tone of voice ( De and Mishra, 2022).

            Face changes during an interaction are the initial signs that send the emotional status which is why many authors are very involved in this modality ( Kumar et al., 2022). Abstracting attributes from one face to another becomes a sensitive and tough task to have a superior classification. The automated FER is mostly studied by authors compared to other modalities of statistics, but it is not a simple task as all persons present their emotions in different ways. Various difficulties and encounters are presented in a zone that one must not be negligent like the gender, background, disparity of head poses, age, and luminosity, in addition to the issue of obstruction caused by skin illness, sunglasses, scarf, etc. ( Vijayvergia and Kumar, 2021). Numerous conventional approaches exist and are utilized for the abstraction of facial features like regular and texture features for instance local directional paradigms, local binary patterns, Gabor wavelet, and facial action units FAC ( Cui et al., 2022). Currently, DL is a very successful and effective method thanks to the outcome gained with its structure that allowed the automatic extraction of classification and features like RNN and CNN; this provoked authors to start utilizing this method for detecting human emotions ( Riza and Charibaldi, 2021). Various studies have been done by authors on the growth of a deep neural network structure that produces very reasonable outcomes in this area.

            This study presents an emotion analysis approach using improved cat swarm optimization with machine learning (EA-ICSOML) technique. The EA-ICSOML technique applies the concepts of computer vision (CV) and DL to identify various types of emotions. For feature vector generation, the ShuffleNet model is used in this work. To adjust the hyperparameters related to the ShuffleNet approach, the ICSO algorithm is used. Finally, the recognition and classification of emotions are performed using the Transient Chaotic Neural Network (TCNN) approach. The performance validation of the EA-ICSOML technique is validated on facial emotion databases.

            RELATED STUDIES

            Modran et al. (2023) focused on forecasting and finding whether music has healing benefits. An ML method can be industrialized, utilizing a multi-class NN for categorizing emotions into four categories and forecasting the output. The NN has three layers: an output layer, an input layer that has many attributes, and a deeply connected hidden layer. To assess the estimator, K-fold cross validation has been utilized. Catania and Garzotto (2023) presented Emoty, a speech-related conversational agent devised for individuals with NDD for training emotional transmission skills. A characteristic of this agent is the expressive expression power of voice. Emoty engages users in minor chats that are requested to reprise sentences and direct some emotions with a suitable vocal tone.

            Garcia-Garcia et al. (2022) presented a new software application as a somber game for teaching progenies with ASD to express and identify emotions. The mechanism incorporated cutting-edge technology for supporting new transmission mechanisms related to tangible user interface and emotion detection from facial expressions. Likewise, children communicate with mechanisms by grabbing substances with their hands and using their faces. Aoki et al. (2022) explored the impacts of speech balloon shapes on sender–receiver agreement concerning the emotionality of text messages. Depending on these outcomes, the author formed a system that generated speech balloons matching linear emotional arousal intensity by ACGAN.

            In Hou (2022), a DL-related human emotion detection framework was devised to assess the probability of digital representation, detection, and prediction of feelings. The presented method analyzed the effect of emotional methods on multimodal detection. The study presents developing works that utilize present approaches like CNN for human emotion detection related to video, language, image, and sound physiological signals. While the discoveries gained are not a province, the evidence gathered specifies that DL can be adequate to categorize facial emotion. Mridha et al. (2022) developed an emotion detection mechanism related to landmarks. This study utilizes a CNN for recognizing facial feelings to understand the state of mind of impaired persons. Only wireless button-related transmission, as per the presented approach, can alert the caregiver that a handicapped individual needs anything.

            THE PROPOSED MODEL

            In this study, we have presented a novel EA-ICSOML technique for the emotion analysis process. The EA-ICSOML technique applies the CV and DL concepts to identify various types of emotions. The working process of the EA-ICSOML technique comprises ShuffleNet feature extraction, ICSO-based hyperparameter optimization, and TCNN-based classification. Figure 1 exemplifies the overall process of the EA-ICSOML algorithm.

            Figure 1:

            Overall process of the EA-ICSOML approach. Abbreviation: EA-ICSOML, emotion analysis approach using improved cat swarm optimization with machine learning.

            Feature extraction: SqueezeNet

            For feature vector generation, the ShuffleNet model is used in this work. Xiangyu Zhang et al. from Megvii ShuffleNet 2018 proposed a neural network structure devised for potential DL on mobile devices ( Ullah et al., 2022). It depends on pointwise group convolution that enables effective computation of mapping features by diminishing the operations needed for convolutional. The network structure embraces a shuffling procedure executed after pointwise group convolution that aids in the rise of the mixing of mapping feature across various channels. This enhances the capability of networks to increase their accuracy and capture complicated features. ShuffleNet has accomplished remarkable performance on benchmark datasets while requiring a low memory footprint and being computationally efficient.

            Hyperparameter tuning: ICSO algorithm

            For adjusting the hyperparameters related to the ShuffleNet approach, the ICSO algorithm is used. The CSO algorithm stimulated through the natural behaviors of felines to resolve complicated optimization problems ( Wang and Han, 2023). The model considers the cat location as the potential solution of the problems to be augmented, attains the estimated potential area of the solution through the prior data of problems, viz., potential area, abstracts the feline actions as a searching pattern, and searches for the optimum solution of problems that improved from the potential area with these patterns. The searching mode of CSO mainly comprises tracing and seeking components; where the amount of populations carrying out two modes is defined as the mixture ratio.

            During the seeking mode, many crucial components are determined like counts of dimension to change (CDC), seeking memory pool (SMP), self-position considering (SPC), and seeking the range of the selected dimension (SRD). SMP is the number of random locations that are produced by all the cats; SRD characterizes the changing range of every cat in all the dimensions, and usually takes values within [0,1]; CDC shows the number of dimensions that change in all the cats, as well as take values within [0,1]; SPC defines whether the existing location of cat is a candidate location for the following iteration. These elements are Boolean values (0 or 1). Its location updating equation can be given as follows:

            (1) Lnew=Lold+(Sr×r1×Lold)

            In Equation (1), L new shows the updated location of the cat; L o1d indicates the original location of the cat; Sr represents the SRD of a cat with values of 0 and 1; and r 1 denotes the arbitrarily created values in 0 and 1. Consider the SPC value of the present locality; if the SPC values are 1, retain the existing place and copy SMP-I copy of the existing place as the memory pool; on the other hand, copy SMP as the memory pool. Figure 2 presents the flowchart of the CSO algorithm.

            Figure 2:

            Flowchart of the CSO algorithm.

            • For every cat in the memory pool, initially, randomly define the individual dimension where the variation takes place as per the CDC values; consequently, updating the position data.

            • Evaluate the fitness value for every individual.

            • Choose the next candidate point position through Roulette probability selection.

            The tracking module corresponding to the local search of optimization method is the same as the location updating of the PSO, mostly by changing the individual velocity in all the dimensions to accomplish the location update as follows:

            (2) {Vk,d=Vk,d+c×r2×(Lbest,dLk,d)Lnewk,d=Loldk,d+Vk,d

            In Equation (2), r 2 indicates the randomly generated value, takes value from 0 to 1; L best,d represents the location of better individuals in the existing iteration in d dimension, V k,d indicates the velocity of kth individuals at d dimension, Lnewk,d denotes the location of the updated individuals at d dimension, L k,d shows the location of the kth individuals at d dimension; c represents the constant that is fixed to 2.05; Loldk,d shows the location beforehand the update.

            The ICSO system assesses a fitness function (FF) to gain better classifier results. It defines a positive integer to exemplify the best result of candidate performances. During this case, the minimized classifier error rate was assumed to be FF, as expressed in Equation (3).

            (3) Fitness(xi)=ClassifierErrorRate(xi)=No.ofmisclassifiedinstancesTotalno.ofinstances×100

            Emotion detection: TCNN model

            Finally, the recognition and classification of emotions are performed using the TCNN model. The self-feedback term with the simulated annealing model was additional based on the Hopfield neural network (HNN), named as transient chaotic neural network (TCNN) and can be given as follows ( Chen et al., 2022):

            (4) yi(t+1)=kyi(t)+α[Nj=1,jiwijxj(t)+Ii][zj(t)g(xi(t)I0)]xi(t)=S1(yi(t))=1/[1+exp(yi(t)/ε)]zj(t+1)=(1β)zi(t)

            In Equation (4), x i ( t) denotes the output of neurons, ε shows the steepness parameter of activation function ( ε > 0), y i ( t) shows the internal state of neurons, α represents a positive proportion parameter, shows the annealing attenuation factor of z i ( t), z i ( t) refers to the self-feedback connection weight, k indicates the damping factor of the neural diaphragm (0 ≤ k ≤ 1), αε[0, ∞) denotes the effect of energy function on chaotic dynamics, and I 0 denotes the positive parameter.

            A positive Lyapunov exponent shows that the method has chaotic features and the stronger the degree of chaos, the larger the Lyapunov exponent and is determined by the subsequent expression:

            (5) λ=limn1nn1i=0log|dy(t+1)dy(t)|

            (6) dy(t+1)dy(t)=kz(t)dx(t)dy(t)=kz(t)dS1(y(t))dy(t)

            (7) dS1(y(t))dy(t)=1εS1(y(t))(1S(y(t)))

            Set parameters: k = 1, β = 0.02, I 0 = 0.65, z(0) = 0.8, α = 0.07, and ε = 0.05. In an earlier phase of the development of the TCNN approach, z i ( t) proceeds a large primary value, and this method was chaotic. Meanwhile, z i ( t) continuously degenerates with time and still, it becomes 0; this method ensures inverse bifurcation transition and decays to HNN with gradient convergence.

            The TCNN model is used for mapping the main function of problems as to the energy function of networks and later changing the dynamics of networks as the objective function. Once the network converges toward the stable point, the neuron output becomes the suboptimal or optimal solution to the problem:

            (8) dyidt=Exi

            Kwok and Smith developed a modified energy function using the following equation:

            (9) E(t)=EHop+H=12Ni=1,ijNj=1,jiwijxi(t)xj(t)Ni=1Iixi(t)+1τiNi=1xi(t)0f1(ξ)dξ+H(xj,wij'Ii)

            For i, j = 1,2,…, N, N refers to the neuron counts, H denotes the energy value of the self-feedback term, and its selective procedure defines the variation feature of chaotic dynamics. X j ( t) indicates the output of the ith neurons at t time, I i shows the threshold of ith neurons, W ij represents the weight connected between neuron ith and jth neurons, f −1(·) shows the inverse function of the activation function, τ i signifies the time constant of the ith neuron, and H denotes the further energy term.

            RESULTS AND DISCUSSION

            In this section, the emotion recognition results of the presented approach are tested on the CK+ Database. The suggested technique is put under simulation by employing the Python 3.6.5 tool on PC i5-8600k, 250GB SSD, GeForce 1050Ti 4GB, 16GB RAM, and 1TB HDD. The setups of the parameters are as follows: learning rate: 0.01, activation: ReLU, epoch count: 50, dropout: 0.5, and size of the batch: 5.

            In Table 1, the overall emotion detection outcomes of the EA-ICSOML technique are demonstrated. Figure 3 represents the results of the EA-ICSOML technique on 80% of the TRP. The results indicate that the EA-ICSOML technique recognizes all types of emotions. In the anger class, the EA-ICSOML technique attains accu y , prec n , reca l , F score , and AUC score of 98.93%, 97.30%, 94.74%, 96%, and 97.16%, respectively. Meanwhile, in the contempt class, the EA-ICSOML method reaches accu y , prec n , reca l , F score , and AUC score of 98.93%, 96.34%, 96.34%, 96.34%, and 97.86%, correspondingly. Eventually, in the fear class, the EA-ICSOML approach accomplishes accu y , prec n , reca l , F score , and AUC score of 98.93%, 96.30%, 96.30%, 96.30%, and 97.83%, respectively. Concurrently, in the happy class, the EA-ICSOML system achieves accu y , prec n , reca l , F score , and AUC score of 99.82%, 100%, 98.73%, 99.36%, and 99.37%, correspondingly. Finally, in the sadness class, the EA-ICSOML method attains accu y , prec n , reca l , F score , and AUC score of 98.75%, 96.15%, 94.94%, 95.54%, and 97.16%, correspondingly.

            Table 1:

            Emotion detection outcome of the EA-ICSOML algorithm on 80:20 of TRP/TSP.

            Class Accu y Prec n Reca l F score AUC score
            Training phase (80%)
             Anger98.9397.3094.7496.0097.16
             Contempt98.9396.3496.3496.3497.86
             Fear98.9396.3096.3096.3097.83
             Disgust99.6498.7598.7598.7599.27
             Happy99.82100.0098.7399.3699.37
             Surprise98.9394.2598.8096.4798.87
             Sadness98.7596.1594.9495.5497.16
             Average99.1397.0196.9496.9798.22
            Testing phase (20%)
             Anger99.29100.0095.8397.8797.92
             Contempt99.29100.0094.4497.1497.22
             Fear100.00100.00100.00100.00100.00
             Disgust100.00100.00100.00100.00100.00
             Happy100.00100.00100.00100.00100.00
             Surprise99.2994.44100.0097.1499.59
             Sadness99.2995.45100.0097.6799.58
             Average99.5998.5698.6198.5599.19

            Abbreviation: EA-ICSOML, emotion analysis approach using improved cat swarm optimization with machine learning.

            Figure 3:

            Emotion detection outcome of the EA-ICSOML algorithm on 80% of TRP. Abbreviation: EA-ICSOML, emotion analysis approach using improved cat swarm optimization with machine learning.

            Figure 4 signifies the outcomes of the EA-ICSOML method on 20% of the TSP. The outcomes inferred that the EA-ICSOML technique recognizes all types of emotions. In the anger class, the EA-ICSOML methodology gains accu y , prec n , reca l , F score , and AUC score of 99.29%, 100%, 95.83%, 97.87%, and 97.92%, correspondingly. In the meantime, in the contempt class, the EA-ICSOML system achieves accu y , prec n , reca l , F score , and AUC score of 99.29%, 100%, 94.44%, 97.14%, and 97.22%, correspondingly. Followed by, in the fear class, the EA-ICSOML algorithm achieves accu y , prec n , reca l , F score , and AUC score of 100%, 100%, 100%, 100%, and 100%, correspondingly. Simultaneously, in the happy class, the EA-ICSOML system attains accu y , prec n , reca l , F score , and AUC score of 100%, 100%, 100%, 100%, and 100% respectively. Lastly, in the sadness class, the EA-ICSOML approach reaches accu y , prec n , reca l , F score , and AUC score of 99.29%, 95.45%, 100%, 97.67%, and 99.58%, correspondingly.

            Figure 4:

            Emotion detection outcome of the EA-ICSOML system on 20% of TSP. Abbreviation: EA-ICSOML, emotion analysis approach using improved cat swarm optimization with machine learning.

            Figure 5 exemplifies the outcomes of the EA-ICSOML technique on 80:20 of TRP/TSP. The outcomes referred to the EA-ICSOML methodology recognizing all types of emotions. With 80% of TRP, the SCADL-AFD system reaches average accu y , prec n , reca l , F score , and AUC score of 99.13%, 97.01%, 96.94%, 96.97%, and 98.22%. Afterwards, with 20% of TSP, the SCADL-AFD method realizes average accu y , prec n , reca l , F score , and AUC score of 99.59%, 98.56%, 98.61%, 98.55%, and 99.19%.

            Figure 5:

            Average outcome of the EA-ICSOML system on 80:20 of TRP/TSP. Abbreviation: EA-ICSOML, emotion analysis approach using improved cat swarm optimization with machine learning.

            Figure 6 inspects the accuracy of the EA-ICSOML method in the training and validation procedure on the test database. The result implies that the EA-ICSOML approach gains higher accuracy values over maximum epochs. Moreover, the enhanced validation accuracy over training accuracy outperforms that the EA-ICSOML approach learns effectively on the test database.

            Figure 6:

            Accuracy curve of the EA-ICSOML approach. Abbreviation: EA-ICSOML, emotion analysis approach using improved cat swarm optimization with machine learning.

            The loss curve of the EA-ICSOML system at the time of training and validation is displayed on the test database in Figure 7. The outcomes show that the EA-ICSOML approach attains near values of training and validation loss. The EA-ICSOML approach learns capably on a test database.

            Figure 7:

            Loss curve of the EA-ICSOML approach. Abbreviation: EA-ICSOML, emotion analysis approach using improved cat swarm optimization with machine learning.

            In Table 2 and Figure 8, the overall comparative outcome of the EA-ICSOML approach is provided. The outcome inferred that the EA-ICSOML method attains higher outcomes under all measures. Concerning accu y , the EA-ICSOML system reaches a higher accu y of 99.59% while the Gaussian NB, QDA, random forest (RF), MLP, support vector machine (SVM), and KNN approaches offer lesser accu y of 84%, 86%, 89%, 94%, 94%, and 97%. Similarly, based on prec n , the EA-ICSOML system attains a maximum prec n of 98.56% while the Gaussian NB, QDA, RF, MLP, SVM, and KNN methods provide lesser prec n of 84%, 85%, 90%, 94%, 94%, and 97%.

            Table 2:

            Comparative outcome of the EA-ICSOML approach with other systems.

            Classifier Accu y Prec n Reca l F score
            Gaussian NB84.0084.0084.0084.00
            QDA model86.0085.0086.0085.00
            Random forest89.0090.0089.0088.00
            MLP model94.0094.0094.0094.00
            SVM model94.0094.0094.0094.00
            KNN model97.0097.0097.0097.00
            EA-ICSOML99.5998.5698.6198.55

            Abbreviations: EA-ICSOML, emotion analysis approach using improved cat swarm optimization with machine learning; SVM, support vector machine.

            Figure 8:

            Comparative outcome of the EA-ICSOML approach with other systems. Abbreviations: EA-ICSOML, emotion analysis approach using improved cat swarm optimization with machine learning; SVM, support vector machine.

            Besides, in terms of reca l , the EA-ICSOML approach attains a superior reca l of 98.61% while the Gaussian NB, QDA, RF, MLP, SVM, and KNN systems offer minimal reca l of 84%, 86%, 89%, 94%, 94%, and 97%. Finally, based on F score , the EA-ICSOML approach accomplishes a maximum F score of 98.55% while the Gaussian NB, QDA, RF, MLP, SVM, and KNN models offer decreased F score of 84%, 85%, 88%, 94%, 94%, and 97%.

            CONCLUSION

            In this manuscript, we have presented a novel EA-ICSOML system for the emotion analysis process. The EA-ICSOML technique applies the concepts of CV and DL concepts to identify various types of emotions. The working process of the EA-ICSOML technique comprises ShuffleNet feature extraction, ICSO-based hyperparameter optimization, and TCNN-based classification. For feature vector generation, the ShuffleNet model is used in this work. To change the hyperparameters compared to the ShuffleNet approach, the ICSO algorithm was used. Finally, the recognition and classification of emotions take place using the TCNN model. The simulation result of the EA-ICSOML system was validated on facial emotion databases. The experimental analysis inferred the improved emotion recognition results of the EA-ICSOML algorithm compared to other recent models in terms of different evaluation measures.

            CONFLICTS OF INTEREST

            The authors declare no conflicts of interest in association with the present study.

            DATA AVAILABILITY STATEMENT

            Data sharing does not apply to this article as no datasets were generated during the current study.

            ETHICS APPROVAL

            This article does not contain any studies with human participants performed by any of the authors.

            REFERENCES

            1. Ahire V, Borse S. 2022. Emotion detection from social media using machine learning techniques: a surveyApplied Information Processing Systems. p. 83–92 . Springer. Singapore:

            2. Aoki T, Chujo R, Matsui K, Choi S, Hautasaari A. 2022. EmoBalloon-conveying emotional arousal in text chats with speech balloonsProceedings of the 2022 CHI Conference on Human Factors in Computing Systems; p. 1–16

            3. Catania F, Garzotto F. 2023. A conversational agent for emotion expression stimulation in persons with neurodevelopmental disorders. Multimed. Tools Appl. Vol. 82(9):12797–12828

            4. Chen L, Hu Z, Zhang F, Guo Z, Jiang K, Pan C, et al.. 2022. Remote wind farm path planning for patrol robot based on the hybrid optimization algorithm. Processes. Vol. 10(10):2101

            5. Cui B, Wang J, Lin H, Zhang Y, Yang L, Xu B. 2022. Emotion-based reinforcement attention network for depression detection on social media: algorithm development and validation. JMIR Med. Inform. Vol. 10(8):e37818

            6. De A, Mishra S. 2022. Augmented intelligence in mental health care: sentiment analysis and emotion detection with health care perspectiveAugmented Intelligence in Healthcare: A Pragmatic and Integrated Analysis. p. 205–235. Springer. Singapore:

            7. Garcia-Garcia JM, Penichet VM, Lozano MD, Fernando A. 2022. Using emotion recognition technologies to teach children with autism spectrum disorder how to identify and express emotions. Univers. Access Inf. Soc. Vol. 21(4):809–825

            8. Hou J. 2022. Deep learning-based human emotion detection framework using facial expressions. J. Interconnect. Netw. Vol. 22 Supp 01:2141018

            9. Kumar S, Prabha R, Samuel S. 2022. Sentiment analysis and emotion detection with healthcare perspectiveAugmented Intelligence in Healthcare: A Pragmatic and Integrated Analysis. p. 189–204. Springer. Singapore:

            10. Modran HA, Chamunorwa T, Ursuțiu D, Samoilă C, Hedeșiu H. 2023. Using deep learning to recognize therapeutic effects of music based on emotions. Sensors. Vol. 23(2):986

            11. Mridha K, Islam MI, Shorna MM, Priyok MA. 2022. ML-DP: a smart emotion detection system for disabled person to develop a smart city2022 10th International Conference on Reliability, Infocom Technologies and Optimization (Trends and Future Directions)(ICRITO); p. 1–6. IEEE.

            12. Nandwani P, Verma R. 2021. A review on sentiment analysis and emotion detection from text. Soc. Netw. Anal. Min. Vol. 11:1–19

            13. Riza MA, Charibaldi N. 2021. Emotion detection in Twitter social media using long short-term memory (LSTM) and fast text. Int. J. Artif. Intell. Robot. Vol. 3:15–26

            14. Sailunaz K, Dhaliwal M, Rokne J, Alhajj R. 2018. Emotion detection from text and speech: a survey. Soc. Netw. Anal. Min. Vol. 8:1–26

            15. Ullah N, Raza A, Khan JA, Khan AA. 2022. An effective approach for automatic COVID-19 detection from multiple image sources using shufflenet convolutional neural network (CNN).

            16. Vasantharajan C, Benhur S, Kumarasen PK, Ponnusamy R, Thangasamy S, Priyadharshini R, et al.. 2022. Tamilemo: finegrained emotion detection dataset for tamil. arXiv. 2202.04725

            17. Vijayvergia A, Kumar K. 2021. Selective shallow models strength integration for emotion detection using GloVe and LSTM. Multimed. Tools Appl. Vol. 80:28349–28363

            18. Wang L, Han S. 2023. Improved cat swarm optimization for parameter estimation of mixed additive and multiplicative random error model. Geod. Geodyn. Vol. 14:385–391

            19. Zad S, Heidari M, James H Jr, Uzuner O. 2021. Emotion detection of textual data: an interdisciplinary surveyProceedings of the 2021 IEEE World AI IoT Congress (AIIoT); Seattle, WA, USA. 10-13 May 2021; p. 255–261. IEEE. Piscataway, NJ, USA:

            Author and article information

            Journal
            jdr
            Journal of Disability Research
            King Salman Centre for Disability Research (Riyadh, Saudi Arabia )
            1658-9912
            28 March 2024
            : 3
            : 3
            : e20240017
            Affiliations
            [1 ] Department of Information Systems, College of Computer and Information Sciences, Princess Nourah Bint Abdulrahman University, Riyadh 11671, Saudi Arabia ( https://ror.org/05b0cyh02)
            [2 ] Department of Computer and Self Development, Preparatory Year Deanship, Prince Sattam Bin Abdulaziz University, AlKharj, Saudi Arabia ( https://ror.org/04jt46d36)
            Author notes
            Author information
            https://orcid.org/0000-0003-3881-087X
            Article
            10.57197/JDR-2024-0017
            9d77db9a-42e1-4738-bf37-3b56662975d7
            Copyright © 2024 The Authors.

            This is an open access article distributed under the terms of the Creative Commons Attribution License (CC BY) 4.0, which permits unrestricted use, distribution and reproduction in any medium, provided the original author and source are credited.

            History
            : 24 May 2023
            : 16 October 2023
            : 20 February 2024
            Page count
            Figures: 8, Tables: 2, References: 19, Pages: 9
            Funding
            Funded by: King Salman Center for Disability Research
            Award ID: KSRG-2023-377
            The authors extend their appreciation to the King Salman Center for Disability Research (funder ID: http://dx.doi.org/10.13039/501100019345) for funding this work through Research Group no KSRG-2023-377.
            Categories

            Computer science
            computer vision,emotion recognition,machine learning,deep learning,speech-impaired people

            Comments

            Comment on this article