INTRODUCTION
A neurodegenerative disease, known as autism spectrum disorder (ASD), results in long-lasting deficiencies in children’s behavior and social communication abilities ( Bahathiq et al., 2024; Nogay and Adeli, 2024). The World Health Organization reports that 0.63% of children have an ASD diagnosis. It first manifests in childhood and then spreads to people of all ages. Most of the time, symptoms start to show up in the first 5 years of life. ASD is a severe neurodevelopmental disease that carries a substantial cost for medical care. Stereotypical behavior and a persistent lack of social contact are hallmarks of ASD ( Kumar et al., 2024), and they are frequently coupled with a generalized decline in communication abilities. There are neurological and genetic components to ASD. Social interaction, an aptitude for imagination and thinking, obsessive behaviors, and issues with communicating with others are all examples of behaviors associated with ASD ( Gao et al., 2024; Saponaro et al., 2024). The precise processes that explain the connection between ASD and neurological problems are not well understood and are believed to include intricate interactions among genetic, environmental, and neurobiological variables. ASD is mostly defined by impairments in social communication and interaction, along with limited and repetitive behavior patterns. However, persons with ASD often have other neurological problems that occur alongside. Many persons with ASD sometimes have difficulties in processing sensory information, which may manifest as either heightened sensitivity or reduced sensitivity to sensory stimuli, such as sound, touch, or light. The presence of sensory processing impairments may have a substantial effect on everyday activities and contribute to the behavioral symptoms associated with ASD. The three categories for social communication skills are deficiencies in interpersonal–emotional cooperation; problems in nonverbal communication actions; and deficiencies in growing, sustaining, and comprehending connections. Additionally, prohibited and monotonous patterns of behavior as well as obsessed interest are considered behavioral deficiencies in children diagnosed with autism. These behaviors are characterized by the following ( Dhamale et al., 2023): (i) stereotypical or recurrent motor actions, (ii) reliance on uniformity, (iii) obstacles adapting to abrupt shifts in everyday or rigid sequences of either nonverbal or verbal conduct, or (iv) distinctive attitude toward visual facets. Hyper- or hypo-sensitivity to a variety of modalities, including hearing, smell, taste, and body awareness, can occur in children with learning problems. For instance, noises from air conditioners may be upsetting to kids who are extremely sensitive to noise. In addition, self-harm, standing on their tiptoe, thrashing of the fingers or firearms, and swaying from side to side are other activities that frequently correspond with perceptual problems ( Wang et al., 2023). Early detection and treatment are crucial. Parents, educators, and other people without specialization or credentials can screen children for autism.
Automated methods for detecting ASD may aid clinicians by simplifying the process of screening and diagnosis, hence decreasing the amount of time and effort needed for evaluation. This enables physicians to concentrate on more intricate situations and provide individualized treatment to persons with ASD. Timely identification of ASD is essential for prompt intervention and treatment, resulting in improved outcomes for persons with ASD. Automated detection systems may expedite the process of identifying potential issues in children, enabling timely access to suitable therapies and support services. ASD detection technologies that are automated may be implemented on platforms that can be scaled, such as mobile apps or web-based platforms, which allows them to be available to a larger audience. This facilitates screening in many locations, such as schools, healthcare institutions, and distant places, hence enhancing the availability of early diagnosis and intervention services.
According to the American Academy of Pediatrics, screening should be performed on all children on a frequent basis to make sure they are receiving enough support to achieve their greatest potential. A healthcare provider can use the data acquired from the evaluation and observation procedure to gain a good knowledge of an individual’s conduct or social skills prior to the diagnosing phase ( Alvarez-Jimenez et al., 2020; Das Biswas et al., 2020). Only medical professionals are qualified to make diagnoses. Using established diagnostic tools, medical professionals will diagnose youngsters with autism. In a conversation with parents or other carers, they will also go over the child’s past developmental trajectory and present behavior. Early assessment and therapy are possible for ASD. Clinical diagnosis of ASD depends critically on early detection. Customized therapies can then be developed based on this diagnosis to enhance the quality of life of children with ASD and their carers. However, using costly testing procedures, the diagnosis of ASD can be a drawn-out process. The rise in ASD cases observed globally in recent times has prompted medical and scientific researchers to develop more efficacious screening techniques ( Mohan and Paramasivam, 2021). Owing to the development of modern technology, we are now able to maintain enormous quantities of data. To make judgments according to the data gathered, data mining is an essential responsibility. It has to do with machine learning. The field of machine learning has made significant strides in the last several years, and applied fields such as medicine and biology are beginning to understand its value. Machine learning techniques are employed or indicated in medical decision-making and diagnosis to help with the interpretation of data ( Lamani and Julian Benadit, 2023). As a result, machine learning-assisted disease evaluation methods are extensively researched. There are currently a number of studies on ASD, although they are not without limits. Some research just looks at medical professionals, and certain studies concentrate on gathering ASD data. It is also crucial to provide a rapid, simple, and easily accessible way to assist with the early identification of ASD. Seeking treatment from specialists is beneficial for families of patients with ASD. Many machine learning and deep learning techniques have been developed for ASD diagnosis and classification in the literature materials ( Epalle et al., 2021; Yang et al., 2022). However, the traditional methods struggle with the particular issues of higher computing expenses, longer execution times, and reduced effectiveness. Therefore, the goal of the proposed study is to build an automated artificial intelligence (AI) tool for ASD detection that combines a number of cutting-edge mining techniques to guarantee the highest level of disease prediction accuracy ( Chen et al., 2020; Fu et al., 2021). Additionally, the public University of California, Irvine (UCI) databases are used in this work to detect ASD. The extensive and widely used ASD datasets have a connection with the clinical diagnosis of ASD in people of all ages. The specific contributions of this work are listed below:
For accurate and efficient ASD detection, this study presents an automated and lightweight framework named attention-based hybrid optimized residual memory network (AHRML).
In order to streamline the identification process for the disabled, a new hybridized Arithmetic Harris Hawks Optimizer (AHHO) is used to reduce the dimensionality of features.
In addition, attention-based residual term memory (ARTM), a clever deep learning algorithm, is created to accurately and less frequently identify ASD from the given data.
Using the well-known ASD dataset that was taken from the UCI repository, the suggested AHRML model’s findings were validated and tested using a variety of parameters.
The remainder of the components of this work are split into the following sections: a thorough review of the literature is provided in the Related Works section to look at several approaches currently in use for ASD classification and identification. It also talks about the issues and difficulties that traditional approaches encounter. The Proposed Methodology section presents a concise synopsis of the suggested methodology, complete with a block diagram, mathematical explanations, and algorithms. In the Results and Discussion section, specific public ASD datasets and assessment metrics are used to validate the performance outcomes and results of the suggested model. In the Conclusion section, the entire study is finally summarized together with the results, conclusions, and next steps.
RELATED WORKS
This section explores a number of current approaches to the identification and categorization of ASDs. It also evaluates the benefits, issues, and difficulties that traditional methods present in light of their performance results.
Khodatars et al. (2021) did a thorough analysis of the literature to look at several deep learning techniques for ASD identification. The study’s authors accurately determined the condition by using neuroimaging technologies. This study presents a comprehensive review of various deep learning strategies, encompassing the following techniques: autoencoders (AEs), convolutional neural networks (CNNs), recurrent neural networks, generative adversarial networks, deep belief networks, and deep Boltzmann machines. The results of this study show that deep learning approaches are effective in detecting ASD because of their strong capacity to handle large datasets with guaranteed accuracy. Sherkatghanad et al. (2020) developed an efficient and automated ASD detection and recognition system using the CNN technique. In this case, the layered architecture is created to carefully examine the inputs in order to identify the ASD with accuracy. Nevertheless, the recommended method takes longer for the validation and training phases, which has a big impact on the system’s overall efficacy. Ke et al. (2020) conducted a thorough evaluation of the literature to evaluate the efficacy and performance of 14 different learning approaches used to diagnose ASD. Deep learning approaches have been adopted in a number of studies recently for the identification of mental disorders in the literature. This is because deep learning techniques have proven to be more effective in complicated pattern analysis and disease diagnosis. Preethi et al. (2022) used a two-dimensional CNN technique to identify ASDs efficiently by examining the input pattern analysis. This work uses the AdaDelta optimization methodology to enhance the detection performance of disorder detection. Qiang et al. (2023) developed a hierarchical brain network method to efficiently identify and recognize ASD. The primary focus of their study was on examining the input feature analysis in order to minimize the mistake rate and forecast disorder. Liu et al. (2021) applied a multi-regional data-driven attention learning strategy for the identification of ASD. Here, the AE and long short-term memory (LSTM) deep learning algorithms are combined to provide the recommended learning strategy. This study has made use of the Autism Brain Image Data Exchange (ABIDE) database to examine the results of the proposed model. Yang et al. (2020) utilized a deep neural network (DNN) classification approach for an effective identification and classification of ASD. The classification of data using both voxels and connectivity functions has been effectively accomplished through the use of DNNs. In order to get enough examples for the DNN study, the multi-layer perceptron is applied with four distinct configurations leveraging the ABIDE repository.
Sewani and Kashef (2020) utilized an AE-based deep learning technique for an effective prediction of autism. AEs are essential for extracting low-dimensional features, which may then be fed into a neural network or learning models for classification problems. This architecture uses nonlinear optimization to reduce dimensionality. With fewer neurons in the hidden layer than in the input, the suggested approach concentrated on exploiting AE to extract pertinent data from the input layer. Zhang et al. (2022) implemented a variational AE model combined with a special feature selection method to efficiently identify and categorize ASD. In order to swap out the primary tanh function with a modified one, this work constructed a pipeline and used a threshold-shifting strategy, which might further improve the efficiency of classification. Furthermore, two restrictions that aid in training models with increased specificity or sensitivity have been applied. This strategy may be a useful supplementary technique for early diagnosis of ASD, given its exceptional classification performance on the heterogeneous dataset. This implies that this technique improves the state-of-the-art approaches even further. Subah et al. (2021) used a DNN to identify ASD from the ABIDE dataset as accurately as possible. Here, the most important characteristics from the provided photos have been extracted using the time series feature extraction model. Several machine learning methods have been used in earlier research to predict ASD, in which some of the most well-liked approaches are discussed below ( Sujatha et al., 2021; Hasan et al., 2022).
Decision tree
One member of the supervised learning algorithm group is this algorithm. Regression and classification issues can be addressed through involvement. By examining fundamental decision rules (training data) drawn from previous data, a decision tree (DT) is used to construct a model for training that may be applied in predicting the type or amount of the variable being targeted. This algorithm has the benefit of requiring less work to set up the data during preliminary processing; however, it also has the downside of becoming unstable if a small change in any component of the data triggers a significant change in the process of decision-making stump.
Random forest
In order to create the class and facilitate learning, random forests (RFs) generate a huge number of DTs ahead of time. Overfitting the training set is optimal for DTs. The usual bootstrap approach of pooling is applied to tree learners by the RF modeling approach. Bagging often selects a random sample in place of a training set and compares trees with samples of this kind.
Support vector machine
It is a linear algorithm that handles classification and regression issues. Support vector machine (SVM) has a straightforward concept: the data are divided into groups by the algorithm, which produces a line or hyper-plane. This algorithm’s effective memory management is one of the advantages it offers, although large datasets are not a good fit for it. Moreover, the SVM is thought to be a classifier that maximizes the difference between specific important variables. As the foundation of this approach is optimizing the distance among the support vectors that produce the optimum hyper-plane, it is more likely to produce more effective data separation. But, it is striving to cross a threshold that raises the margin, necessitating the application of an optimizing process and requires that every point be on the appropriate segment of the hyper-plane.
Naive Bayes
This statistical classification technique is based on the Bayes theorem. This is one of the simplest supervised learning algorithms and is fast, precise, and trustworthy. On massive datasets, naive Bayes classifiers offer great speed as well as precision. One of the most potent and adaptable learning algorithms for data analysis and algorithmic learning, it has a high prediction efficiency. In a controlled setting, it is successfully trainable. This can be used with different attributes for small and large training data collection, and it often contains very little training data. This sorting algorithm has the knack of handling missing values.
The majority of earlier works ( Bala et al., 2022; Farooq et al., 2023) target to establish an efficient intelligence-based learning methodology for detecting ASD from available data, according to the literature evaluation and analysis. However, the main issues with conventional techniques are their lower detection rate, longer prediction times, and mathematical modeling complexity. There is a lack of research on deep learning models for longitudinal monitoring and progress tracking of individuals with ASD over time. Developing models that can track changes in behavior and symptom severity longitudinally could provide valuable insights for personalized intervention planning and treatment evaluation. Therefore, the goal of the proposed research project is to put into practice a novel, lightweight AI-based detection framework for the precise identification and prediction of ASD based on the provided medical data.
PROPOSED METHODOLOGY
A detailed explanation of the proposed ASD detection methodology, including an overview, a block diagram, and algorithmic details, is given in this section. Typically, the ASD is a chronic neurological disease. In order to provide proper medical care, ASD requires early, fast, precise, and successful detection. As an alternative to the traditional approaches, which focus on a quicker and cheaper patient diagnosis, data mining techniques have been used to identify the injured person. The traditional processes that have been developed and carried out in the past have been costly and laborious. Although much research in this area has been conducted with effectiveness in assessing ASD, these studies have not yet achieved outstanding performance in a time that is less expensive. The design and implementation of a unique and distinctive framework, known as AHRML for the identification and categorization of ASD from the provided medical data, constitutes the work’s original contribution. Data preprocessing, feature optimization, ASD prediction, and performance estimation are the most significant aspects of the proposed AHRML system, which is presented in Figure 1. For system development and validation, the public open-source dataset is taken into consideration in this work initially. Following the acquisition of the data, typical preprocessing steps—which include attribute balancing, scaling, outlier removal, and missing value imputation—are used to produce normalized data. As a consequence, the previously processed information is used to determine the crucial and required features, which reduces the training and validation time and the computational load of classification. This is accomplished by implementing the novel hybrid optimization technique, AHHO. Subsequently, the acquired features are passed to an intelligence-based classifier, ARTM, which uses processing speed and detection rate to reliably identify ASD. Lastly, a thorough simulation and performance analysis are done to look at the results of the suggested AHRML system.

Overview of the proposed AHRML model. Abbreviations: AHRML, attention-based hybrid optimized residual memory network; ASD, autism spectrum disorder; LSTM, long short-term memory; ResNet, residual network.
The integration of residual connections in a hybrid optimized residual memory network enhances interpretability by establishing unambiguous routes for the transmission of information inside the network. Furthermore, the use of memory processes may improve the comprehensibility of the model by allowing it to store and recall pertinent information from previous observations. This facilitates doctors’ understanding and confidence in the model’s predictions. Prior studies have shown that the hybrid optimized residual memory network is very successful in a range of machine learning tasks, such as medical diagnosis and prediction. The hybrid optimized residual memory network utilizes its hybrid architecture and optimization methodologies to produce exceptional performance in automated ASD diagnosis, beyond the capabilities of traditional deep learning models. This enhances the accuracy and reliability of diagnostic outcomes.
Dataset characteristics
The dataset comprises several aspects including demographic information (such as age and gender), behavioral factors (including social skills and communication abilities), and diagnostic information pertaining to ASD. The dataset has a certain quantity of instances or samples, with each instance representing a child. The target variable frequently denotes the existence or nonexistence of ASD, functioning as the label for classification tasks. The dataset is often used for the purpose of researching and developing machine learning models pertaining to the diagnosis, prediction, and categorization of autism. Preprocessing stages include several tasks such as managing missing data, encoding categorical variables, and scaling numerical characteristics, among other tasks.
AHHO for feature selection
At this point, the most essential and useful information qualities have been selected from the provided data using the cutting-edge hybrid optimization technique, AHHO. The ASD detection system’s overall processing efficiency and speed are greatly increased by using this optimizer. The development of this hybridized model involves the integration of two distinct and widely used optimization techniques, namely the Harris hawks optimization (HHO) algorithm and the arithmetic algorithm (AA). As a result, many optimization strategies for feature selection are being explored in the literature.
HHO is known for its ability to perform global optimization by leveraging the collaborative behavior of Harris hawks. In the context of dimensionality reduction, global optimization can help identify a low-dimensional embedding that captures the most informative features of the dataset while minimizing information loss or distortion. Many dimensionality reduction algorithms involve nonconvex optimization problems, where traditional optimization techniques may struggle to find globally optimal solutions. HHO’s capacity to escape local optima and explore diverse regions of the search space can be advantageous for optimizing nonconvex objective functions commonly encountered in dimensionality reduction. The key advantages of the AA and HHO algorithms, in comparison to others, are their higher searching efficiency, low number of iterations needed to locate the optimal value, and quick solution determination time. Therefore, in order to develop a novel and distinctive optimizer for feature selection, the proposed work combines these two models. The architecture is composed of a primary layer at the top of the hierarchy that contains a number of HHO agents for searching and a supplementary layer at the bottom that comprises subgroups with a number of arithmetic optimization algorithm (AOA) populations. The procedure of upgrading the search agents’ locations is started by the AOA operation. It is essential to make sure that the locations of every searching agent in the highest layer are synchronized with the most effective solution found by the group that corresponds in the bottom layer in order to locate the best possible solution. A solution that is even more efficient can be achieved by developing new equations for each of the exploration and exploitation stages. This method greatly enhances the outcome and enables an in-depth examination of the problem at hand. The system takes its information from how Harris hawks hunt, making use of their phenomenal ability to locate and identify the prey with accuracy. Sometimes, though, a potential victim might not seem easily identifiable, so the hawks have to stay observant and keep an eye on the target area for a while before they may spot their next supper. In order to evaluate the most suitable option with the expected outcome, the system expertly utilizes Harris hawks to symbolize multiple options. By positioning the hawks in strategic positions and employing a variety of techniques, the algorithm accurately mimics the hawks’ movements until the intended outcome is obtained. Each tactic is applied with equal care and attention to guarantee the greatest efficiency. At first, the exploration strategy of AHHO is executed by updating the location of searching agents in the topmost layer, which is mathematically represented as shown below:
where δij+1 indicates the location of searching agents in the topmost layer, j indicates the present iteration, α, β, γ, ϑ, x, 𝔎 are the random numbers in the range of 0 to 1, h r denotes the random hawk, h b defines the best location of the hawk, h 𝕞 indicates the average mean of the hawk, l j , u j are the lower and upper bounds respectively, and mop represents the math optimizer probability coefficient. It is estimated as shown in the following equations:
where 𝔥 is the parameter that determines the precision of exploitation, k and K are the current and maximum number of iterations, ℵ defines the total number of hawks in the population, and h i ( j) denotes the position of the hawk. After that, the transition is performed that balances the modes of exploration and exploitation according to the energy level of the prey, which is computed based on the following equation:
where ℰ is the energy level, and ℰ 0 denotes the initial energy. Harris hawks use an unanticipated swoop technique in the attacking phase to seize their prey, which is detected in the earlier stage. The rabbit has enough energy to escape by making erratic and cunning leaps during |ℰ| ≥ 0.5. Moreover, the rabbit grows more and more exhausted as the Harris hawks close in on it, and then they ambush attack, which is illustrated by the following equations:
The hybrid algorithm makes use of the concept of incorporating the Levy flight pattern in order to statistically represent the leapfrog motions of the predator and the escape patterns of the prey. It mimics the unpredictable, abrupt, and swift dives that hawks make when they approach their intended prey, as well as the circular movements of the prey—especially rabbits—during their flight. Then, the position update is performed based on the following equations:
The Harris hawks will encircle the rabbit fiercely and make surprise hits in an effort to capture it. This phase of the resilient attack is described in the following equations:
This process culminates in the output of the best optimal value, which is utilized to select the most important and necessary characteristics from the normalized dataset. The suggested model’s overall ASD detection performance and results are significantly enhanced by using this hybridized optimizer.
ARTM classification
In this stage, a novel ARTM classification strategy is used to predict ASD using the features that were acquired in the previous stage. For the diagnosis of ASD, a range of learning algorithms are typically created in earlier literature publications. Deep learning techniques work well with a high disease diagnosis rate, making them superior to machine learning algorithms. However, the typical issues of increased error rate, high time complexity, and complexity in data training and validation plague standard deep learning methods. Therefore, the goal of the presented study is to design a deep learning system for ASD diagnosis that is both efficient and lightweight. The attention block, residual network (ResNet) block, and LSTM block are combined to develop the proposed ARTM approach. The human attention process is modeled after the biological mechanisms that focus promptly on the most unique and significant aspects of the problem at hand. After training, typical neural networks frequently acquire stable weight matrices, which do not alter even if the network is fed entirely new inputs. As a result, classic neural network models find it challenging to adjust to rapidly changing settings.
ARTM allows the model to selectively attend to relevant features and ignore irrelevant ones during different stages of processing. By dynamically adjusting the attention weights, the model can focus on important features while suppressing noise or irrelevant information, leading to more effective feature representation and improved discrimination between classes. The memory component in ARTM enables the model to maintain information over multiple time steps or layers. This is particularly beneficial for tasks involving sequential data or long-range dependencies, such as natural language processing or time series analysis. By preserving the context and temporal relationships, ARTM can capture complex patterns and dependencies that may be crucial for accurate prediction or classification.
A neural network could create self-modifying weights and focus on the most important parts of the input through the incorporation of the attention mechanism. The idea put forward makes use of the temporal attention structure and the channel attention mechanism, which are two of the attention mechanism approaches that have been put forth at this point. By combining spatial and channel-wise data through convolution processes, a ResNet can obtain useful features from layered data. It has been shown that by clearly modeling the interrelated relationships between the channels of attribute maps, the efficiency of a network can be significantly enhanced. Consequently, by assigning each channel of feature maps a weight parameter, we incorporate the channel attention mechanism with ResNet in order to adapt and adjust channel-wise information. The backpropagation method can still be used to train such weight factors. These procedures are made possible by the inclusion of the squeeze-and-excitation blocks. The architecture model of the proposed ARTM model is shown in Figure 2.

Architecture of the ARTM model. Abbreviations: ARTM, attention-based residual term memory; FC, fully connected; GAP, global average pooling; LSTM, long short-term memory; ResNet, residual network.
Channel-wise information gets generated by means of global average pooling in the squeezing phase. Two fully connected (FC) layers are used in the following excitation process to identify channel interactions and restrict cost. Specifically, one FC layer with a reduction ratio lowers the inputs’ scale, and the additional FC layer eventually recovers the channel dimension to its initial value. Implementing a channel-wise convolution among stream weights and feature maps produced by the ResNet yields the block’s final outcome. Putting the last hidden state as the output continues to be the default in a standard LSTM prediction process. Nevertheless, as the input data sequence grows longer, the prediction accuracy will drop. In order to address this issue, we incorporate a temporal attention layer into the final LSTM layer, enabling the consideration of all concealed states. The last time step’s concealed state is designated as a typical state since it has historical data in it. After that, it is compared with all hidden states in order to determine their respective scores. Higher weights will be allocated to the hidden state with the greatest score. As a result, the system has the potential to effectively highlight the informative elements of various inputs and can adjust to input sequences that are both short- and long-term. The temporal attention operations are mathematically illustrated below:
where 𝒽 𝓈 indicates the hidden state at time step 𝓈, 𝒽 𝒮 denotes the final hidden state at time step 𝒮, ϖ 𝓈 represents the attention weight value, 𝓍 𝒮 defines the learnable matrix, and 𝒽 𝑜 is the final output of the temporal attention layer. Then, the final prediction result is obtained as shown in the following equation:
where ℴ denotes the predicted result, ϖ i defines the coefficient variable of ith transformed variable, and ϖ 0 represents the optional bias term. This hybrid learning model improves performance outcomes by accurately predicting ASD in the suggested framework.
RESULTS AND DISCUSSION
This section uses the widely used ASD dataset and assessment metrics to validate the effectiveness and results of the proposed AHRML system. A Windows 10 PC with specifications including a CPU speed of 2.9 GHz core i7, GPU of Intel 620, RAM of 12 GB, and a spare hard drive space of a minimum of 5 GB has been utilized for the experiments in order to successfully run and validate the proposed configuration. Although the prevalence of ASD is rising globally, there are very few publicly accessible statistics dedicated to the disorder’s research. There are currently too few clinical screening datasets for autism, and the majority of resources are genetics-focused. The ASD in the adult and children’s datasets, which comprised 704 instances and 21 attributes in the adult dataset and 292 instances and 21 attributes in the children’s dataset, was taken from a public UCI repository ( Hossain et al., 2021; Sujatha et al., 2021) and employed in the present study. This children’s dataset contains 292 observations, with ages ranging from 4 to 11. Working with the ASD dataset presents many challenges due to the multitude of respondents, and the dataset details are given in Table 1.
ASD dataset description.
S. No | Data information | Type of attributes | No. of samples | Yes (Y) or no (N) |
---|---|---|---|---|
1 | ASD screening for children | Binary, categorical, and continuous | 292 | Y—141; N—151 |
2 | ASD screening for adult | Binary, categorical, and continuous | 704 | Y—189; N—515; |
3 | ASD screening for combined information | Binary, categorical, and continuous | 996 | Y—330; N—666 |
Abbreviation: ASD, autism spectrum disorder.
Accuracy in predictive modeling refers to the percentage of total data points that are successfully estimated. It shows the proportion of accurate forecasts in a simulated set of data. It is crucial since AI learns on its own, making it challenging to assess the accuracy of supplied data:
The ratio of positive events to all genuine positive examples is known as precision. To put it plainly, precision tells us how true a model is when it says it is correct. In order to compute the same value, apply the following equation:
Recall, which is calculated as the ratio of true positives to the total number of real positive occurrences, quantifies the percentage of actual positive events that a system or model properly detected. The outcome is computed using the formula below.
The F1-score is the combined precision and recall metric, with 1 denoting the best possible outcome and 0 the worst. An accurate disease detection system that strikes a balance between false positives and false negatives is indicated by an increased F1-score.
where 𝒯𝒫 indicates the true positives, 𝒯𝒩 indicates true negatives, ℱ𝒫 represents the false positives, and ℱ𝒩 represents the false negatives. Using a dataset of children with ASD, Tables 2 and 3 assess and contrast the overall effectiveness of the suggested learning techniques with the standard approaches. Next, corresponding graphical illustrations for each of them are shown in Figures 3– 5. The following factors were taken into account for this analysis: F1-score, kappa coefficient, accuracy, precision, recall, log loss value, and area under the curve. This comparative analysis shows that, in comparison to the other current classification approaches, the suggested AHRML model performs well and offers better performance yields.
Performance comparative study for children’s ASD dataset.
Methods | Accuracy (%) | Precision (%) | Recall (%) | F1-score (%) |
---|---|---|---|---|
NB | 94.25 | 90.91 | 97.56 | 94.12 |
KNN | 86.21 | 87.18 | 82.93 | 85 |
SVM | 98 | 98.5 | 99 | 99 |
RF | 96.55 | 93.18 | 99 | 96.47 |
DT | 87.36 | 87.50 | 85.37 | 86.42 |
XGB | 97.70 | 97.56 | 97.56 | 97.5 |
LR | 98 | 98.5 | 98 | 98 |
ANN | 98.85 | 97.62 | 98.8 | 98.8 |
Proposed | 99 | 98.9 | 99 | 99.2 |
Abbreviations: ANN, artificial neural network; ASD, autism spectrum disorder; DT, decision tree; KNN, k-nearest neighbor; NB, naive Bayes; RF, random forest; SVM, support vector machine; LR, linear regression; XGB, XGBoost.
Comparison based on AUC, kappa, and log loss value for children’s ASD dataset.
Methods | AUC | Kappa | Log loss value |
---|---|---|---|
NB | 0.94 | 0.88 | 2.071 |
KNN | 0.86 | 0.72 | 4.972 |
SVM | 0.85 | 0.84 | 2.698 |
RF | 0.96 | 0.93 | 1.243 |
DT | 0.87 | 0.74 | 4.557 |
XGB | 0.97 | 0.91 | 0.829 |
LR | 0.88 | 0.91 | 0.878 |
ANN | 0.98 | 0.97 | 0.414 |
Proposed | 0.99 | 0.99 | 0.152 |
Abbreviations: ANN, artificial neural network; ASD, autism spectrum disorder; AUC, area under the curve; DT, decision tree; KNN, k-nearest neighbor; NB, naive Bayes; RF, random forest; SVM, support vector machine.

Overall performance analysis using children’s ASD dataset. Abbreviations: ANN, artificial neural network; ASD, autism spectrum disorder; DT, decision tree; KNN, k-nearest neighbor; NB, naive Bayes; RF, random forest; SVM, support vector machine.

AUC and kappa analysis for children’s ASD dataset. Abbreviations: ANN, artificial neural network; ASD, autism spectrum disorder; AUC, area under the curve; DT, decision tree; KNN, k-nearest neighbor; NB, naive Bayes; RF, random forest; SVM, support vector machine.

Log loss value for children’s ASD dataset. Abbreviations: ANN, artificial neural network; ASD, autism spectrum disorder; DT, decision tree; KNN, k-nearest neighbor; NB, naive Bayes; RF, random forest; SVM, support vector machine.
In the same way, the adult ASD dataset’s overall performance comparison study is carried out, as indicated in Tables 4 and 5. Next, their corresponding graphical representations are shown in Figures 6– 8. These results also show that, in comparison to the other classification methods, the AHRML model yields better results for ASD detection.
Performance comparative study for the adult ASD dataset.
Methods | Accuracy (%) | Precision (%) | Recall (%) | F1-score (%) |
---|---|---|---|---|
NB | 96.19 | 95 | 91.94 | 93.44 |
KNN | 89.05 | 84.21 | 77.42 | 80.67 |
SVM | 96.19 | 93.55 | 93.55 | 93.55 |
RF | 95.71 | 94.92 | 90.32 | 92.56 |
DT | 87.14 | 84.31 | 69.35 | 76.11 |
XGB | 96.19 | 95 | 91.94 | 93.44 |
LR | 97.14 | 96.67 | 93.55 | 95.05 |
ANN | 96.67 | 91.04 | 98.39 | 94.57 |
Proposed | 99 | 98.8 | 98.9 | 99.1 |
Abbreviations: ANN, artificial neural network; ASD, autism spectrum disorder; DT, decision tree; KNN, k-nearest neighbor; NB, naive Bayes; RF, random forest; SVM, support vector machine.
Comparison based on AUC, kappa, and log loss value for the adult ASD dataset.
Methods | AUC | Kappa | Log loss value |
---|---|---|---|
NB | 0.94 | 0.90 | 1.373 |
KNN | 0.85 | 0.73 | 3.948 |
SVM | 0.95 | 0.90 | 1.373 |
RF | 0.94 | 0.89 | 1.545 |
DT | 0.81 | 0.80 | 4.634 |
XGB | 0.94 | 0.90 | 1.373 |
LR | 0.96 | 0.93 | 1.029 |
ANN | 0.97 | 0.92 | 1.201 |
Proposed | 0.99 | 0.989 | 0.211 |
Abbreviations: ANN, artificial neural network; ASD, autism spectrum disorder; AUC, area under the curve; DT, decision tree; KNN, k-nearest neighbor; NB, naive Bayes; RF, random forest; SVM, support vector machine.

Overall performance analysis using the adult ASD dataset. Abbreviations: ANN, artificial neural network; ASD, autism spectrum disorder; DT, decision tree; KNN, k-nearest neighbor; NB, naive Bayes; RF, random forest; SVM, support vector machine.

AUC and kappa analysis for the adult ASD dataset. Abbreviations: ANN, artificial neural network; ASD, autism spectrum disorder; AUC, area under the curve; DT, decision tree; KNN, k-nearest neighbor; NB, naive Bayes; RF, random forest; SVM, support vector machine.

Log loss value for the adult ASD dataset. Abbreviations: ANN, artificial neural network; ASD, autism spectrum disorder; DT, decision tree; KNN, k-nearest neighbor; NB, naive Bayes; RF, random forest; SVM, support vector machine.
Furthermore, as indicated in Tables 6 and 7, the performance outcomes are estimated and contrasted for the combined children + adult dataset. Figures 9– 11 provide the relevant graphical depictions of them. When compared to all other current techniques, the comparison data show that the suggested AHRML produces an ASD prediction result with a lower log loss value because adopting hybrid AHHO is the primary means of achieving better categorization outcomes.
Performance comparative study for the combined ASD dataset.
Methods | Accuracy (%) | Precision (%) | Recall (%) | F1-score (%) |
---|---|---|---|---|
NB | 93.27 | 86.41 | 93.68 | 89.89 |
KNN | 83.16 | 72.73 | 75.79 | 74.23 |
SVM | 91.58 | 88.04 | 85.26 | 86.63 |
RF | 93.27 | 92.13 | 86.32 | 89.13 |
DT | 91.92 | 87.37 | 87.37 | 87.37 |
XGB | 92.59 | 91.01 | 85.26 | 88.04 |
LR | 91.58 | 91.67 | 81.05 | 86.03 |
ANN | 94.28 | 89.80 | 92.63 | 91.19 |
Proposed | 99.1 | 99 | 98.8 | 99 |
Abbreviations: ANN, artificial neural network; ASD, autism spectrum disorder; DT, decision tree; KNN, k-nearest neighbor; NB, naive Bayes; RF, random forest; SVM, support vector machine.
Comparison based on AUC, kappa, and log loss value for the combined ASD dataset.
Methods | AUC | Kappa | Log loss value |
---|---|---|---|
NB | 0.93 | 0.84 | 2.42 |
KNN | 0.81 | 0.61 | 6.068 |
SVM | 0.89 | 0.80 | 3.033 |
RF | 0.91 | 0.84 | 2.427 |
DT | 0.90 | 0.81 | 2.913 |
XGB | 0.90 | 0.82 | 2.669 |
LR | 0.88 | 0.80 | 3.034 |
ANN | 0.93 | 0.86 | 2.063 |
Proposed | 0.99 | 0.99 | 0.154 |
Abbreviations: ANN, artificial neural network; ASD, autism spectrum disorder; AUC, area under the curve; DT, decision tree; KNN, k-nearest neighbor; NB, naive Bayes; RF, random forest; SVM, support vector machine.

Overall performance analysis using the combined ASD dataset. Abbreviations: ANN, artificial neural network; ASD, autism spectrum disorder; DT, decision tree; KNN, k-nearest neighbor; NB, naive Bayes; RF, random forest; SVM, support vector machine.
CONCLUSION
The study has proposed a model to identify and diagnose ASD using advanced deep learning methods. It addressed the existing limitations by presenting a model for detecting ASD symptoms over time. The ASD is typically a long-term neurological condition. ASD needs to be detected early, quickly, precisely, and successfully in order to receive the appropriate medical therapy. Rather than relying on more expensive and time-consuming traditional methods, data mining techniques have been employed to identify the damaged party. The conventional procedures that were created and used in the past have been expensive and time-consuming. While some studies in this field have been undertaken with success in evaluating ASD, these investigations have not yet reached exceptional results at a lower cost. The work’s original contribution is the formulation and application of the distinctive and one-of-a-kind AHRML framework for the diagnosis and classification of ASD from supplied medical data. The most important components of the suggested AHRML system are feature optimization, performance estimates, ASD prediction, and data preprocessing. In this work, the public open-source dataset is first considered for system development and validation. The normalized data are obtained via standard preprocessing procedures such as attribute balancing, scaling, outlier removal, and missing value imputation, which are applied after the data are acquired. Consequently, this lowers the training and validation time as well as the classification computing burden by using the preprocessed data to identify the important and necessary features. The AHHO, a novel hybrid optimization technique, is used to achieve this. The obtained attributes are then fed into an intelligence-based classifier called ARTM, which employs detection rate and processing speed to accurately identify ASD. Finally, a comprehensive performance analysis and simulation are conducted to examine the outcomes of the proposed AHRML system. When compared to all current approaches, the total analysis concludes that the proposed AHRML offers better detection performance and outcomes because of appropriate feature reduction, training, and testing operations. The results improve our comprehension of the intricate connections between behavioral characteristics and ASD, opening up possibilities for future study in uncovering the fundamental causes of the illness. Deep learning models for ASD diagnosis are useful tools for doctors, providing help in diagnosing and planning therapy. These models enhance diagnostic accuracy and consistency by using objective and standardized tests that work with clinical experience. The effectiveness of deep learning models in detecting ASD is highly dependent on the quality and amount of annotated data that are accessible for training purposes. The restricted availability of extensive, top-notch datasets could limit the ability of models to generalize and perform well.