Processing math: 100%
437
views
0
recommends
+1 Recommend
1 collections
    0
    shares

      King Salman Center for Disability Research is pleased to invite you to submit your scientific research to the Journal of Disability Research. JDR contributes to the Center's strategy to maximize the impact of the field, by supporting and publishing scientific research on disability and related issues, which positively affect the level of services, rehabilitation, and care for individuals with disabilities.
      JDR is an Open Access scientific journal that takes the lead in covering disability research in all areas of health and society at the regional and international level.

      scite_
      0
      0
      0
      0
      Smart Citations
      0
      0
      0
      0
      Citing PublicationsSupportingMentioningContrasting
      View Citations

      See how this article has been cited at scite.ai

      scite shows how a scientific paper has been cited by providing the context of the citation, a classification describing whether it supports, mentions, or contrasts the cited claim, and a label indicating in which section the citation was made.

       
      • Record: found
      • Abstract: found
      • Article: found
      Is Open Access

      Enhancing Navigation and Object Recognition for Visually Impaired Individuals: A Gradient Support Vector Boosting-based Crossover Golden Jackal Algorithm Approach

      Published
      research-article
      Bookmark

            Abstract

            On a global scale, individuals with vision impairments encounter various limitations when it comes to moving around and finding their way independently. Their daily activities are impeded by their limited understanding of their environment while moving about both indoors and outside, where situations are constantly changing. Recent technological breakthroughs have made it possible to create several electronic devices that help visually impaired and disabled people with navigation. These devices encompass navigation systems, obstacle avoidance systems, object localization devices, and orientation assistance systems. They are designed to enhance or substitute conventional aids like guide dogs and white canes. This research work proposes a solution based on the gradient support vector boosting-based crossover golden jackal (GSB-CGJ) algorithm, which integrates various assistive technologies focused on navigation and object recognition, providing intelligent feedback to the user. The developed model focuses on guiding visually impaired individuals, preventing unwanted collisions with obstacles, and generating active feedback. The proposed method consists of three distinct phases. In the input phase, images are acquired from the Image and Video Dataset for Visually Impaired using Intel RealSense Camera. The next stage entails object recognition, which is efficiently carried out using the GSB-CGJ algorithm. The hyperparameters of the support vector machine and adaptive boosting methods are optimized using the golden jackal optimization method, enhancing object recognition ability. At the end, the output phase delivers feedback to the user. The experimental and assessment results validate that the model demonstrates high accuracy in recognizing objects and precision in localizing them. This approach effectively delivers remarkable real-time implementation capability, showcasing better adaptability and reliability while reducing execution time.

            Main article text

            INTRODUCTION

            According to the World Health Organization (WHO), about 15% of the total world population has several disabilities such as visual impairment, hearing loss, and locomotor issues. WHO also reveals that there will be an increase in this number because of the gradual increase in the chronic health conditions of aged people (de Freitas et al., 2022). To support these types of disabled people, assistive technology (AT) is used (Kbar et al., 2016, 2017). AT provides various services particularly for aged people to compensate for their functional difficulties. Active participation in life, which includes minimizing the necessity of caregivers and reducing the cost of health care, is also supported by AT. The eyeglasses, prostheses, wheeling chairs, and hearing aids are the few examples of the assistive products (Abdi et al., 2021). Hence, life quality is improved by the technology-based smart and intelligent assistant agents. There are several principles that are stated by “The United Nations Principles for Older Persons” to improve the life quality of the aged people. The major objective of these principles is to focus on the importance of developing assistive surroundings that are highly utilized for the improvement of elder people in terms of mental and physical well-being (Thakur and Han, 2021).

            People with several disabilities are assisted by AT in different ways that comprise allowing access to data and education, enhancing the ability to communicate, and performing daily tasks. It also helps an individual to evolve and enhance their skills as well as social inclusion (Ramirez, 2023). Based on individual characteristics as well as personal aspirations, the merits of the assistive products are varied from one person to another. The complication in the AT services leads to the implementation process of the AT services being more complex (Baucas et al., 2021). Various studies state that the efficiency of the information and communication technology (ICT)-based AT is established for the larger set of people with cognitive dysfunction that includes autism spectrum disorder, intellectual disability, acquired brain injury, and developmental disability. It is revealed that ICT-based assistive products such as digital voice recorders and personal digital assistants are employed for individuals with impaired cognition (Brandt et al., 2020). The progress of technological advancements makes the AT more cost-effective by the availability of important ATs, and they are also integrated with the social innovation setting (Nierling and Maia, 2020).

            Artificial intelligence (AI)-based solutions are gaining popularity in several fields (Abidi et al., 2022a, b). The ATs of the people who are visually impaired are not easily surrendered without minimizing their independence. The functional needs of the individuals are satisfied by the ATs without causing any social embarrassment. The obstacles are detected by the white cane, and it is also referred to as the symbol by which the blindness of the individuals is communicated. Thus, the white canes are not often utilized by the visually impaired individuals, and the attraction of unnecessary attraction as well as the stigmatization is prevented in which the contact is limited with the individuals who are not blind (dos Santos et al., 2022). The AI-driven solution is exhibited by the OrCam devices that are utilized for visually impaired people in which the individuals can read easily. Thus, AT ranges from the Braille display as well as the books to text-to-speech platforms or wheelchairs (Fiorini et al., 2021). Thus, it is revealed that AT is utilized by people in all views of their daily life including fitness, education, and other daily activities (Söderström et al., 2021). In this work, the gradient support vector boosting-based crossover golden jackal (GSB-CGJ) algorithm is proposed for visual recognition using ATs. The main contributions of the paper are as follows:

            Novel architecture: A robust architecture with gradient boosting and a support vector machine (SVM) is created for effective visual recognition. It also uses the best visual impairment dataset for easy processing.

            Enhanced recognition: The improvement in recognition is provided by the proposed GSB-CGJ architecture. It processes the data provided efficiently and provides accurate recognition.

            Optimization with golden jackal algorithm: The crossover-integrated golden jackal optimization (GJO) fine-tunes the overall performance of the architecture by selecting the optimal hyperparameters.

            The remaining sections of the paper are organized as follows: the following section presents the literature review conducted on various works related to the topic of ATs for patients with disabilities; a detailed description of the proposed methodology is given in the Methodology section; the next section provides a detailed explanation of the results and discussions, and the paper is finally concluded with the Conclusions section.

            LITERATURE REVIEW

            Researchers around the world are working with an aim to make the life of people with disabilities better. With the advent of AI, it is becoming easier, and numerous solutions have been developed. For the paralyzed patients’ revival in the connected communities, Jacob et al. (2021) developed an exoskeleton mechanism with the integration of Internet of Things (IoT) and AI. There was an increase in the number of paralyzed people across the world due to the increase in nervous disorders, spinal cord defects, injuries, and stroke. One of the most popular revivals for the paralyzed patients was the exoskeletons. Constant control, adaptability, and instant response were the challenges faced by conventional exoskeletons. To solve the aforementioned challenges, an AI-enabled smart and lightweight exoskeleton system (AI-SES) mechanism was proposed in this article. This mechanism used the data collected by multiple sensors. From the evaluation results, it was found that the proposed mechanism eliminated the challenges faced by the traditional mechanisms and performed well in the navigation functioning. Only a limited number of technologies were explored in this work.

            For the revival of neurologically ill individuals, Casey et al. (2021) established a robotic arm functioning with brain–computer interface (BCI). People with neurological disorders were highly dependent on others for their daily activities. Hence, an assistive robotic arm with BCI was a preferred choice. In this work, the arm was controlled with the readings obtained from electromyogram (EMG) and electroencephalogram (EEG). From the simulations done, it was observed that EMG exhibited superior performance and this arm was applied to healthcare use. Various performance enhancements to EEG were not explored in this work.

            Based on eye tracking, Paing et al. (2022) developed an assistive mechanism for disabled patients. The assistive mechanisms were used on a large scale for the betterment of disabled individuals. An eye-tracking-based assistive mechanism was proposed in this article to help disabled patients in daily activities. This mechanism was divided into two sections: mobile and stationary assistive mechanisms. The disabled patients were capable of controlling the wheelchairs and home appliances, and sending messages to the caretakers, utilizing their eye movement. From the experiments conducted, it was found that the proposed mechanism showed better performance in assisting disabled patients. The high cost of the mechanism was the limitation of this work.

            To incorporate multimodal interfaces, IoT devices, and robotic systems in the medical sector, Brunete et al. (2021) established an intelligent assistive framework. In this work, a multimodal and intuitive interaction mechanism that assists individuals with disabilities was presented. Various home appliances and supportive robots were controlled using this smart assistive framework. For controlling the interactive mechanism, a tablet device was provided in the framework. From the evaluations done, it was found that the usage of mobile phones and eye-based control was preferred among the users.

            Based on designing intelligent assistive devices, Wang et al. (2022) illustrated that smart wearable sensor devices help in promoting iteration and digitization of healthcare components used in better manufacturing systems used in rehabilitation assistive devices (RADs). The primary objective of this method was to analyze and improve the item life cycle of the executives and activity system. Research on personalized product-service systems (PSS) for RADs for special populations was scarce, limiting their development. To provide better healthcare services, this study provides a basic overview of IoT-based RAD and PSS-based methods. A few benefits include better efficiency, data availability, and optimization. Some of the drawbacks include complexity issues, high costs, and technical challenges.

            Bouteraa (2021) developed a support system for blind and visually impaired people (BVIP). The primary objective of this approach was to enhance safety for BVIP using sensor data. The navigation system utilizes wearable obstacle detection systems controlled by installed controller settings, employing robot operating system planning. It provides a blended mixed haptic-voice message to train visually impaired persons. A few benefits include better scalability, enhanced tracking capability, and good detection.

            Malcangi (2021) illustrated that rarely affected diseases provide greater attention in healthcare services due to their impact on nations’ economies. The primary objective of this method is to offer various treatments for affected individuals, as neuromotor cell impairments in the brain can lead to serious handicaps. Recent AI advances facilitate early neurodegenerative disease prediction using cutting-edge computational technologies, including advanced artificial neural networks with rapid training and online learning facilities. A few downsides include affecting the overall acceptance of privacy and raising data security issues.

            Ciubotaru et al. (2023) developed a preventive method for detecting various health-related effects in the field of research and development. The major goal of this approach was to enhance IoT-based systems in finding frailty. This study helped frame the overall methodology for analyzing various data protocols, advocating for an interdisciplinary approach to interpreting wearable sensor data. It emphasizes visual portrayals and AI models for exhaustive examination. Some of the benefits include better classification, assessability, and good computational analysis. However, there are drawbacks that include affecting system performance, privacy, and security concerns.

            Said et al. (2023) presented an obstacle detection system for visually impaired people using deep learning technique specifically modified neural architecture search. The results revealed that the proposed model offered a 2.6% increase in average precision with acceptable computation complexity. Al-Zahrani and Al-Baity also presented a deep learning-based method for object detection. The major contribution from their side is that they proposed a system with Arabic annotation instead of English. The model was based on the Mask region-based convolutional neural network algorithm. The results reported 83.9% accuracy for the model. Tarik et al. (2023) proposed a low cost device for object detection for visually impaired called I-CANe. It provides real-time voice feedback also. Islam et al. (2023) proposed an objection detection system based on TensorFlow object detection application programming interface and SSDLite MobileNetV2. Moreover, gradient particle swarm optimization technique was used in this work to optimize the final layers. The results reported a detection accuracy of 88.89%, a precision of 0.892, and a recall of 0.89.

            Research gaps

            The usage of AI and sensors in visual recognition for AT among individuals with disabilities yields more benefits compared to other techniques. However, existing works have various limitations such as high implementation costs, insufficient exploration of advanced technologies, lack of comprehensive datasets, and increased execution time. Nature-inspired algorithms have proven to be efficient in several fields (Abidi et al., 2020a, b). To overcome these limitations, this paper proposes the nature-inspired algorithm GSB-CGJ. Based on the literature, the following research gaps are highlighted:

            Execution time: Execution time is a crucial parameter for ensuring optimal performance in any mechanism. The GSB-CGJ framework employs proper optimization techniques to fine-tune the performance of the architecture and reduce the execution time.

            Recognition accuracy: The accuracy is the main parameter that defines the effectiveness of any framework. The existing works show reduced accuracy due to insufficient datasets. The GSB-CGJ presents a clean dataset and provides better accuracy.

            Cost of implementation: The high cost of implementation is the parameter that affects the easy availability of any architecture. GSB-CGJ has a simple architecture and reduced execution time, which make the proposed technique cost-effective.

            METHODOLOGY

            The workflow of the proposed methodology is shown in Figure 1. Various phases in the process of visual recognition utilizing ATs are presented as follows:

            Figure 1:

            Workflow of the proposed methodology.

            Phase I: the input phase

            Visually impaired individuals determine the class of an object based on parameters such as its shape and size, which they identify using their sense of touch. To acquire feedback on the object’s content, sound is used. The dataset provided as input is the Image and Video Dataset for Visually Impaired using Intel RealSense Camera.

            Phase II: the visual recognition phase using assistive technologies

            This phase involves identifying specific visual features from the object by analyzing images from the video stream and the information provided by visually impaired individuals. To execute this task, various image descriptors are analyzed, among which the bag-of-features method demonstrates higher performance in recognition.

            Gradient boosting

            To solve regression-based tasks, the boosting algorithm plays a major role. T = {fr , br } indicating the training dataset is boosted by gradient approach (Douiba et al., 2023). The instances f are plotted based on their predicted values W(f) of the loss function, with W(f) representing their respective output values. The weighted sum of functions present in the gradient boosting is shown based on the following:

            (1) Wn(f)=Wn1(f)+σnsn(f),

            where the nth function sn (f)’s weight is denoted as σn by developing various iterative processes that help in constructing the approximation. The following equation illustrates the constant approximation:

            (2) W0(f)=argminδMr=1O(br,δ).

            The predicate decreed in the subsequent approach is shown below:

            (3) (σn,sn(f))=argminσ,sMr=1O(br,Wn1,(fr)+δs(fr)).

            Whereas T{fr,W(f)}Mr=1 helps in training every model, the subsequent equation illustrates the computation of pseudo-residuals inr :

            (4) inr=[O(br,W(f))W(f)]W(f)=Wn1(f).

            The line search optimization issues are solved by further evaluating σn .

            Support vector machine

            An SVM algorithm is applied, which employs both regression and utilization effort. It treats every point as the points present in multi-dimensional space, not entirely determined by the features of the data (Wang et al., 2023). Classification is available by detecting the distinct hyperplanes by separating various groups of scattering data points. The following equation indicates the SVM process:

            (5) 12wUw+TQj=1ξj

            Based on the requirements,

            (6) zj(wUφ(yj)+a)1ξjandξj0,j=1,Q,

            where T signifies the constant capacity by utilizing vector coefficients denoted as w, managing non-separable data indicated as ξ. Here, the maximum iterations D present in the index j. The class labels and the independent variables are indicated as zj and yj , respectively. To avoid more penalties for errors, increased T values lead to a narrower tolerance.

            Golden jackal optimization

            The golden jackal algorithm is referred to as the swarm intelligence algorithm that was developed by Chopra and Mohsin Ansari (2022). Usually, the golden jackals hunt with both females and males. The hunting behavior of the golden jackal consists of three stages: searching and moving closer to the prey, encircling and irritating the prey, and pouncing toward the prey. In this section, GJO is described and explained step by step. The initial process is to build the solution population, expressed in the equation below:

            (7) Aj=lb+t×(ublb),j=1,2,,M.

            From the above equation, the random value is represented by t[0,1] , whereas lb and ub represent parameter limits in the search space. The for each Aj , Aj = 1,2,…,M, the fitness value is evaluated to find effective male and female jackal solutions. The matrix of prey is constructed by utilizing the following equation:

            (8) Prey=[A11A12A1eA21A22A2eAM1AM2AMe]

            Exploration stage

            The process of hunting is referred to as the model of exploration, and it is based on the female jackal (AGN ) as well as the male jackal (AN ) by updating their position in accordance with (AN ); this process is expressed by the equations given below.

            (9) A1(u)=AN(u)F×|AN(u)tk×prey (u)|

            (10) A2(u)=AGN(u)F×|AGN(u)tk×prey(u)|

            From the above equations, AN and AGN are updated to A1 and A2 , respectively, with respect to the uth iteration. The prey’s energy is represented by F and updated using the equation below:

            (11) F=F0×F1,

            where F 0 and F 1 denote the initial and decreasing energy values from the above equation, respectively. Thus, the values of F 0 and F 1 are updated by utilizing the following equations:

            (12) F0=2×t1

            (13) F1=d1×(1(u/U))

            From the above equations, the maximum number of generations is represented by U, whereas t1 represents the random value that is developed based on the Levy distribution by utilizing the equation below.

            (14) t1=0.05×Levy.

            The jackal positions are updated by applying the following equation:

            (15) A(u+1)=A1(u)+A2(u)2×levy, levy=m|w×η||τ|1/δ,η=(Γ(1+δ)×sin(φδ2)Γ(1+δ2)×δ×2(δ12)).

            From the above equation, the random numbers are represented by w and τ, whereas m = 0.01 and δ = 1.5 are referred to as constants.

            Exploitation stage

            In the exploitation phase, GJO provides feasible solutions within the areas that are explored. The hunting of the male and female jackals is represented by this process in GJO, and it is derived by utilizing the equations below.

            (16) A1(u)=AN(u)F×|t1×AN(u)prey(u)|

            (17) A2(u)=AGN(u)F×|t2×AGN(u)prey(u)|

            Crossover strategy

            The hybridization of the two parent parameters is acquired by employing the crossover operator, and the offspring is attained. It consists of both the parent solution features. The following equation is the crossover that is employed in the present work:

            (18) wk={ykiifciCRzikotherwise.

            From the above equation, the crossover rate is indicated by CR , whereas the two parent solutions are indicated by zik and yki . The probability of crossover is 0.3 in the present work.

            The Proposed GSB-CGJ algorithm

            AT facilitates active participation in daily life, reducing the reliance on caregivers and lowering healthcare costs. The effectiveness of assistive products varies depending on individual characteristics and personal goals. In order to improve the overall performance of assisting visually challenged individuals in object identification, AI-based assistive mechanisms are employed.

            The hybrid model is applied in which more than two distinct and relevant models are employed and the results are combined to form a single value. This paper proposes the GSB-CGJ algorithm, combined with gradient boosting and SVM, to assist visually challenged people in identifying objects by utilizing the AI-based assistive mechanisms. Utilizing the proposed GSB-CGJ helps visually challenged people identify objects, which is improved from the acquired datasets. Adding the diverse models with few hypotheses with respect to the class labels is not seen in other models. Therefore, assisting visually challenged people in identifying objects is performed with high accuracy. It is noted that selecting effective and optimal hyperparameters of GSB is important in assisting visually challenged people in identifying objects. Thus, CGJ is utilized in this paper to optimize the parameters of the GSB algorithm. Then, the transform operator and step size value are multiplied to balance both the local search and global searches. By applying this method, an effective global solution is acquired. Therefore, the attained global solution has a high optimization interval time and a minimum step size value. On the other hand, the CGJ algorithm is subjected to several convergence problems and is easily trapped by the local optima problems. The crossover strategy is added to overcome these limitations to alter the CGJ optimization. The hyperparameters of the GSB algorithm are tuned using the CGJ algorithm to rectify all these difficulties by selecting the best candidate solution and optimal solution. The CGJ algorithm tunes the hyperparameters of the GSB algorithm to overcome all these limitations by selecting the most effective candidate solution. By utilizing the advantages of the CGJ algorithm, the tuning of hyperparameters is carried out. Figure 2 illustrates the flowchart representation of the proposed GSB-CGJ algorithm.

            Figure 2:

            Proposed GSB-CGJ algorithm. Abbreviation: GSB-CGJ, gradient support vector boosting-based crossover golden jackal.

            Phase III: the output phase

            Finally, in the output phase, the output obtained by the visual recognition phase is presented as feedback in this phase.

            The proposed algorithm, GSB-CGJ, was chosen for several reasons. First, it combines the advantages of SVM and gradient boosting, which are both well-established and widely used machine learning techniques. SVMs are known for their ability to handle high-dimensional data and produce accurate classification results, while gradient boosting is effective at improving the performance of weak learners through ensemble learning. By combining these two methods, GSB-CGJ aims to achieve both high accuracy and robustness in object recognition.

            Additionally, the crossover GJO method was chosen for hyperparameter optimization due to its efficiency in finding optimal solutions in complex search spaces. This method is inspired by the hunting behavior of golden jackals, which involves a combination of searching, encircling, and pouncing toward prey. This approach allows for the efficient tuning of hyperparameters in the GSB-CGJ algorithm, further enhancing its performance.

            RESULTS AND DISCUSSION

            The developed architecture of GSB-CGJ for supporting visual recognition is analyzed in this section for its effectiveness. Here, some evaluation measures are considered for this, and certain existing disability assistance strategies such as evolving fuzzy neural network (EFuNN), AI-IoT-SES, and SVM are used for comparison. The experiments were conducted using the MATLAB platform, and the results are briefly discussed below.

            System configuration

            The experiments were conducted using a 12th generation Intel Core i7 processor in an HP laptop with 8 GB RAM (random access memory), 512 GB SSD (solid state disk) hard disk memory space, and Windows 11 operating system. The experiments were performed in the MATLAB simulation platform.

            Parameter setting

            This paper develops a GSB-CGJ algorithm for recognizing objects visually using some ATs. Here, the hyperparameters of the gradient boosting and SVM are optimized using the golden jackal algorithm. Table 1 provides the hyperparameters of these algorithms and their values.

            Table 1:

            Hyperparameters of the GSB-CGJ method.

            MethodsHyperparameterValues
            Gradient boostingEstimator quantity100
            Rate of learning1
            Maximum depth10
            Maximum number of features7
            Maximum samples to split20
            Maximum number of leaf samples20
            SVMSVM typeOne class
            Kernal typeLinear
            Gamma0.506
            C312.5
            Iteration100
            Tolerance error0.00001
            GJOSize of the population10
            Iteration100

            Abbreviations: GJO, golden jackal optimization; GSB-CGJ, gradient support vector boosting-based crossover golden jackal; SVM, support vector machine.

            Dataset description

            The Image and Video Dataset for Visually Impaired using Intel RealSense Camera constitutes two videos (Mehta and Mehendale, 2023). Both videos are recorded using a 1080p Intel RealSense depth camera: one is captured by blindfolding and the other without blindfolding. The dataset aims to utilize this for machine vision to permit them to operate operations like instance segmentation. To mimic the highly careful gait of a visually impaired person, the captured videos and extracted images with a blindfold were utilized, which supports the generation of the dataset. For instance, the walking activities of a human were recorded by fixing a hat on the head of a blindfolded human. There are 25 videos in the dataset with the frame rate of 30 fps. The image resolution was 1920 × 1080 pixels. For video chunk selection, parameters such as duration (e.g., 10 seconds per chunk) and overlap (e.g., 50% overlap between chunks) were used. The GSB-CGJ algorithm utilized pixels from specific regions of interest within the images, as determined during preprocessing.

            Performance measures

            The model’s performance was evaluated using several common metrics such as accuracy, precision, recall, F1-score, and area under ROC curve (AUC), providing a comprehensive assessment of its ability to recognize objects and its precision in localizing them. The mathematical modeling of these metrics is explained below.

            Accuracy

            It indicates the values of the ratio of the total correct predictions over the total predictions made by the model. The formula is given below:

            (19) Ac=αT+βTαT+βt+αF+βF.

            Precision

            The positive prediction accuracy is computed as a precision value. The precision is computed using the formula given below:

            (20) Pr=αTαT+αF.

            Recall

            The capacity to capture all related instances is indicated by the term recall. It is specified as below:

            (21) Sn=αTαT+βF

            F1-score

            The harmonic mean of precision and recall are referred to as the F1-score. The formula is given below:

            (22) F1scr=2×pre×senpre+sen.

            AUC curve

            The AUC value is measured by plotting the true-positive rate (TPR) against the false-positive rate (FPR) values.

            Performance analysis

            The performance of the GSB-CGJ model is assessed to measure the effectiveness of visual recognition, and the results are given in Table 2. Here, the accuracy of the model scores is 94.55%. The precision, recall, and F1-score of the model are 92%, 91.45%, and 94.75%, respectively. Higher performance scores in all the measures indicate higher performance. Based on the figure, it is confirmed that the developed model has admirable performance.

            Table 2:

            The performance of the GSB-CGJ model.

            Accuracy (%)Precision (%)Recall (%)F1-score (%)AUC (%)
            94.5592.591.4594.7594

            Abbreviations: AUC, area under receiver operating characteristic curve; GSB-CGJ, gradient support vector boosting-based crossover golden jackal.

            Comparative analysis

            The accuracy of the GSB-CGJ model is compared with that of some existing disability assistant methods, shown in Figure 3. The accuracy of the model described here is the effectiveness of the system in detecting objects and recognizing them in indoor and outdoor environments for fully visually impaired patients. This figure shows that the developed model has the highest accuracy with 94.55%, which is superior to that of all the existing methods.

            Figure 3:

            Analysis of visual recognition accuracy of the GSB-CGJ method. Abbreviations: GSB-CGJ, gradient support vector boosting-based crossover golden jackal; IoT, Internet of Things; SES, smart and lightweight exoskeleton system; SVM, support vector machine.

            Figure 4 analyzes the precision power of the developed GSB-CGJ method by comparing it with certain baselining methods. The precision in this context measures the percentage of the positive prediction generated by the methodology. A higher precision score shows higher visual recognition effectiveness and automatically shows the highest reliability of the model. The observation from the figure implies that the developed model has a higher level of visual recognition precision performance, with a score of 92.5% compared to other underlying methods.

            Figure 4:

            Comparing the precision performance of all disability-assistive methods. Abbreviations: GSB-CGJ, gradient support vector boosting-based crossover golden jackal; IoT, Internet of Things; SES, smart and lightweight exoskeleton system; SVM, support vector machine.

            The recall performance produced by the developed GSB-CGJ visual recognition model is compared with that of other methods as shown in Figure 5. Here, recall indicates the whole related instances which are also referred to as the TPR. Greater performance is indicated by a higher recall score. This figure shows that the developed model has the highest recall value with a score of 91.45%. Figure 6 shows the F1-score of the proposed GSB-CGJ methodology. F1-score shows the mean values of recall and precision. Thus the greater values of precision and recall which simultaneously indicates that the higher F1-score performance. From the figure, it is observed that the proposed visual recognition method has a higher level of F1-score performance compared to other existing methods. The area beneath the receiver operating characteristic (ROC) curve is another measure that is utilized to analyze the visual recognition potential by comparing it with other underlying techniques. The range of the AUC score should lie between 0.5 and 1 for the productive strategy. This analysis demonstrated that the developed GSB-CGJ model has a superior level of performance with a higher AUC score of 94% (Figure 7).

            Figure 5:

            Recall the performance of the developed GSB-CGJ visual recognition strategy. Abbreviations: GSB-CGJ, gradient support vector boosting-based crossover golden jackal; IoT, Internet of Things; SES, smart and lightweight exoskeleton system; SVM, support vector machine.

            Figure 6:

            Analysis of the F1-score of the various disability-assistive methodologies. Abbreviations: GSB-CGJ, gradient support vector boosting-based crossover golden jackal; IoT, Internet of Things; SES, smart and lightweight exoskeleton system; SVM, support vector machine.

            Figure 7:

            AUC score analysis of the proposed GSB-CGJ methodology. Abbreviations: AUC, area under receiver operating characteristic curve; GSB-CGJ, gradient support vector boosting-based crossover golden jackal; IoT, Internet of Things; SES, smart and lightweight exoskeleton system; SVM, support vector machine.

            The errors made by the developed GSB-CGJ methodology are analyzed using two metrics: FPR and false-negative rate (FNR). The FPR is shown in Figure 8 and the FNR is shown in Figure 9. The lesser error values, such as the low FNR and FPR scores compared to all models, posed excellent performance. Figures 8 and 9 show that the developed model has the lowest scores of FPR and FNR, which indicate excellent visually impaired assistive performance.

            Figure 8:

            Comparison of the FPR values of disability-assistive techniques. Abbreviations: FPR, false-positive rate; GSB-CGJ, gradient support vector boosting-based crossover golden jackal; IoT, Internet of Things; SES, smart and lightweight exoskeleton system; SVM, support vector machine.

            Figure 9:

            Comparison of the FNR values of all disability-assistive strategies. Abbreviations: FNR, false-negative rate; GSB-CGJ, gradient support vector boosting-based crossover golden jackal; IoT, Internet of Things; SES, smart and lightweight exoskeleton system; SVM, support vector machine.

            The time taken by the disability-assistive methods to finish their tasks successfully is referred to as the execution time. The time for execution is influenced by multiple factors, such as the algorithm’s complexity, the input’s size, etc. The proposed model has excellent visual detection performance and less complexity. This automatically decreases the processing time. The execution times of various methods are shown in Figure 10.

            Figure 10:

            Comparison of the execution time of various disorder-assistive methods. Abbreviations: GSB-CGJ, gradient support vector boosting-based crossover golden jackal; IoT, Internet of Things; SES, smart and lightweight exoskeleton system; SVM, support vector machine.

            The performance of the GSB-CGJ model is shown in Table 3. Here, some requirements that should be satisfied by the AI-based ATs for visually impaired people are taken. Some existing strategies are also checked to determine whether they satisfy these requirements. The analysis showed that the developed model has an excellent assistive capacity to support visually impaired persons compared to the other existing disability-assistive AI strategies. In this table, H denotes high, M denotes moderate, and L denotes low. Furthermore, Y specifies yes, which is the availability of the corresponding feature, and N means no, indicates the absence of this feature.

            Table 3:

            Comparison of various disability-assistive strategies.

            MethodsAdaptabilityReliabilityFlexibilityReal-time implementation
            GSB-CGJHHYY
            EFuNNMMYN
            AI-IoT-SESMMNY
            SVMLMNY

            Abbreviations: AI, artificial intelligence; GSB-CGJ, gradient support vector boosting-based crossover golden jackal; H, high; IoT, Internet of Things; L, low; M, moderate; N, no; SES, smart and lightweight exoskeleton system; SVM, support vector machine; Y, yes.

            CONCLUSIONS

            The rapidly moving life in the recent decade has caused visually impaired people to face multiple difficulties due to the crowding of the obstacles in the surroundings. Thus, it is necessary for them to depend on external assistance offered by humans, dogs, and multiple advanced electronic support systems to make decisions. A large number of blind individuals seek the development of assistive devices to support and facilitate the normal lives of visually impaired people. Many existing strategies have the disadvantage of detecting obstacles at ground or waist levels, but they do not have the capacity to detect them simultaneously. Therefore, this study develops GSB-CGJ to visually recognize objects with the assistance of various assistive strategies. The developed method gained input images from the Image and Video Dataset for Visually Impaired using an Intel RealSense Camera. Finally, GSB-CGJ recognizes the objects. The ability of the model to be analyzed with some common metrics and three already available assistive AI strategies for disabled persons, such as EFuNN, AI-IoT-SES, and SVM, is compared. The results showed the developed model’s higher level of performance with higher visual recognition capability and greater adaptability, trustworthiness, and object localization precision for detecting the obstacles precisely. It lessened the time and cost considerably, which is another advantage.

            Once trained, the model can be deployed on a device equipped with a camera, such as a smartphone or wearable device. In real time, the device captures images of the user’s surroundings and processes them using the trained model to recognize objects and provide feedback to the user. The feedback may include audio cues, vibrations, or text-to-speech output, depending on the user’s preferences. The practical implementation also involves considerations for scalability, computational efficiency, and user interface design. The model should be lightweight enough to run on resource-constrained devices while still maintaining high accuracy. Additionally, the user interface should be intuitive and accessible to visually impaired individuals, allowing them to easily interact with the system and receive meaningful feedback about their environment.

            While our proposed approach shows promising results, there are several limitations to consider. First, the performance of the model may vary depending on the quality and consistency of the input data. Variations in lighting conditions, object textures, and background clutter can impact the accuracy of object recognition. Additionally, the model may struggle with detecting objects that are occluded or partially obscured. Furthermore, the computational resources required to train and deploy the model may be prohibitive for some users, particularly those in resource-constrained environments. Finally, the model’s performance may degrade when applied to real-world scenarios that differ significantly from the training data. Despite these limitations, our approach represents an important step toward improving navigation and object recognition for visually impaired individuals.

            In the future, this model will be extended to include multiple data modalities such as video, audio, etc. Furthermore, the model will be examined with 3D images instead of 2D to evaluate precision further.

            ACKNOWLEDGMENTS

            The authors extend their appreciation to the King Salman Center for Disability Research for funding this work through Research Group no. KSRG-2023-451.

            COMPETING INTEREST

            The authors declare no conflicts of interest in association with the present study.

            REFERENCES

            1. Abdi S, Kitsara I, Hawley MS, de Witte LP. 2021. Emerging technologies and their potential for generating new assistive technologies. Assist. Technol. Vol. 33:17–26. [Cross Ref]

            2. Abidi MH, Alkhalefah H, Mohammed MK, Umer U, Qudeiri JEA. 2020a. Optimal scheduling of flexible manufacturing system using improved lion-based hybrid machine learning approach. IEEE Access. Vol. 8:96088–96114. [Cross Ref]

            3. Abidi MH, Umer U, Mohammed MK, Aboudaif MK, Alkhalefah H. 2020b. Automated maintenance data classification using recurrent neural network: enhancement by spotted hyena-based whale optimization. Mathematics. Vol. 8:2008

            4. Abidi MH, Alkhalefah H, Umer U. 2022a. Fuzzy harmony search based optimal control strategy for wireless cyber physical system with industry 4.0. J. Intell. Manuf. Vol. 33:1795–1812. [Cross Ref]

            5. Abidi MH, Mohammed MK, Alkhalefah H. 2022b. Predictive maintenance planning for industry 4.0 using machine learning for sustainable manufacturing. Sustainability. Vol. 14:3387

            6. Baucas MJ, Spachos P, Gregori S. 2021. Internet-of-things devices and assistive technologies for health care: applications, challenges, and opportunities. IEEE Signal Process. Mag. Vol. 38:65–77. [Cross Ref]

            7. Bouteraa Y. 2021. Design and development of a wearable assistive device integrating a fuzzy decision support system for blind and visually impaired people. Micromachines (Basel). Vol. 12:1082[Cross Ref]

            8. Brandt Å, Jensen MP, Søberg MS, Andersen SD, Sund T. 2020. Information and communication technology-based assistive technology to compensate for impaired cognition in everyday life: a systematic review. Disabil. Rehabil. Assist. Technol. Vol. 15:810–824. [Cross Ref]

            9. Brunete A, Gambao E, Hernando M, Cedazo R. 2021. Smart assistive architecture for the integration of IoT devices, robotic systems, and multimodal interfaces in healthcare environments. Sensors. Vol. 21:2212

            10. Casey A, Azhar H, Grzes M, Sakel M. 2021. BCI controlled robotic arm as assistance to the rehabilitation of neurologically disabled patients. Disabil. Rehabil. Assist. Technol. Vol. 16:525–537. [Cross Ref]

            11. Chopra N, Mohsin Ansari M. 2022. Golden jackal optimization: a novel nature-inspired optimizer for engineering applications. Expert Syst. Appl. Vol. 198:116924. [Cross Ref]

            12. Ciubotaru B-I, Sasu G-V, Goga N, Vasilățeanu A, Marin I, Goga M, et al.. 2023. Prototype results of an internet of things system using wearables and artificial intelligence for the detection of frailty in elderly people. Appl. Sci. Vol. 13:8702

            13. de Freitas MP, Piai VA, Farias RH, Fernandes AMR, de Moraes Rossetto AG, Leithardt VR. 2022. Artificial intelligence of things applied to assistive technology: a systematic literature review. Sensors. Vol. 22:8531[Cross Ref]

            14. dos Santos ADP, Ferrari ALM, Medola FO, Sandnes FE. 2022. Aesthetics and the perceived stigma of assistive technology for visual impairment. Disabil. Rehabil. Assist. Technol. Vol. 17:152–158. [Cross Ref]

            15. Douiba M, Benkirane S, Guezzaz A, Azrour M. 2023. Anomaly detection model based on gradient boosting and decision tree for IoT environments security. J. Reliable Intell. Environ. Vol. 9:421–432. [Cross Ref]

            16. Fiorini L, De Mul M, Fabbricotti I, Limosani R, Vitanza A, D’Onofrio G, et al.. 2021. Assistive robots to improve the independent living of older persons: results from a needs study. Disabil. Rehabil. Assist. Technol. Vol. 16:92–102. [Cross Ref]

            17. Islam RB, Akhter S, Iqbal F, Saif Ur Rahman M, Khan R. 2023. Deep learning based object detection and surrounding environment description for visually impaired people. Heliyon. Vol. 9:e16924. [Cross Ref]

            18. Jacob S, Alagirisamy M, Xi C, Balasubramanian V, Srinivasan R, Parvathi R, et al.. 2021. AI and IoT-enabled smart exoskeleton system for rehabilitation of paralyzed people in connected communities. IEEE Access. Vol. 9:80340–80350. [Cross Ref]

            19. Kbar G, Al-Daraiseh A, Mian SH, Abidi MH. 2016. Utilizing sensors networks to develop a smart and context-aware solution for people with disabilities at the workplace (design and implementation). Int. J. Distrib. Sens. Netw. Vol. 12:1550147716658606. [Cross Ref]

            20. Kbar G, Al-Daraiseh A, Aly S, Abidi MH, Mian SH. 2017. Assessment of technologies relevant for people with motor hearing and speech impairment. IETE Tech. Rev. Vol. 34:254–264. [Cross Ref]

            21. Malcangi M. 2021. AI-based methods and technologies to develop wearable devices for prosthetics and predictions of degenerative diseases. Methods Mol. Biol. Vol. 2190:337–354. [Cross Ref]

            22. Mehta P, Mehendale N. 2023. Image and video dataset for visually impaired using intel realsense camera. IEEE Dataport. [Cross Ref]

            23. Nierling L, Maia M. 2020. Assistive technologies: social barriers and socio-technical pathways. Societies. Vol. 10:41

            24. Paing MP, Juhong A, Pintavirooj C. 2022. Design and development of an assistive system based on eye tracking. Electronics. Vol. 11:535[Cross Ref]

            25. Ramirez ARG. 2023. Introductory chapter: trends in assistive technologyTrends in Assistive Technologies. Alejandro Rafael G-R. IntechOpen. Rijeka: p. 1–8

            26. Said Y, Atri M, Albahar MA, Ben Atitallah A, Alsariera YA. 2023. Obstacle detection system for navigation assistance of visually impaired people based on deep learning techniques. Sensors. Vol. 23:5262

            27. Söderström S, Østby M, Bakken H, Ellingsen KE. 2021. How using assistive technology for cognitive impairments improves the participation and self-determination of young adults with intellectual developmental disabilities. J. Intellect. Disabil. Vol. 25:168–182. [Cross Ref]

            28. Tarik H, Hassan S, Naqvi RA, Rubab S, Tariq U, Hamdi M, et al.. 2023. Empowering and conquering infirmity of visually impaired using AI-technology equipped with object detection and real-time voice feedback system in healthcare application. CAAI Trans. Intell. Technol. 1–14. [Cross Ref]

            29. Thakur N, Han CY. 2021. A review of assistive technologies for activities of daily living of elderly. arXiv. [Cross Ref]

            30. Wang Z, Cui L, Guo W, Zhao L, Yuan X, Gu X, et al.. 2022. A design method for an intelligent manufacturing and service system for rehabilitation assistive devices and special groups. Adv. Eng. Inform. Vol. 51:101504. [Cross Ref]

            31. Wang H, Li G, Wang Z. 2023. Fast SVM classifier for large-scale classification problems. Inform. Sci. Vol. 642:119136. [Cross Ref]

            Author and article information

            Journal
            jdr
            Journal of Disability Research
            King Salman Centre for Disability Research (Riyadh, Saudi Arabia )
            1658-9912
            31 May 2024
            : 3
            : 5
            : e20240057
            Affiliations
            [1 ] Advanced Manufacturing Institute, King Saud University, Riyadh 11421, Saudi Arabia ( https://ror.org/02f81g417)
            [2 ] King Salman Center for Disability Research, Riyadh 11614, Saudi Arabia ( https://ror.org/01ht2b307)
            [3 ] Department of Mechanical Engineering, Jamia Millia Islamia, New Delhi 110025, India ( https://ror.org/00pnhhv55)
            Author notes
            Correspondence to: Mustufa Haider Abidi*, e-mail: mabidi@ 123456ksu.edu.sa , Tel: +966-11-4698773, Fax: +966-11-4670969
            Author information
            https://orcid.org/0000-0001-5609-5705
            https://orcid.org/0000-0003-1909-2534
            https://orcid.org/0000-0002-3573-8385
            Article
            10.57197/JDR-2024-0057
            975c4162-f5b6-4d2b-8785-00ac747592f6
            Copyright © 2024 The Authors.

            This is an open access article distributed under the terms of the Creative Commons Attribution License (CC BY) 4.0, which permits unrestricted use, distribution and reproduction in any medium, provided the original author and source are credited.

            History
            : 04 March 2024
            : 02 May 2024
            : 02 May 2024
            Page count
            Figures: 10, Tables: 3, References: 31, Pages: 11
            Funding
            Funded by: King Salman Center for Disability Research
            Award ID: KSRG-2023-451
            The authors extend their appreciation to the King Salman Center for Disability Research (funder ID: http://dx.doi.org/10.13039/501100019345) for funding this work through Research Group no. KSRG-2023-451.

            Social policy & Welfare,Political science,Education & Public policy,Special education,Civil law,Social & Behavioral Sciences
            visual impairment,gradient boosting,visual recognition techniques,support vector machine,golden jackal optimization

            Comments

            Comment on this article