Power Shell has been widely deployed in fileless malware and advanced persistent threat(APT)attacks due to its high stealthiness and live-off-theland technique.However,existing works mainly focus on deobfuscation and ...Power Shell has been widely deployed in fileless malware and advanced persistent threat(APT)attacks due to its high stealthiness and live-off-theland technique.However,existing works mainly focus on deobfuscation and malicious detection,lacking the malicious Power Shell families classification and behavior analysis.Moreover,the state-of-the-art methods fail to capture fine-grained features and semantic relationships,resulting in low robustness and accuracy.To this end,we propose Power Detector,a novel malicious Power Shell script detector based on multimodal semantic fusion and deep learning.Specifically,we design four feature extraction methods to extract key features from character,token,abstract syntax tree(AST),and semantic knowledge graph.Then,we intelligently design four embeddings(i.e.,Char2Vec,Token2Vec,AST2Vec,and Rela2Vec) and construct a multi-modal fusion algorithm to concatenate feature vectors from different views.Finally,we propose a combined model based on transformer and CNN-Bi LSTM to implement Power Shell family detection.Our experiments with five types of Power Shell attacks show that PowerDetector can accurately detect various obfuscated and stealth PowerShell scripts,with a 0.9402 precision,a 0.9358 recall,and a 0.9374 F1-score.Furthermore,through singlemodal and multi-modal comparison experiments,we demonstrate that PowerDetector’s multi-modal embedding and deep learning model can achieve better accuracy and even identify more unknown attacks.展开更多
Artificial intelligence can be indirectly applied to the repair of peripheral nerve injury.Specifically,it can be used to analyze and process data regarding peripheral nerve injury and repair,while study findings on p...Artificial intelligence can be indirectly applied to the repair of peripheral nerve injury.Specifically,it can be used to analyze and process data regarding peripheral nerve injury and repair,while study findings on peripheral nerve injury and repair can provide valuable data to enrich artificial intelligence algorithms.To investigate advances in the use of artificial intelligence in the diagnosis,rehabilitation,and scientific examination of peripheral nerve injury,we used CiteSpace and VOSviewer software to analyze the relevant literature included in the Web of Science from 1994–2023.We identified the following research hotspots in peripheral nerve injury and repair:(1)diagnosis,classification,and prognostic assessment of peripheral nerve injury using neuroimaging and artificial intelligence techniques,such as corneal confocal microscopy and coherent anti-Stokes Raman spectroscopy;(2)motion control and rehabilitation following peripheral nerve injury using artificial neural networks and machine learning algorithms,such as wearable devices and assisted wheelchair systems;(3)improving the accuracy and effectiveness of peripheral nerve electrical stimulation therapy using artificial intelligence techniques combined with deep learning,such as implantable peripheral nerve interfaces;(4)the application of artificial intelligence technology to brain-machine interfaces for disabled patients and those with reduced mobility,enabling them to control devices such as networked hand prostheses;(5)artificial intelligence robots that can replace doctors in certain procedures during surgery or rehabilitation,thereby reducing surgical risk and complications,and facilitating postoperative recovery.Although artificial intelligence has shown many benefits and potential applications in peripheral nerve injury and repair,there are some limitations to this technology,such as the consequences of missing or imbalanced data,low data accuracy and reproducibility,and ethical issues(e.g.,privacy,data security,research transparency).Future research should address the issue of data collection,as large-scale,high-quality clinical datasets are required to establish effective artificial intelligence models.Multimodal data processing is also necessary,along with interdisciplinary collaboration,medical-industrial integration,and multicenter,large-sample clinical studies.展开更多
Human hand detection in uncontrolled environments is a challenging visual recognition task due to numerous variations of hand poses and background image clutter.To achieve highly accurate results as well as provide re...Human hand detection in uncontrolled environments is a challenging visual recognition task due to numerous variations of hand poses and background image clutter.To achieve highly accurate results as well as provide real-time execution,we proposed a deep transfer learning approach over the state-of-the-art deep learning object detector.Our method,denoted as YOLOHANDS,is built on top of the You Only Look Once(YOLO)deep learning architecture,which is modified to adapt to the single class hand detection task.The model transfer is performed by modifying the higher convolutional layers including the last fully connected layer,while initializing lower non-modified layers with the generic pre-trained weights.To address robustness issues,we introduced a comprehensive augmentation procedure over the training image dataset,specifically adapted for the hand detection problem.Experimental evaluation of the proposed method,which is performed on a challenging public dataset,has demonstrated highly accurate results,comparable to the state-of-the-art methods.展开更多
Sign language recognition can be treated as one of the efficient solu-tions for disabled people to communicate with others.It helps them to convey the required data by the use of sign language with no issues.The lates...Sign language recognition can be treated as one of the efficient solu-tions for disabled people to communicate with others.It helps them to convey the required data by the use of sign language with no issues.The latest develop-ments in computer vision and image processing techniques can be accurately uti-lized for the sign recognition process by disabled people.American Sign Language(ASL)detection was challenging because of the enhancing intraclass similarity and higher complexity.This article develops a new Bayesian Optimiza-tion with Deep Learning-Driven Hand Gesture Recognition Based Sign Language Communication(BODL-HGRSLC)for Disabled People.The BODL-HGRSLC technique aims to recognize the hand gestures for disabled people’s communica-tion.The presented BODL-HGRSLC technique integrates the concepts of compu-ter vision(CV)and DL models.In the presented BODL-HGRSLC technique,a deep convolutional neural network-based residual network(ResNet)model is applied for feature extraction.Besides,the presented BODL-HGRSLC model uses Bayesian optimization for the hyperparameter tuning process.At last,a bidir-ectional gated recurrent unit(BiGRU)model is exploited for the HGR procedure.A wide range of experiments was conducted to demonstrate the enhanced perfor-mance of the presented BODL-HGRSLC model.The comprehensive comparison study reported the improvements of the BODL-HGRSLC model over other DL models with maximum accuracy of 99.75%.展开更多
Hybrid Machine Learning (HML) is a kind of advanced algorithm in the field of intelligent information process. It combines the induced learning based-on decision-making tree with the blocking neural network. And it pr...Hybrid Machine Learning (HML) is a kind of advanced algorithm in the field of intelligent information process. It combines the induced learning based-on decision-making tree with the blocking neural network. And it provides a useful intelligent knowledge-based data mining technique. Its core algorithm is ID3 and Field Theory based ART (FTART). The paper introduces the principals of hybrid machine learning firstly, and then applies it into analyzing family apparel expenditures and their influencing factors systematically. Finally, compared with those from the traditional statistic methods, the results from HML is more friendly and easily to be understood. Besides, the forecasting by HML is more correct than by the traditional ways.展开更多
It has been proved by the ancient and modemacupuncture practices that needling techniques are ofvital importance in the acupuncture treatment ofdiseases.At present,much has been reported aboutthe researches on acupunc...It has been proved by the ancient and modemacupuncture practices that needling techniques are ofvital importance in the acupuncture treatment ofdiseases.At present,much has been reported aboutthe researches on acupuncture techniques,which canbe roughly divided into the following three types:reinforcing,reducing,and even methods.Throughlong years of study and clinical practice,Dr.展开更多
This paper aims to verify the family support situation for primary school children with intellectual disabilities learning in regular class and to explore various educational strategies to promote their development.A ...This paper aims to verify the family support situation for primary school children with intellectual disabilities learning in regular class and to explore various educational strategies to promote their development.A self-made questionnaire was used in this survey,and the parents of 380 intellectual disabled students were the subjects of this survey.It turns out that the overall family support for intellectual disabled children learning in regular class in China is good,but it is affected by the degree of obstacles.Factors such as grade,gender,and parental education had no significant effect on family support.It is the shared responsibility of the government,schools,and parents to promote the level of family support.Governments at all levels must implement family support projects,schools must carry out family education guidance to impart scientific parenting knowledge,and parents must take note of their own responsibilities,so as to promote the physical and mental development of children with intellectual disabilities.展开更多
In this paper we present a CNN based approach for a real time 3 D-hand pose estimation from the depth sequence.Prior discriminative approaches have achieved remarkable success but are facing two main challenges:Firstl...In this paper we present a CNN based approach for a real time 3 D-hand pose estimation from the depth sequence.Prior discriminative approaches have achieved remarkable success but are facing two main challenges:Firstly,the methods are fully supervised hence require large numbers of annotated training data to extract the dynamic information from a hand representation.Secondly,unreliable hand detectors based on strong assumptions or a weak detector which often fail in several situations like complex environment and multiple hands.In contrast to these methods,this paper presents an approach that can be considered as semi-supervised by performing predictive coding of image sequences of hand poses in order to capture latent features underlying a given image without supervision.The hand is modelled using a novel latent tree dependency model(LDTM)which transforms internal joint location to an explicit representation.Then the modeled hand topology is integrated with the pose estimator using data dependent method to jointly learn latent variables of the posterior pose appearance and the pose configuration respectively.Finally,an unsupervised error term which is a part of the recurrent architecture ensures smooth estimations of the final pose.Experiments on three challenging public datasets,ICVL,MSRA,and NYU demonstrate the significant performance of the proposed method which is comparable or better than state-of-the-art approaches.展开更多
Parkinson’s disease manifests in movement disorder symptoms, such as hand tremor. There exists an assortment of therapy interventions. In particular deep brain stimulation offers considerable efficacy for the treatme...Parkinson’s disease manifests in movement disorder symptoms, such as hand tremor. There exists an assortment of therapy interventions. In particular deep brain stimulation offers considerable efficacy for the treatment of Parkinson’s disease. However, a considerable challenge is the convergence toward an optimal configuration of tuning parameters. Quantified feedback from a wearable and wireless system consisting of an accelerometer and gyroscope can be enabled through a novel software application on a smartphone. The smartphone with its internal accelerometer and gyroscope can record the quantified attributes of Parkinson’s disease and tremor through mounting the smartphone about the dorsum of the hand. The recorded data can be then wirelessly transmitted as an email attachment to an Internet derived resource for subsequent post-processing. The inertial sensor data can be consolidated into a feature set for machine learning classification. A multilayer perceptron neural network has been successfully applied to attain considerable classification accuracy between deep brain stimulation “On” and “Off” scenarios for a subject with Parkinson’s disease. The findings establish the foundation for the broad objective of applying wearable and wireless systems for the development of closed-loop optimization of deep brain stimulation parameters in the context of cloud computing with machine learning classification.展开更多
Split hand/foot malformation (SHFM), formerly known as ectrodactyly is a rare congénital anomaly, its incidence varies from 1/8.500 to 1/25.000 live birth. It mainly affects the development of the limbs, its clin...Split hand/foot malformation (SHFM), formerly known as ectrodactyly is a rare congénital anomaly, its incidence varies from 1/8.500 to 1/25.000 live birth. It mainly affects the development of the limbs, its clinical variability is standard, can present as an isolated feature or as a syndrome associated with other congenital anomalies. Our objective was to present the two cases of SHFM, and to review the literature on the clinical aspects and discuss a probable origin. The father went to school and is a driver because the malformations concerned only the fingers, were less severe, and did not prevent the realization of certain simple gestures of the daily life. On the other hand, the malformations of the fingers of the newborn were severe and the absence of the thumbs compromised the later prehension function. Also the association of a microglossia and a cleft palate contributed to a weight loss that justified hospitalization. The clinical presentation of split hands and feet is variable and the prognosis depends on the type of anomaly. Familial cases suggest a probable genetic origin. Genetic testing is necessary to establish genetic counseling.展开更多
Hand Gesture Recognition(HGR)is a promising research area with an extensive range of applications,such as surgery,video game techniques,and sign language translation,where sign language is a complicated structured for...Hand Gesture Recognition(HGR)is a promising research area with an extensive range of applications,such as surgery,video game techniques,and sign language translation,where sign language is a complicated structured form of hand gestures.The fundamental building blocks of structured expressions in sign language are the arrangement of the fingers,the orientation of the hand,and the hand’s position concerning the body.The importance of HGR has increased due to the increasing number of touchless applications and the rapid growth of the hearing-impaired population.Therefore,real-time HGR is one of the most effective interaction methods between computers and humans.Developing a user-free interface with good recognition performance should be the goal of real-time HGR systems.Nowadays,Convolutional Neural Network(CNN)shows great recognition rates for different image-level classification tasks.It is challenging to train deep CNN networks like VGG-16,VGG-19,Inception-v3,and Efficientnet-B0 from scratch because only some significant labeled image datasets are available for static hand gesture images.However,an efficient and robust hand gesture recognition system of sign language employing finetuned Inception-v3 and Efficientnet-Bo network is proposed to identify hand gestures using a comparative small HGR dataset.Experiments show that Inception-v3 achieved 90%accuracy and 0.93%precision,0.91%recall,and 0.90%f1-score,respectively,while EfficientNet-B0 achieved 99%accuracy and 0.98%,0.97%,0.98%,precision,recall,and f1-score respectively.展开更多
AIM:To describe the learning curves of hand-assisted laparoscopic D2 radical gastrectomy(HALG) for the treatment of gastric cancer.METHODS:The HALG surgical procedure consists of three stages:surgery under direct visi...AIM:To describe the learning curves of hand-assisted laparoscopic D2 radical gastrectomy(HALG) for the treatment of gastric cancer.METHODS:The HALG surgical procedure consists of three stages:surgery under direct vision via the port for hand assistance,hand-assisted laparoscopicsurgery,and gastrointestinal tract reconstruction.According to the order of the date of surgery,patients were divided into 6 groups(A-F) with 20 cases in each group.All surgeries were performed by the same group of surgeons.We performed a comprehensive and indepth retrospective comparative analysis of the clinical data of all patients,with the clinical data including general patient information and intraoperative and postoperative observation indicators.RESULTS:There were no differences in the basic information among the patient groups(P > 0.05).The operative time of the hand-assisted surgery stage in group A was 8-10 min longer than the other groups,with the difference being statistically significant(P = 0.01).There were no differences in total operative time between the groups(P = 0.30).Postoperative intestinal function recovery time in group A was longer than that of other groups(P = 0.02).Lengths of hospital stay and surgical quality indicators(such as intraoperative blood loss,numbers of detected lymph nodes,intraoperative side injury,postoperative complications,reoperation rate,and readmission rate 30 d after surgery) were not significantly different among the groups.CONCLUSION:HALG is a surgical procedure that can be easily mastered,with a learning curve closely related to the operative time of the hand-assisted laparoscopic surgery stage.展开更多
A multidisciplinary approach for developing an intelligent sign multi-language recognition system to greatly enhance deaf-mute communication will be discussed and implemented. This involves designing a low-cost glove-...A multidisciplinary approach for developing an intelligent sign multi-language recognition system to greatly enhance deaf-mute communication will be discussed and implemented. This involves designing a low-cost glove-based sensing system, collecting large and diverse datasets, preprocessing the data, and using efficient machine learning models. Furthermore, the glove is integrated with a user-friendly mobile application called “Life-sign” for this system. The main goal of this work is to minimize the processing time of machine learning classifiers while maintaining higher accuracy performance. This is achieved by using effective preprocessing algorithms to handle noisy and inconsistent data. Testing and iterating approaches have been applied to various classifiers to refine and improve their accuracy in the recognition process. Additionally, the Extra Trees (ET) classifier has been identified as the best algorithm, with results proving successful gesture prediction at an average accuracy of about 99.54%. A smart optimization feature has been implemented to control the size of data transferred via Bluetooth, allowing for fast recognition of consecutive gestures. Real-time performance has been measured through extensive experimental testing on various consecutive gestures, specifically referring to Arabic Sign Language (ArSL). The results have demonstrated that the system guarantees consecutive gesture recognition with a lower delay of 50 milliseconds.展开更多
The concept of Network Centric Therapy represents an amalgamation of wearable and wireless inertial sensor systems and machine learning with access to a Cloud computing environment. The advent of Network Centric Thera...The concept of Network Centric Therapy represents an amalgamation of wearable and wireless inertial sensor systems and machine learning with access to a Cloud computing environment. The advent of Network Centric Therapy is highly relevant to the treatment of Parkinson’s disease through deep brain stimulation. Originally wearable and wireless systems for quantifying Parkinson’s disease involved the use a smartphone to quantify hand tremor. Although originally novel, the smartphone has notable issues as a wearable application for quantifying movement disorder tremor. The smartphone has evolved in a pathway that has made the smartphone progressively more cumbersome to mount about the dorsum of the hand. Furthermore, the smartphone utilizes an inertial sensor package that is not certified for medical analysis, and the trial data access a provisional Cloud computing environment through an email account. These concerns are resolved with the recent development of a conformal wearable and wireless inertial sensor system. This conformal wearable and wireless system mounts to the hand with the profile of a bandage by adhesive and accesses a secure Cloud computing environment through a segmented wireless connectivity strategy involving a smartphone and tablet. Additionally, the conformal wearable and wireless system is certified by the FDA of the United States of America for ascertaining medical grade inertial sensor data. These characteristics make the conformal wearable and wireless system uniquely suited for the quantification of Parkinson’s disease treatment through deep brain stimulation. Preliminary evaluation of the conformal wearable and wireless system is demonstrated through the differentiation of deep brain stimulation set to “On” and “Off” status. Based on the robustness of the acceleration signal, this signal was selected to quantify hand tremor for the prescribed deep brain stimulation settings. Machine learning classification using the Waikato Environment for Knowledge Analysis (WEKA) was applied using the multilayer perceptron neural network. The multilayer perceptron neural network achieved considerable classification accuracy for distinguishing between the deep brain stimulation system set to “On” and “Off” status through the quantified acceleration signal data obtained by this recently developed conformal wearable and wireless system. The research achievement establishes a progressive pathway to the future objective of achieving deep brain stimulation capabilities that promote closed-loop acquisition of configuration parameters that are uniquely optimized to the individual through extrinsic means of a highly conformal wearable and wireless inertial sensor system and machine learning with access to Cloud computing resources.展开更多
Recent hand pose estimation methods require large numbers of annotated training data to extract the dynamic information from a hand representation.Nevertheless,precise and dense annotation on the real data is difficul...Recent hand pose estimation methods require large numbers of annotated training data to extract the dynamic information from a hand representation.Nevertheless,precise and dense annotation on the real data is difficult to come by and the amount of information passed to the training algorithm is significantly higher.This paper presents an approach to developing a hand pose estimation system which can accurately regress a 3D pose in an unsupervised manner.The whole process is performed in three stages.Firstly,the hand is modelled by a novel latent tree dependency model (LTDM) which transforms internal joints location to an explicit representation.Secondly,we perform predictive coding of image sequences of hand poses in order to capture latent features underlying a given image without supervision.A mapping is then performed between an image depth and a generated representation.Thirdly,the hand joints are regressed using convolutional neural networks to finally estimate the latent pose given some depth map.Finally,an unsupervised error term which is a part of the recurrent architecture ensures smooth estimation of the final pose.To demonstrate the performance of the proposed system,a complete experiment was conducted on three challenging public datasets,ICVL,MSRA,and NYU.The empirical results show the significant performance of our method which is comparable or better than the state-of-the-art approaches.展开更多
The evident change in the design of the autopilot system produced massive help for the aviation industry and it required frequent upgrades.Reinforcement learning delivers appropriate outcomes when considering a contin...The evident change in the design of the autopilot system produced massive help for the aviation industry and it required frequent upgrades.Reinforcement learning delivers appropriate outcomes when considering a continuous environment where the controlling Unmanned Aerial Vehicle(UAV)required maximum accuracy.In this paper,we designed a hybrid framework,which is based on Reinforcement Learning and Deep Learning where the traditional electronic flight controller is replaced by using 3D hand gestures.The algorithm is designed to take the input from 3D hand gestures and integrate with the Deep Deterministic Policy Gradient(DDPG)to receive the best reward and take actions according to 3D hand gestures input.The UAV consist of a Jetson Nano embedded testbed,Global Positioning System(GPS)sensor module,and Intel depth camera.The collision avoidance system based on the polar mask segmentation technique detects the obstacles and decides the best path according to the designed reward function.The analysis of the results has been observed providing best accuracy and computational time using novel design framework when compared with traditional Proportional Integral Derivatives(PID)flight controller.There are six reward functions estimated for 2500,5000,7500,and 10000 episodes of training,which have been normalized between 0 to−4000.The best observation has been captured on 2500 episodes where the rewards are calculated for maximum value.The achieved training accuracy of polar mask segmentation for collision avoidance is 86.36%.展开更多
In this research work,an efficient sign language recognition tool for e-learning has been proposed with a new type of feature set based on angle and lines.This feature set has the ability to increase the overall perfo...In this research work,an efficient sign language recognition tool for e-learning has been proposed with a new type of feature set based on angle and lines.This feature set has the ability to increase the overall performance of machine learning algorithms in an efficient way.The hand gesture recognition based on these features has been implemented for usage in real-time.The feature set used hand landmarks,which were generated using media-pipe(MediaPipe)and open computer vision(openCV)on each frame of the incoming video.The overall algorithm has been tested on two well-known ASLalphabet(American Sign Language)and ISL-HS(Irish Sign Language)sign language datasets.Different machine learning classifiers including random forest,decision tree,and naïve Bayesian have been used to classify hand gestures using this unique feature set and their respective results have been compared.Since the random forest classifier performed better,it has been selected as the base classifier for the proposed system.It showed 96.7%accuracy with ISL-HS and 93.7%accuracy with ASL-alphabet dataset using the extracted features.展开更多
Brain-computer interfaces (BCIs) records brain activity using electroencephalogram (EEG) headsets in the form of EEG signals;these signals can berecorded, processed and classified into different hand movements, which...Brain-computer interfaces (BCIs) records brain activity using electroencephalogram (EEG) headsets in the form of EEG signals;these signals can berecorded, processed and classified into different hand movements, which can beused to control other IoT devices. Classification of hand movements will beone step closer to applying these algorithms in real-life situations using EEGheadsets. This paper uses different feature extraction techniques and sophisticatedmachine learning algorithms to classify hand movements from EEG brain signalsto control prosthetic hands for amputated persons. To achieve good classificationaccuracy, denoising and feature extraction of EEG signals is a significant step. Wesaw a considerable increase in all the machine learning models when the movingaverage filter was applied to the raw EEG data. Feature extraction techniques likea fast fourier transform (FFT) and continuous wave transform (CWT) were usedin this study;three types of features were extracted, i.e., FFT Features, CWTCoefficients and CWT scalogram images. We trained and compared differentmachine learning (ML) models like logistic regression, random forest, k-nearestneighbors (KNN), light gradient boosting machine (GBM) and XG boost onFFT and CWT features and deep learning (DL) models like VGG-16, DenseNet201 and ResNet50 trained on CWT scalogram images. XG Boost with FFTfeatures gave the maximum accuracy of 88%.展开更多
With the advancement of technology and the increase in user demands, gesture recognition played a pivotal role in the field of human-computer interaction. Among various sensing devices, Time-of-Flight (ToF) sensors we...With the advancement of technology and the increase in user demands, gesture recognition played a pivotal role in the field of human-computer interaction. Among various sensing devices, Time-of-Flight (ToF) sensors were widely applied due to their low cost. This paper explored the implementation of a human hand posture recognition system using ToF sensors and residual neural networks. Firstly, this paper reviewed the typical applications of human hand recognition. Secondly, this paper designed a hand gesture recognition system using a ToF sensor VL53L5. Subsequently, data preprocessing was conducted, followed by training the constructed residual neural network. Then, the recognition results were analyzed, indicating that gesture recognition based on the residual neural network achieved an accuracy of 98.5% in a 5-class classification scenario. Finally, the paper discussed existing issues and future research directions.展开更多
Deep brain stimulation offers an advanced means of treating Parkinson’s disease in a patient specific context. However, a considerable challenge is the process of ascertaining an optimal parameter configuration. Impe...Deep brain stimulation offers an advanced means of treating Parkinson’s disease in a patient specific context. However, a considerable challenge is the process of ascertaining an optimal parameter configuration. Imperative for the deep brain stimulation parameter optimization process is the quantification of response feedback. As a significant improvement to traditional ordinal scale techniques is the advent of wearable and wireless systems. Recently conformal wearable and wireless systems with a profile on the order of a bandage have been developed. Previous research endeavors have successfully differentiated between deep brain stimulation “On” and “Off” status through quantification using wearable and wireless inertial sensor systems. However, the opportunity exists to further evolve to an objectively quantified response to an assortment of parameter configurations, such as the variation of amplitude, for the deep brain stimulation system. Multiple deep brain stimulation amplitude settings are considered inclusive of “Off” status as a baseline, 1.0 mA, 2.5 mA, and 4.0 mA. The quantified response of this assortment of amplitude settings is acquired through a conformal wearable and wireless inertial sensor system and consolidated using Python software automation to a feature set amenable for machine learning. Five machine learning algorithms are evaluated: J48 decision tree, K-nearest neighbors, support vector machine, logistic regression, and random forest. The performance of these machine learning algorithms is established based on the classification accuracy to distinguish between the deep brain stimulation amplitude settings and the time to develop the machine learning model. The support vector machine achieves the greatest classification accuracy, which is the primary performance parameter, and <span style="font-family:Verdana;">K-nearest neighbors achieves considerable classification accuracy with minimal time to develop the machine learning model.</span>展开更多
基金This work was supported by National Natural Science Foundation of China(No.62172308,No.U1626107,No.61972297,No.62172144,and No.62062019).
文摘Power Shell has been widely deployed in fileless malware and advanced persistent threat(APT)attacks due to its high stealthiness and live-off-theland technique.However,existing works mainly focus on deobfuscation and malicious detection,lacking the malicious Power Shell families classification and behavior analysis.Moreover,the state-of-the-art methods fail to capture fine-grained features and semantic relationships,resulting in low robustness and accuracy.To this end,we propose Power Detector,a novel malicious Power Shell script detector based on multimodal semantic fusion and deep learning.Specifically,we design four feature extraction methods to extract key features from character,token,abstract syntax tree(AST),and semantic knowledge graph.Then,we intelligently design four embeddings(i.e.,Char2Vec,Token2Vec,AST2Vec,and Rela2Vec) and construct a multi-modal fusion algorithm to concatenate feature vectors from different views.Finally,we propose a combined model based on transformer and CNN-Bi LSTM to implement Power Shell family detection.Our experiments with five types of Power Shell attacks show that PowerDetector can accurately detect various obfuscated and stealth PowerShell scripts,with a 0.9402 precision,a 0.9358 recall,and a 0.9374 F1-score.Furthermore,through singlemodal and multi-modal comparison experiments,we demonstrate that PowerDetector’s multi-modal embedding and deep learning model can achieve better accuracy and even identify more unknown attacks.
基金supported by the Capital’s Funds for Health Improvement and Research,No.2022-2-2072(to YG).
文摘Artificial intelligence can be indirectly applied to the repair of peripheral nerve injury.Specifically,it can be used to analyze and process data regarding peripheral nerve injury and repair,while study findings on peripheral nerve injury and repair can provide valuable data to enrich artificial intelligence algorithms.To investigate advances in the use of artificial intelligence in the diagnosis,rehabilitation,and scientific examination of peripheral nerve injury,we used CiteSpace and VOSviewer software to analyze the relevant literature included in the Web of Science from 1994–2023.We identified the following research hotspots in peripheral nerve injury and repair:(1)diagnosis,classification,and prognostic assessment of peripheral nerve injury using neuroimaging and artificial intelligence techniques,such as corneal confocal microscopy and coherent anti-Stokes Raman spectroscopy;(2)motion control and rehabilitation following peripheral nerve injury using artificial neural networks and machine learning algorithms,such as wearable devices and assisted wheelchair systems;(3)improving the accuracy and effectiveness of peripheral nerve electrical stimulation therapy using artificial intelligence techniques combined with deep learning,such as implantable peripheral nerve interfaces;(4)the application of artificial intelligence technology to brain-machine interfaces for disabled patients and those with reduced mobility,enabling them to control devices such as networked hand prostheses;(5)artificial intelligence robots that can replace doctors in certain procedures during surgery or rehabilitation,thereby reducing surgical risk and complications,and facilitating postoperative recovery.Although artificial intelligence has shown many benefits and potential applications in peripheral nerve injury and repair,there are some limitations to this technology,such as the consequences of missing or imbalanced data,low data accuracy and reproducibility,and ethical issues(e.g.,privacy,data security,research transparency).Future research should address the issue of data collection,as large-scale,high-quality clinical datasets are required to establish effective artificial intelligence models.Multimodal data processing is also necessary,along with interdisciplinary collaboration,medical-industrial integration,and multicenter,large-sample clinical studies.
基金financed by the Ministry of Education,Science and Technological Development of the Republic of Serbia.
文摘Human hand detection in uncontrolled environments is a challenging visual recognition task due to numerous variations of hand poses and background image clutter.To achieve highly accurate results as well as provide real-time execution,we proposed a deep transfer learning approach over the state-of-the-art deep learning object detector.Our method,denoted as YOLOHANDS,is built on top of the You Only Look Once(YOLO)deep learning architecture,which is modified to adapt to the single class hand detection task.The model transfer is performed by modifying the higher convolutional layers including the last fully connected layer,while initializing lower non-modified layers with the generic pre-trained weights.To address robustness issues,we introduced a comprehensive augmentation procedure over the training image dataset,specifically adapted for the hand detection problem.Experimental evaluation of the proposed method,which is performed on a challenging public dataset,has demonstrated highly accurate results,comparable to the state-of-the-art methods.
基金The authors extend their appreciation to the King Salman centre for Disability Research for funding this work through Research Group no KSRG-2022-017.
文摘Sign language recognition can be treated as one of the efficient solu-tions for disabled people to communicate with others.It helps them to convey the required data by the use of sign language with no issues.The latest develop-ments in computer vision and image processing techniques can be accurately uti-lized for the sign recognition process by disabled people.American Sign Language(ASL)detection was challenging because of the enhancing intraclass similarity and higher complexity.This article develops a new Bayesian Optimiza-tion with Deep Learning-Driven Hand Gesture Recognition Based Sign Language Communication(BODL-HGRSLC)for Disabled People.The BODL-HGRSLC technique aims to recognize the hand gestures for disabled people’s communica-tion.The presented BODL-HGRSLC technique integrates the concepts of compu-ter vision(CV)and DL models.In the presented BODL-HGRSLC technique,a deep convolutional neural network-based residual network(ResNet)model is applied for feature extraction.Besides,the presented BODL-HGRSLC model uses Bayesian optimization for the hyperparameter tuning process.At last,a bidir-ectional gated recurrent unit(BiGRU)model is exploited for the HGR procedure.A wide range of experiments was conducted to demonstrate the enhanced perfor-mance of the presented BODL-HGRSLC model.The comprehensive comparison study reported the improvements of the BODL-HGRSLC model over other DL models with maximum accuracy of 99.75%.
文摘Hybrid Machine Learning (HML) is a kind of advanced algorithm in the field of intelligent information process. It combines the induced learning based-on decision-making tree with the blocking neural network. And it provides a useful intelligent knowledge-based data mining technique. Its core algorithm is ID3 and Field Theory based ART (FTART). The paper introduces the principals of hybrid machine learning firstly, and then applies it into analyzing family apparel expenditures and their influencing factors systematically. Finally, compared with those from the traditional statistic methods, the results from HML is more friendly and easily to be understood. Besides, the forecasting by HML is more correct than by the traditional ways.
文摘It has been proved by the ancient and modemacupuncture practices that needling techniques are ofvital importance in the acupuncture treatment ofdiseases.At present,much has been reported aboutthe researches on acupuncture techniques,which canbe roughly divided into the following three types:reinforcing,reducing,and even methods.Throughlong years of study and clinical practice,Dr.
基金supported by The Final Achievement of the 13th Five-Year Plan of Philosophy and Social Sciences in Guangdong Province in 2020“Research on the Relationship Between Family Support,School Support and School Adaptation of Regular Primary School Students(No.:GD20XJY27).
文摘This paper aims to verify the family support situation for primary school children with intellectual disabilities learning in regular class and to explore various educational strategies to promote their development.A self-made questionnaire was used in this survey,and the parents of 380 intellectual disabled students were the subjects of this survey.It turns out that the overall family support for intellectual disabled children learning in regular class in China is good,but it is affected by the degree of obstacles.Factors such as grade,gender,and parental education had no significant effect on family support.It is the shared responsibility of the government,schools,and parents to promote the level of family support.Governments at all levels must implement family support projects,schools must carry out family education guidance to impart scientific parenting knowledge,and parents must take note of their own responsibilities,so as to promote the physical and mental development of children with intellectual disabilities.
基金supported in part by the Fundamental Research Funds for the Central Universities(WK2350000002)。
文摘In this paper we present a CNN based approach for a real time 3 D-hand pose estimation from the depth sequence.Prior discriminative approaches have achieved remarkable success but are facing two main challenges:Firstly,the methods are fully supervised hence require large numbers of annotated training data to extract the dynamic information from a hand representation.Secondly,unreliable hand detectors based on strong assumptions or a weak detector which often fail in several situations like complex environment and multiple hands.In contrast to these methods,this paper presents an approach that can be considered as semi-supervised by performing predictive coding of image sequences of hand poses in order to capture latent features underlying a given image without supervision.The hand is modelled using a novel latent tree dependency model(LDTM)which transforms internal joint location to an explicit representation.Then the modeled hand topology is integrated with the pose estimator using data dependent method to jointly learn latent variables of the posterior pose appearance and the pose configuration respectively.Finally,an unsupervised error term which is a part of the recurrent architecture ensures smooth estimations of the final pose.Experiments on three challenging public datasets,ICVL,MSRA,and NYU demonstrate the significant performance of the proposed method which is comparable or better than state-of-the-art approaches.
文摘Parkinson’s disease manifests in movement disorder symptoms, such as hand tremor. There exists an assortment of therapy interventions. In particular deep brain stimulation offers considerable efficacy for the treatment of Parkinson’s disease. However, a considerable challenge is the convergence toward an optimal configuration of tuning parameters. Quantified feedback from a wearable and wireless system consisting of an accelerometer and gyroscope can be enabled through a novel software application on a smartphone. The smartphone with its internal accelerometer and gyroscope can record the quantified attributes of Parkinson’s disease and tremor through mounting the smartphone about the dorsum of the hand. The recorded data can be then wirelessly transmitted as an email attachment to an Internet derived resource for subsequent post-processing. The inertial sensor data can be consolidated into a feature set for machine learning classification. A multilayer perceptron neural network has been successfully applied to attain considerable classification accuracy between deep brain stimulation “On” and “Off” scenarios for a subject with Parkinson’s disease. The findings establish the foundation for the broad objective of applying wearable and wireless systems for the development of closed-loop optimization of deep brain stimulation parameters in the context of cloud computing with machine learning classification.
文摘Split hand/foot malformation (SHFM), formerly known as ectrodactyly is a rare congénital anomaly, its incidence varies from 1/8.500 to 1/25.000 live birth. It mainly affects the development of the limbs, its clinical variability is standard, can present as an isolated feature or as a syndrome associated with other congenital anomalies. Our objective was to present the two cases of SHFM, and to review the literature on the clinical aspects and discuss a probable origin. The father went to school and is a driver because the malformations concerned only the fingers, were less severe, and did not prevent the realization of certain simple gestures of the daily life. On the other hand, the malformations of the fingers of the newborn were severe and the absence of the thumbs compromised the later prehension function. Also the association of a microglossia and a cleft palate contributed to a weight loss that justified hospitalization. The clinical presentation of split hands and feet is variable and the prognosis depends on the type of anomaly. Familial cases suggest a probable genetic origin. Genetic testing is necessary to establish genetic counseling.
基金This research work was supported by the National Research Foundation of Korea(NRF)grant funded by the Korean government(MSIT)(NRF-2022R1A2C1004657).
文摘Hand Gesture Recognition(HGR)is a promising research area with an extensive range of applications,such as surgery,video game techniques,and sign language translation,where sign language is a complicated structured form of hand gestures.The fundamental building blocks of structured expressions in sign language are the arrangement of the fingers,the orientation of the hand,and the hand’s position concerning the body.The importance of HGR has increased due to the increasing number of touchless applications and the rapid growth of the hearing-impaired population.Therefore,real-time HGR is one of the most effective interaction methods between computers and humans.Developing a user-free interface with good recognition performance should be the goal of real-time HGR systems.Nowadays,Convolutional Neural Network(CNN)shows great recognition rates for different image-level classification tasks.It is challenging to train deep CNN networks like VGG-16,VGG-19,Inception-v3,and Efficientnet-B0 from scratch because only some significant labeled image datasets are available for static hand gesture images.However,an efficient and robust hand gesture recognition system of sign language employing finetuned Inception-v3 and Efficientnet-Bo network is proposed to identify hand gestures using a comparative small HGR dataset.Experiments show that Inception-v3 achieved 90%accuracy and 0.93%precision,0.91%recall,and 0.90%f1-score,respectively,while EfficientNet-B0 achieved 99%accuracy and 0.98%,0.97%,0.98%,precision,recall,and f1-score respectively.
文摘AIM:To describe the learning curves of hand-assisted laparoscopic D2 radical gastrectomy(HALG) for the treatment of gastric cancer.METHODS:The HALG surgical procedure consists of three stages:surgery under direct vision via the port for hand assistance,hand-assisted laparoscopicsurgery,and gastrointestinal tract reconstruction.According to the order of the date of surgery,patients were divided into 6 groups(A-F) with 20 cases in each group.All surgeries were performed by the same group of surgeons.We performed a comprehensive and indepth retrospective comparative analysis of the clinical data of all patients,with the clinical data including general patient information and intraoperative and postoperative observation indicators.RESULTS:There were no differences in the basic information among the patient groups(P > 0.05).The operative time of the hand-assisted surgery stage in group A was 8-10 min longer than the other groups,with the difference being statistically significant(P = 0.01).There were no differences in total operative time between the groups(P = 0.30).Postoperative intestinal function recovery time in group A was longer than that of other groups(P = 0.02).Lengths of hospital stay and surgical quality indicators(such as intraoperative blood loss,numbers of detected lymph nodes,intraoperative side injury,postoperative complications,reoperation rate,and readmission rate 30 d after surgery) were not significantly different among the groups.CONCLUSION:HALG is a surgical procedure that can be easily mastered,with a learning curve closely related to the operative time of the hand-assisted laparoscopic surgery stage.
文摘A multidisciplinary approach for developing an intelligent sign multi-language recognition system to greatly enhance deaf-mute communication will be discussed and implemented. This involves designing a low-cost glove-based sensing system, collecting large and diverse datasets, preprocessing the data, and using efficient machine learning models. Furthermore, the glove is integrated with a user-friendly mobile application called “Life-sign” for this system. The main goal of this work is to minimize the processing time of machine learning classifiers while maintaining higher accuracy performance. This is achieved by using effective preprocessing algorithms to handle noisy and inconsistent data. Testing and iterating approaches have been applied to various classifiers to refine and improve their accuracy in the recognition process. Additionally, the Extra Trees (ET) classifier has been identified as the best algorithm, with results proving successful gesture prediction at an average accuracy of about 99.54%. A smart optimization feature has been implemented to control the size of data transferred via Bluetooth, allowing for fast recognition of consecutive gestures. Real-time performance has been measured through extensive experimental testing on various consecutive gestures, specifically referring to Arabic Sign Language (ArSL). The results have demonstrated that the system guarantees consecutive gesture recognition with a lower delay of 50 milliseconds.
文摘The concept of Network Centric Therapy represents an amalgamation of wearable and wireless inertial sensor systems and machine learning with access to a Cloud computing environment. The advent of Network Centric Therapy is highly relevant to the treatment of Parkinson’s disease through deep brain stimulation. Originally wearable and wireless systems for quantifying Parkinson’s disease involved the use a smartphone to quantify hand tremor. Although originally novel, the smartphone has notable issues as a wearable application for quantifying movement disorder tremor. The smartphone has evolved in a pathway that has made the smartphone progressively more cumbersome to mount about the dorsum of the hand. Furthermore, the smartphone utilizes an inertial sensor package that is not certified for medical analysis, and the trial data access a provisional Cloud computing environment through an email account. These concerns are resolved with the recent development of a conformal wearable and wireless inertial sensor system. This conformal wearable and wireless system mounts to the hand with the profile of a bandage by adhesive and accesses a secure Cloud computing environment through a segmented wireless connectivity strategy involving a smartphone and tablet. Additionally, the conformal wearable and wireless system is certified by the FDA of the United States of America for ascertaining medical grade inertial sensor data. These characteristics make the conformal wearable and wireless system uniquely suited for the quantification of Parkinson’s disease treatment through deep brain stimulation. Preliminary evaluation of the conformal wearable and wireless system is demonstrated through the differentiation of deep brain stimulation set to “On” and “Off” status. Based on the robustness of the acceleration signal, this signal was selected to quantify hand tremor for the prescribed deep brain stimulation settings. Machine learning classification using the Waikato Environment for Knowledge Analysis (WEKA) was applied using the multilayer perceptron neural network. The multilayer perceptron neural network achieved considerable classification accuracy for distinguishing between the deep brain stimulation system set to “On” and “Off” status through the quantified acceleration signal data obtained by this recently developed conformal wearable and wireless system. The research achievement establishes a progressive pathway to the future objective of achieving deep brain stimulation capabilities that promote closed-loop acquisition of configuration parameters that are uniquely optimized to the individual through extrinsic means of a highly conformal wearable and wireless inertial sensor system and machine learning with access to Cloud computing resources.
文摘Recent hand pose estimation methods require large numbers of annotated training data to extract the dynamic information from a hand representation.Nevertheless,precise and dense annotation on the real data is difficult to come by and the amount of information passed to the training algorithm is significantly higher.This paper presents an approach to developing a hand pose estimation system which can accurately regress a 3D pose in an unsupervised manner.The whole process is performed in three stages.Firstly,the hand is modelled by a novel latent tree dependency model (LTDM) which transforms internal joints location to an explicit representation.Secondly,we perform predictive coding of image sequences of hand poses in order to capture latent features underlying a given image without supervision.A mapping is then performed between an image depth and a generated representation.Thirdly,the hand joints are regressed using convolutional neural networks to finally estimate the latent pose given some depth map.Finally,an unsupervised error term which is a part of the recurrent architecture ensures smooth estimation of the final pose.To demonstrate the performance of the proposed system,a complete experiment was conducted on three challenging public datasets,ICVL,MSRA,and NYU.The empirical results show the significant performance of our method which is comparable or better than the state-of-the-art approaches.
文摘The evident change in the design of the autopilot system produced massive help for the aviation industry and it required frequent upgrades.Reinforcement learning delivers appropriate outcomes when considering a continuous environment where the controlling Unmanned Aerial Vehicle(UAV)required maximum accuracy.In this paper,we designed a hybrid framework,which is based on Reinforcement Learning and Deep Learning where the traditional electronic flight controller is replaced by using 3D hand gestures.The algorithm is designed to take the input from 3D hand gestures and integrate with the Deep Deterministic Policy Gradient(DDPG)to receive the best reward and take actions according to 3D hand gestures input.The UAV consist of a Jetson Nano embedded testbed,Global Positioning System(GPS)sensor module,and Intel depth camera.The collision avoidance system based on the polar mask segmentation technique detects the obstacles and decides the best path according to the designed reward function.The analysis of the results has been observed providing best accuracy and computational time using novel design framework when compared with traditional Proportional Integral Derivatives(PID)flight controller.There are six reward functions estimated for 2500,5000,7500,and 10000 episodes of training,which have been normalized between 0 to−4000.The best observation has been captured on 2500 episodes where the rewards are calculated for maximum value.The achieved training accuracy of polar mask segmentation for collision avoidance is 86.36%.
基金This research was supported by a Grant(2021R1F1A1063634)of the Basic Science Research Program through the National Research Foundation(NRF)funded by the Ministry of Education,Republic of Korea.
文摘In this research work,an efficient sign language recognition tool for e-learning has been proposed with a new type of feature set based on angle and lines.This feature set has the ability to increase the overall performance of machine learning algorithms in an efficient way.The hand gesture recognition based on these features has been implemented for usage in real-time.The feature set used hand landmarks,which were generated using media-pipe(MediaPipe)and open computer vision(openCV)on each frame of the incoming video.The overall algorithm has been tested on two well-known ASLalphabet(American Sign Language)and ISL-HS(Irish Sign Language)sign language datasets.Different machine learning classifiers including random forest,decision tree,and naïve Bayesian have been used to classify hand gestures using this unique feature set and their respective results have been compared.Since the random forest classifier performed better,it has been selected as the base classifier for the proposed system.It showed 96.7%accuracy with ISL-HS and 93.7%accuracy with ASL-alphabet dataset using the extracted features.
文摘Brain-computer interfaces (BCIs) records brain activity using electroencephalogram (EEG) headsets in the form of EEG signals;these signals can berecorded, processed and classified into different hand movements, which can beused to control other IoT devices. Classification of hand movements will beone step closer to applying these algorithms in real-life situations using EEGheadsets. This paper uses different feature extraction techniques and sophisticatedmachine learning algorithms to classify hand movements from EEG brain signalsto control prosthetic hands for amputated persons. To achieve good classificationaccuracy, denoising and feature extraction of EEG signals is a significant step. Wesaw a considerable increase in all the machine learning models when the movingaverage filter was applied to the raw EEG data. Feature extraction techniques likea fast fourier transform (FFT) and continuous wave transform (CWT) were usedin this study;three types of features were extracted, i.e., FFT Features, CWTCoefficients and CWT scalogram images. We trained and compared differentmachine learning (ML) models like logistic regression, random forest, k-nearestneighbors (KNN), light gradient boosting machine (GBM) and XG boost onFFT and CWT features and deep learning (DL) models like VGG-16, DenseNet201 and ResNet50 trained on CWT scalogram images. XG Boost with FFTfeatures gave the maximum accuracy of 88%.
文摘With the advancement of technology and the increase in user demands, gesture recognition played a pivotal role in the field of human-computer interaction. Among various sensing devices, Time-of-Flight (ToF) sensors were widely applied due to their low cost. This paper explored the implementation of a human hand posture recognition system using ToF sensors and residual neural networks. Firstly, this paper reviewed the typical applications of human hand recognition. Secondly, this paper designed a hand gesture recognition system using a ToF sensor VL53L5. Subsequently, data preprocessing was conducted, followed by training the constructed residual neural network. Then, the recognition results were analyzed, indicating that gesture recognition based on the residual neural network achieved an accuracy of 98.5% in a 5-class classification scenario. Finally, the paper discussed existing issues and future research directions.
文摘Deep brain stimulation offers an advanced means of treating Parkinson’s disease in a patient specific context. However, a considerable challenge is the process of ascertaining an optimal parameter configuration. Imperative for the deep brain stimulation parameter optimization process is the quantification of response feedback. As a significant improvement to traditional ordinal scale techniques is the advent of wearable and wireless systems. Recently conformal wearable and wireless systems with a profile on the order of a bandage have been developed. Previous research endeavors have successfully differentiated between deep brain stimulation “On” and “Off” status through quantification using wearable and wireless inertial sensor systems. However, the opportunity exists to further evolve to an objectively quantified response to an assortment of parameter configurations, such as the variation of amplitude, for the deep brain stimulation system. Multiple deep brain stimulation amplitude settings are considered inclusive of “Off” status as a baseline, 1.0 mA, 2.5 mA, and 4.0 mA. The quantified response of this assortment of amplitude settings is acquired through a conformal wearable and wireless inertial sensor system and consolidated using Python software automation to a feature set amenable for machine learning. Five machine learning algorithms are evaluated: J48 decision tree, K-nearest neighbors, support vector machine, logistic regression, and random forest. The performance of these machine learning algorithms is established based on the classification accuracy to distinguish between the deep brain stimulation amplitude settings and the time to develop the machine learning model. The support vector machine achieves the greatest classification accuracy, which is the primary performance parameter, and <span style="font-family:Verdana;">K-nearest neighbors achieves considerable classification accuracy with minimal time to develop the machine learning model.</span>