While progress has been made in information source localization,it has overlooked the prevalent friend and adversarial relationships in social networks.This paper addresses this gap by focusing on source localization ...While progress has been made in information source localization,it has overlooked the prevalent friend and adversarial relationships in social networks.This paper addresses this gap by focusing on source localization in signed network models.Leveraging the topological characteristics of signed networks and transforming the propagation probability into effective distance,we propose an optimization method for observer selection.Additionally,by using the reverse propagation algorithm we present a method for information source localization in signed networks.Extensive experimental results demonstrate that a higher proportion of positive edges within signed networks contributes to more favorable source localization,and the higher the ratio of propagation rates between positive and negative edges,the more accurate the source localization becomes.Interestingly,this aligns with our observation that,in reality,the number of friends tends to be greater than the number of adversaries,and the likelihood of information propagation among friends is often higher than among adversaries.In addition,the source located at the periphery of the network is not easy to identify.Furthermore,our proposed observer selection method based on effective distance achieves higher operational efficiency and exhibits higher accuracy in information source localization,compared with three strategies for observer selection based on the classical full-order neighbor coverage.展开更多
Deaf people or people facing hearing issues can communicate using sign language(SL),a visual language.Many works based on rich source language have been proposed;however,the work using poor resource language is still ...Deaf people or people facing hearing issues can communicate using sign language(SL),a visual language.Many works based on rich source language have been proposed;however,the work using poor resource language is still lacking.Unlike other SLs,the visuals of the Urdu Language are different.This study presents a novel approach to translating Urdu sign language(UrSL)using the UrSL-CNN model,a convolutional neural network(CNN)architecture specifically designed for this purpose.Unlike existingworks that primarily focus on languageswith rich resources,this study addresses the challenge of translating a sign language with limited resources.We conducted experiments using two datasets containing 1500 and 78,000 images,employing a methodology comprising four modules:data collection,pre-processing,categorization,and prediction.To enhance prediction accuracy,each sign image was transformed into a greyscale image and underwent noise filtering.Comparative analysis with machine learning baseline methods(support vectormachine,GaussianNaive Bayes,randomforest,and k-nearest neighbors’algorithm)on the UrSL alphabets dataset demonstrated the superiority of UrSL-CNN,achieving an accuracy of 0.95.Additionally,our model exhibited superior performance in Precision,Recall,and F1-score evaluations.This work not only contributes to advancing sign language translation but also holds promise for improving communication accessibility for individuals with hearing impairments.展开更多
Sign language,a visual-gestural language used by the deaf and hard-of-hearing community,plays a crucial role in facilitating communication and promoting inclusivity.Sign language recognition(SLR),the process of automa...Sign language,a visual-gestural language used by the deaf and hard-of-hearing community,plays a crucial role in facilitating communication and promoting inclusivity.Sign language recognition(SLR),the process of automatically recognizing and interpreting sign language gestures,has gained significant attention in recent years due to its potential to bridge the communication gap between the hearing impaired and the hearing world.The emergence and continuous development of deep learning techniques have provided inspiration and momentum for advancing SLR.This paper presents a comprehensive and up-to-date analysis of the advancements,challenges,and opportunities in deep learning-based sign language recognition,focusing on the past five years of research.We explore various aspects of SLR,including sign data acquisition technologies,sign language datasets,evaluation methods,and different types of neural networks.Convolutional Neural Networks(CNN)and Recurrent Neural Networks(RNN)have shown promising results in fingerspelling and isolated sign recognition.However,the continuous nature of sign language poses challenges,leading to the exploration of advanced neural network models such as the Transformer model for continuous sign language recognition(CSLR).Despite significant advancements,several challenges remain in the field of SLR.These challenges include expanding sign language datasets,achieving user independence in recognition systems,exploring different input modalities,effectively fusing features,modeling co-articulation,and improving semantic and syntactic understanding.Additionally,developing lightweight network architectures for mobile applications is crucial for practical implementation.By addressing these challenges,we can further advance the field of deep learning for sign language recognition and improve communication for the hearing-impaired community.展开更多
Sign language recognition is vital for enhancing communication accessibility among the Deaf and hard-of-hearing communities.In Japan,approximately 360,000 individualswith hearing and speech disabilities rely on Japane...Sign language recognition is vital for enhancing communication accessibility among the Deaf and hard-of-hearing communities.In Japan,approximately 360,000 individualswith hearing and speech disabilities rely on Japanese Sign Language(JSL)for communication.However,existing JSL recognition systems have faced significant performance limitations due to inherent complexities.In response to these challenges,we present a novel JSL recognition system that employs a strategic fusion approach,combining joint skeleton-based handcrafted features and pixel-based deep learning features.Our system incorporates two distinct streams:the first stream extracts crucial handcrafted features,emphasizing the capture of hand and body movements within JSL gestures.Simultaneously,a deep learning-based transfer learning stream captures hierarchical representations of JSL gestures in the second stream.Then,we concatenated the critical information of the first stream and the hierarchy of the second stream features to produce the multiple levels of the fusion features,aiming to create a comprehensive representation of the JSL gestures.After reducing the dimensionality of the feature,a feature selection approach and a kernel-based support vector machine(SVM)were used for the classification.To assess the effectiveness of our approach,we conducted extensive experiments on our Lab JSL dataset and a publicly available Arabic sign language(ArSL)dataset.Our results unequivocally demonstrate that our fusion approach significantly enhances JSL recognition accuracy and robustness compared to individual feature sets or traditional recognition methods.展开更多
Research on Chinese Sign Language(CSL)provides convenience and support for individuals with hearing impairments to communicate and integrate into society.This article reviews the relevant literature on Chinese Sign La...Research on Chinese Sign Language(CSL)provides convenience and support for individuals with hearing impairments to communicate and integrate into society.This article reviews the relevant literature on Chinese Sign Language Recognition(CSLR)in the past 20 years.Hidden Markov Models(HMM),Support Vector Machines(SVM),and Dynamic Time Warping(DTW)were found to be the most commonly employed technologies among traditional identificationmethods.Benefiting from the rapid development of computer vision and artificial intelligence technology,Convolutional Neural Networks(CNN),3D-CNN,YOLO,Capsule Network(CapsNet)and various deep neural networks have sprung up.Deep Neural Networks(DNNs)and their derived models are integral tomodern artificial intelligence recognitionmethods.In addition,technologies thatwerewidely used in the early days have also been integrated and applied to specific hybrid models and customized identification methods.Sign language data collection includes acquiring data from data gloves,data sensors(such as Kinect,LeapMotion,etc.),and high-definition photography.Meanwhile,facial expression recognition,complex background processing,and 3D sign language recognition have also attracted research interests among scholars.Due to the uniqueness and complexity of Chinese sign language,accuracy,robustness,real-time performance,and user independence are significant challenges for future sign language recognition research.Additionally,suitable datasets and evaluation criteria are also worth pursuing.展开更多
Objective:This study aimed to determine the effectiveness of aromatherapy with lavender essential oil compared to progressive muscle relaxation(PMR)on anxiety and vital signs of patients under spinal anesthesia.Materi...Objective:This study aimed to determine the effectiveness of aromatherapy with lavender essential oil compared to progressive muscle relaxation(PMR)on anxiety and vital signs of patients under spinal anesthesia.Materials and Methods:This clinical trial was conducted on 120 spinal anesthesia candidates who were randomly assigned into three groups of 40 including control,PMR(Jacobsen group),and aromatherapy.The state-trait anxiety inventory was completed on surgery day and 15 min after the end of the intervention by the samples of all three groups,and at the same time as completing the questionnaire,vital signs were also measured and recorded.Results:The mean score of anxiety after intervention was lower than that before the intervention in the aromatherapy group(P<0.001).The mean score of anxiety in the aromatherapy group was significantly lower than that in the Jacobsen group(P<0.001).Moreover,data analysis showed a significant decrease in the mean arterial blood pressure scores of the PMR(P=008)and aromatherapy(P<0.001)groups and a statistically significant increase in the mean heart rate scores in the control group(P=0.002).Conclusion:The use of aromatherapy with lavender is more effective than PMR therapy in reducing the anxiety level of patients undergoing spinal anesthesia.Due to the high level of anxiety and its serious effects on the patient’s hemodynamics,aromatherapy with lavender can be used as an easy and cheap method to reduce anxiety in operation rooms.展开更多
This paper is concerned with the following fourth-order three-point boundary value problem , where , we discuss the existence of positive solutions to the above problem by applying to the fixed point theory in cones a...This paper is concerned with the following fourth-order three-point boundary value problem , where , we discuss the existence of positive solutions to the above problem by applying to the fixed point theory in cones and iterative technique.展开更多
There are few studies regarding imaging markers for predicting postoperative rebleeding after stereotactic minimally invasive surgery(MIS)for hypertensive intracerebral haemorrhage(ICH),and little is known about the r...There are few studies regarding imaging markers for predicting postoperative rebleeding after stereotactic minimally invasive surgery(MIS)for hypertensive intracerebral haemorrhage(ICH),and little is known about the relationship between satellite sign on computed tomography(CT)scans and postoperative rebleeding after MIS.This study aimed to determine the value of the CT satellite sign in predicting postoperative rebleeding in patients with hypertensive ICH who undergo stereotactic MIS.We retrospectively examined and analysed 105 patients with hypertensive ICH who underwent standard stereotactic MIS for hematoma evacuation within 72 h following admission.Postoperative rebleeding occurred in 14 of 65(21.5%)patients with the satellite sign on baseline CT,and in 5 of the 40(12.5%)patients without the satellite sign.This diiTerence was statistically significant.Positive and negative values of the satellite sign for predicting postoperative rebleeding were 21.5%and 87.5%,respectively.Multivariate logistic regression analysis verified that baseline ICH volume and intraventricular rupture were independent predictors of postoperative rebleeding.In conclusion,the satellite sign on baseline CT scans may not predict postoperative rebleeding following stereotactic MIS for hypertensive ICH.展开更多
The infrastructure and construction of roads are crucial for the economic and social development of a region,but traffic-related challenges like accidents and congestion persist.Artificial Intelligence(AI)and Machine ...The infrastructure and construction of roads are crucial for the economic and social development of a region,but traffic-related challenges like accidents and congestion persist.Artificial Intelligence(AI)and Machine Learning(ML)have been used in road infrastructure and construction,particularly with the Internet of Things(IoT)devices.Object detection in Computer Vision also plays a key role in improving road infrastructure and addressing trafficrelated problems.This study aims to use You Only Look Once version 7(YOLOv7),Convolutional Block Attention Module(CBAM),the most optimized object-detection algorithm,to detect and identify traffic signs,and analyze effective combinations of adaptive optimizers like Adaptive Moment estimation(Adam),Root Mean Squared Propagation(RMSprop)and Stochastic Gradient Descent(SGD)with the YOLOv7.Using a portion of German traffic signs for training,the study investigates the feasibility of adopting smaller datasets while maintaining high accuracy.The model proposed in this study not only improves traffic safety by detecting traffic signs but also has the potential to contribute to the rapid development of autonomous vehicle systems.The study results showed an impressive accuracy of 99.7%when using a batch size of 8 and the Adam optimizer.This high level of accuracy demonstrates the effectiveness of the proposed model for the image classification task of traffic sign recognition.展开更多
Aiming at solving the problem of missed detection and low accuracy in detecting traffic signs in the wild, an improved method of YOLOv8 is proposed. Firstly, combined with the characteristics of small target objects i...Aiming at solving the problem of missed detection and low accuracy in detecting traffic signs in the wild, an improved method of YOLOv8 is proposed. Firstly, combined with the characteristics of small target objects in the actual scene, this paper further adds blur and noise operation. Then, the asymptotic feature pyramid network (AFPN) is introduced to highlight the influence of key layer features after feature fusion, and simultaneously solve the direct interaction of non-adjacent layers. Experimental results on the TT100K dataset show that compared with the YOLOv8, the detection accuracy and recall are higher. .展开更多
Continuous sign language recognition(CSLR)is challenging due to the complexity of video background,hand gesture variability,and temporal modeling difficulties.This work proposes a CSLR method based on a spatialtempora...Continuous sign language recognition(CSLR)is challenging due to the complexity of video background,hand gesture variability,and temporal modeling difficulties.This work proposes a CSLR method based on a spatialtemporal graph attention network to focus on essential features of video series.The method considers local details of sign language movements by taking the information on joints and bones as inputs and constructing a spatialtemporal graph to reflect inter-frame relevance and physical connections between nodes.The graph-based multihead attention mechanism is utilized with adjacent matrix calculation for better local-feature exploration,and short-term motion correlation modeling is completed via a temporal convolutional network.We adopted BLSTM to learn the long-termdependence and connectionist temporal classification to align the word-level sequences.The proposed method achieves competitive results regarding word error rates(1.59%)on the Chinese Sign Language dataset and the mean Jaccard Index(65.78%)on the ChaLearn LAP Continuous Gesture Dataset.展开更多
Rapid advancement of intelligent transportation systems(ITS)and autonomous driving(AD)have shown the importance of accurate and efficient detection of traffic signs.However,certain drawbacks,such as balancing accuracy...Rapid advancement of intelligent transportation systems(ITS)and autonomous driving(AD)have shown the importance of accurate and efficient detection of traffic signs.However,certain drawbacks,such as balancing accuracy and real-time performance,hinder the deployment of traffic sign detection algorithms in ITS and AD domains.In this study,a novel traffic sign detection algorithm was proposed based on the bidirectional Res2Net architecture to achieve an improved balance between accuracy and speed.An enhanced backbone network module,called C2Net,which uses an upgraded bidirectional Res2Net,was introduced to mitigate information loss in the feature extraction process and to achieve information complementarity.Furthermore,a squeeze-and-excitation attention mechanism was incorporated within the channel attention of the architecture to perform channel-level feature correction on the input feature map,which effectively retains valuable features while removing non-essential features.A series of ablation experiments were conducted to validate the efficacy of the proposed methodology.The performance was evaluated using two distinct datasets:the Tsinghua-Tencent 100K and the CSUST Chinese traffic sign detection benchmark 2021.On the TT100K dataset,the method achieves precision,recall,and Map0.5 scores of 83.3%,79.3%,and 84.2%,respectively.Similarly,on the CCTSDB 2021 dataset,the method achieves precision,recall,and Map0.5 scores of 91.49%,73.79%,and 81.03%,respectively.Experimental results revealed that the proposed method had superior performance compared to conventional models,which includes the faster region-based convolutional neural network,single shot multibox detector,and you only look once version 5.展开更多
The subway is the primary travel tool for urban residents in China. Due to the complex structure of the subway and high personnel density in rush hours, subway evacuation capacity is critical. The subway evacuation mo...The subway is the primary travel tool for urban residents in China. Due to the complex structure of the subway and high personnel density in rush hours, subway evacuation capacity is critical. The subway evacuation model is explored in this work by combining the improved social force model with the view radius using the Vicsek model. The pedestrians are divided into two categories based on different force models. The first category is sensitive pedestrians who have normal responses to emergency signs. The second category is insensitive pedestrians. By simulating different proportions of the insensitive pedestrians, we find that the escape time is directly proportional to the number of insensitive pedestrians and inversely proportional to the view radius. However, when the view radius is large enough, the escape time does not change significantly, and the evacuation of people in a small view radius environment tends to be integrated. With the improvement of view radius conditions, the escape time changes more obviously with the proportion of insensitive pedestrians. A new emergency sign layout is proposed, and the simulations show that the proposed layout can effectively reduce the escape time in a small view radius environment. However, the evacuation effect of the new escape sign layout on the large view radius environment is not apparent. In this case, the exit setting emerges as an additional factor affecting the escape time.展开更多
With advancements in computing powers and the overall quality of images captured on everyday cameras,a much wider range of possibilities has opened in various scenarios.This fact has several implications for deaf and ...With advancements in computing powers and the overall quality of images captured on everyday cameras,a much wider range of possibilities has opened in various scenarios.This fact has several implications for deaf and dumb people as they have a chance to communicate with a greater number of people much easier.More than ever before,there is a plethora of info about sign language usage in the real world.Sign languages,and by extension the datasets available,are of two forms,isolated sign language and continuous sign language.The main difference between the two types is that in isolated sign language,the hand signs cover individual letters of the alphabet.In continuous sign language,entire words’hand signs are used.This paper will explore a novel deep learning architecture that will use recently published large pre-trained image models to quickly and accurately recognize the alphabets in the American Sign Language(ASL).The study will focus on isolated sign language to demonstrate that it is possible to achieve a high level of classification accuracy on the data,thereby showing that interpreters can be implemented in the real world.The newly proposed Mobile-NetV2 architecture serves as the backbone of this study.It is designed to run on end devices like mobile phones and infer signals(what does it infer)from images in a relatively short amount of time.With the proposed architecture in this paper,the classification accuracy of 98.77%in the Indian Sign Language(ISL)and American Sign Language(ASL)is achieved,outperforming the existing state-of-the-art systems.展开更多
Communication between people with disabilities and people who do not understand sign language is a growing social need and can be a tedious task.One of the main functions of sign language is to communicate with each o...Communication between people with disabilities and people who do not understand sign language is a growing social need and can be a tedious task.One of the main functions of sign language is to communicate with each other through hand gestures.Recognition of hand gestures has become an important challenge for the recognition of sign language.There are many existing models that can produce a good accuracy,but if the model test with rotated or translated images,they may face some difficulties to make good performance accuracy.To resolve these challenges of hand gesture recognition,we proposed a Rotation,Translation and Scale-invariant sign word recognition system using a convolu-tional neural network(CNN).We have followed three steps in our work:rotated,translated and scaled(RTS)version dataset generation,gesture segmentation,and sign word classification.Firstly,we have enlarged a benchmark dataset of 20 sign words by making different amounts of Rotation,Translation and Scale of the ori-ginal images to create the RTS version dataset.Then we have applied the gesture segmentation technique.The segmentation consists of three levels,i)Otsu Thresholding with YCbCr,ii)Morphological analysis:dilation through opening morphology and iii)Watershed algorithm.Finally,our designed CNN model has been trained to classify the hand gesture as well as the sign word.Our model has been evaluated using the twenty sign word dataset,five sign word dataset and the RTS version of these datasets.We achieved 99.30%accuracy from the twenty sign word dataset evaluation,99.10%accuracy from the RTS version of the twenty sign word evolution,100%accuracy from thefive sign word dataset evaluation,and 98.00%accuracy from the RTS versionfive sign word dataset evolution.Furthermore,the influence of our model exists in competitive results with state-of-the-art methods in sign word recognition.展开更多
We study the following Schrodinger-Poisson system where (Pλ){-△u+ V(x)u+λФ(x)u^p=x∈R^3,-△Ф=u^2,lim│x│→∞Ф(x) =0,u〉0,where λ≥0 is a parameter,1 〈 p 〈 +∞, V(x) and Q(x)=1 ,D.Ruiz[19] prov...We study the following Schrodinger-Poisson system where (Pλ){-△u+ V(x)u+λФ(x)u^p=x∈R^3,-△Ф=u^2,lim│x│→∞Ф(x) =0,u〉0,where λ≥0 is a parameter,1 〈 p 〈 +∞, V(x) and Q(x)=1 ,D.Ruiz[19] proved that(Pλ)with p∈ (2, 5) has always a positive radial solution, but (Pλ) with p E (1, 2] has solution only if λ 〉 0 small enough and no any nontrivial solution if λ≥1/4.By using sub-supersolution method,we prove that there exists λ0〉0 such that(Pλ)with p ∈(1+∞)has alaways a bound state(H^1(R^3)solution for λ∈[0,λ0)and certain functions V(x)and Q(x)in L^∞(R^3).Moreover,for every λ∈[0,λ0),the solutions uλ of (Pλ)converges,along a subsequence,to a solution of (P0)in H^1 as λ→0展开更多
In this article, we study the following nonhomogeneous Schrodinger-Poissone quations{-△u+λV(x)u+K(x)Фu=f(x,u)+g(x),x∈R^3,-△Ф=k(x)u^2, x∈R^3}where λ 〉 0 is a parameter. Under some suitable assumpt...In this article, we study the following nonhomogeneous Schrodinger-Poissone quations{-△u+λV(x)u+K(x)Фu=f(x,u)+g(x),x∈R^3,-△Ф=k(x)u^2, x∈R^3}where λ 〉 0 is a parameter. Under some suitable assumptions on 11, K, f and g, the existence of multiple solutions is proved by using the Ekeland's variational principle and the Mountain Pass Theorem in critical point theory. In particular, the potential V is allowed to be signchanging.展开更多
基金Project supported by the National Natural Science Foundation of China(Grant Nos.62103375 and 62006106)the Zhejiang Provincial Philosophy and Social Science Planning Project(Grant No.22NDJC009Z)+1 种基金the Education Ministry Humanities and Social Science Foundation of China(Grant Nos.19YJCZH056 and 21YJC630120)the Natural Science Foundation of Zhejiang Province of China(Grant Nos.LY23F030003 and LQ21F020005).
文摘While progress has been made in information source localization,it has overlooked the prevalent friend and adversarial relationships in social networks.This paper addresses this gap by focusing on source localization in signed network models.Leveraging the topological characteristics of signed networks and transforming the propagation probability into effective distance,we propose an optimization method for observer selection.Additionally,by using the reverse propagation algorithm we present a method for information source localization in signed networks.Extensive experimental results demonstrate that a higher proportion of positive edges within signed networks contributes to more favorable source localization,and the higher the ratio of propagation rates between positive and negative edges,the more accurate the source localization becomes.Interestingly,this aligns with our observation that,in reality,the number of friends tends to be greater than the number of adversaries,and the likelihood of information propagation among friends is often higher than among adversaries.In addition,the source located at the periphery of the network is not easy to identify.Furthermore,our proposed observer selection method based on effective distance achieves higher operational efficiency and exhibits higher accuracy in information source localization,compared with three strategies for observer selection based on the classical full-order neighbor coverage.
文摘Deaf people or people facing hearing issues can communicate using sign language(SL),a visual language.Many works based on rich source language have been proposed;however,the work using poor resource language is still lacking.Unlike other SLs,the visuals of the Urdu Language are different.This study presents a novel approach to translating Urdu sign language(UrSL)using the UrSL-CNN model,a convolutional neural network(CNN)architecture specifically designed for this purpose.Unlike existingworks that primarily focus on languageswith rich resources,this study addresses the challenge of translating a sign language with limited resources.We conducted experiments using two datasets containing 1500 and 78,000 images,employing a methodology comprising four modules:data collection,pre-processing,categorization,and prediction.To enhance prediction accuracy,each sign image was transformed into a greyscale image and underwent noise filtering.Comparative analysis with machine learning baseline methods(support vectormachine,GaussianNaive Bayes,randomforest,and k-nearest neighbors’algorithm)on the UrSL alphabets dataset demonstrated the superiority of UrSL-CNN,achieving an accuracy of 0.95.Additionally,our model exhibited superior performance in Precision,Recall,and F1-score evaluations.This work not only contributes to advancing sign language translation but also holds promise for improving communication accessibility for individuals with hearing impairments.
基金supported from the National Philosophy and Social Sciences Foundation(Grant No.20BTQ065).
文摘Sign language,a visual-gestural language used by the deaf and hard-of-hearing community,plays a crucial role in facilitating communication and promoting inclusivity.Sign language recognition(SLR),the process of automatically recognizing and interpreting sign language gestures,has gained significant attention in recent years due to its potential to bridge the communication gap between the hearing impaired and the hearing world.The emergence and continuous development of deep learning techniques have provided inspiration and momentum for advancing SLR.This paper presents a comprehensive and up-to-date analysis of the advancements,challenges,and opportunities in deep learning-based sign language recognition,focusing on the past five years of research.We explore various aspects of SLR,including sign data acquisition technologies,sign language datasets,evaluation methods,and different types of neural networks.Convolutional Neural Networks(CNN)and Recurrent Neural Networks(RNN)have shown promising results in fingerspelling and isolated sign recognition.However,the continuous nature of sign language poses challenges,leading to the exploration of advanced neural network models such as the Transformer model for continuous sign language recognition(CSLR).Despite significant advancements,several challenges remain in the field of SLR.These challenges include expanding sign language datasets,achieving user independence in recognition systems,exploring different input modalities,effectively fusing features,modeling co-articulation,and improving semantic and syntactic understanding.Additionally,developing lightweight network architectures for mobile applications is crucial for practical implementation.By addressing these challenges,we can further advance the field of deep learning for sign language recognition and improve communication for the hearing-impaired community.
基金supported by the Competitive Research Fund of the University of Aizu,Japan.
文摘Sign language recognition is vital for enhancing communication accessibility among the Deaf and hard-of-hearing communities.In Japan,approximately 360,000 individualswith hearing and speech disabilities rely on Japanese Sign Language(JSL)for communication.However,existing JSL recognition systems have faced significant performance limitations due to inherent complexities.In response to these challenges,we present a novel JSL recognition system that employs a strategic fusion approach,combining joint skeleton-based handcrafted features and pixel-based deep learning features.Our system incorporates two distinct streams:the first stream extracts crucial handcrafted features,emphasizing the capture of hand and body movements within JSL gestures.Simultaneously,a deep learning-based transfer learning stream captures hierarchical representations of JSL gestures in the second stream.Then,we concatenated the critical information of the first stream and the hierarchy of the second stream features to produce the multiple levels of the fusion features,aiming to create a comprehensive representation of the JSL gestures.After reducing the dimensionality of the feature,a feature selection approach and a kernel-based support vector machine(SVM)were used for the classification.To assess the effectiveness of our approach,we conducted extensive experiments on our Lab JSL dataset and a publicly available Arabic sign language(ArSL)dataset.Our results unequivocally demonstrate that our fusion approach significantly enhances JSL recognition accuracy and robustness compared to individual feature sets or traditional recognition methods.
基金supported by National Social Science Foundation Annual Project“Research on Evaluation and Improvement Paths of Integrated Development of Disabled Persons”(Grant No.20BRK029)the National Language Commission’s“14th Five-Year Plan”Scientific Research Plan 2023 Project“Domain Digital Language Service Resource Construction and Key Technology Research”(YB145-72)the National Philosophy and Social Sciences Foundation(Grant No.20BTQ065).
文摘Research on Chinese Sign Language(CSL)provides convenience and support for individuals with hearing impairments to communicate and integrate into society.This article reviews the relevant literature on Chinese Sign Language Recognition(CSLR)in the past 20 years.Hidden Markov Models(HMM),Support Vector Machines(SVM),and Dynamic Time Warping(DTW)were found to be the most commonly employed technologies among traditional identificationmethods.Benefiting from the rapid development of computer vision and artificial intelligence technology,Convolutional Neural Networks(CNN),3D-CNN,YOLO,Capsule Network(CapsNet)and various deep neural networks have sprung up.Deep Neural Networks(DNNs)and their derived models are integral tomodern artificial intelligence recognitionmethods.In addition,technologies thatwerewidely used in the early days have also been integrated and applied to specific hybrid models and customized identification methods.Sign language data collection includes acquiring data from data gloves,data sensors(such as Kinect,LeapMotion,etc.),and high-definition photography.Meanwhile,facial expression recognition,complex background processing,and 3D sign language recognition have also attracted research interests among scholars.Due to the uniqueness and complexity of Chinese sign language,accuracy,robustness,real-time performance,and user independence are significant challenges for future sign language recognition research.Additionally,suitable datasets and evaluation criteria are also worth pursuing.
基金financially supported by Arak University of Medical Sciences.
文摘Objective:This study aimed to determine the effectiveness of aromatherapy with lavender essential oil compared to progressive muscle relaxation(PMR)on anxiety and vital signs of patients under spinal anesthesia.Materials and Methods:This clinical trial was conducted on 120 spinal anesthesia candidates who were randomly assigned into three groups of 40 including control,PMR(Jacobsen group),and aromatherapy.The state-trait anxiety inventory was completed on surgery day and 15 min after the end of the intervention by the samples of all three groups,and at the same time as completing the questionnaire,vital signs were also measured and recorded.Results:The mean score of anxiety after intervention was lower than that before the intervention in the aromatherapy group(P<0.001).The mean score of anxiety in the aromatherapy group was significantly lower than that in the Jacobsen group(P<0.001).Moreover,data analysis showed a significant decrease in the mean arterial blood pressure scores of the PMR(P=008)and aromatherapy(P<0.001)groups and a statistically significant increase in the mean heart rate scores in the control group(P=0.002).Conclusion:The use of aromatherapy with lavender is more effective than PMR therapy in reducing the anxiety level of patients undergoing spinal anesthesia.Due to the high level of anxiety and its serious effects on the patient’s hemodynamics,aromatherapy with lavender can be used as an easy and cheap method to reduce anxiety in operation rooms.
文摘This paper is concerned with the following fourth-order three-point boundary value problem , where , we discuss the existence of positive solutions to the above problem by applying to the fixed point theory in cones and iterative technique.
文摘There are few studies regarding imaging markers for predicting postoperative rebleeding after stereotactic minimally invasive surgery(MIS)for hypertensive intracerebral haemorrhage(ICH),and little is known about the relationship between satellite sign on computed tomography(CT)scans and postoperative rebleeding after MIS.This study aimed to determine the value of the CT satellite sign in predicting postoperative rebleeding in patients with hypertensive ICH who undergo stereotactic MIS.We retrospectively examined and analysed 105 patients with hypertensive ICH who underwent standard stereotactic MIS for hematoma evacuation within 72 h following admission.Postoperative rebleeding occurred in 14 of 65(21.5%)patients with the satellite sign on baseline CT,and in 5 of the 40(12.5%)patients without the satellite sign.This diiTerence was statistically significant.Positive and negative values of the satellite sign for predicting postoperative rebleeding were 21.5%and 87.5%,respectively.Multivariate logistic regression analysis verified that baseline ICH volume and intraventricular rupture were independent predictors of postoperative rebleeding.In conclusion,the satellite sign on baseline CT scans may not predict postoperative rebleeding following stereotactic MIS for hypertensive ICH.
文摘The infrastructure and construction of roads are crucial for the economic and social development of a region,but traffic-related challenges like accidents and congestion persist.Artificial Intelligence(AI)and Machine Learning(ML)have been used in road infrastructure and construction,particularly with the Internet of Things(IoT)devices.Object detection in Computer Vision also plays a key role in improving road infrastructure and addressing trafficrelated problems.This study aims to use You Only Look Once version 7(YOLOv7),Convolutional Block Attention Module(CBAM),the most optimized object-detection algorithm,to detect and identify traffic signs,and analyze effective combinations of adaptive optimizers like Adaptive Moment estimation(Adam),Root Mean Squared Propagation(RMSprop)and Stochastic Gradient Descent(SGD)with the YOLOv7.Using a portion of German traffic signs for training,the study investigates the feasibility of adopting smaller datasets while maintaining high accuracy.The model proposed in this study not only improves traffic safety by detecting traffic signs but also has the potential to contribute to the rapid development of autonomous vehicle systems.The study results showed an impressive accuracy of 99.7%when using a batch size of 8 and the Adam optimizer.This high level of accuracy demonstrates the effectiveness of the proposed model for the image classification task of traffic sign recognition.
文摘Aiming at solving the problem of missed detection and low accuracy in detecting traffic signs in the wild, an improved method of YOLOv8 is proposed. Firstly, combined with the characteristics of small target objects in the actual scene, this paper further adds blur and noise operation. Then, the asymptotic feature pyramid network (AFPN) is introduced to highlight the influence of key layer features after feature fusion, and simultaneously solve the direct interaction of non-adjacent layers. Experimental results on the TT100K dataset show that compared with the YOLOv8, the detection accuracy and recall are higher. .
基金supported by the Key Research&Development Plan Project of Shandong Province,China(No.2017GGX10127).
文摘Continuous sign language recognition(CSLR)is challenging due to the complexity of video background,hand gesture variability,and temporal modeling difficulties.This work proposes a CSLR method based on a spatialtemporal graph attention network to focus on essential features of video series.The method considers local details of sign language movements by taking the information on joints and bones as inputs and constructing a spatialtemporal graph to reflect inter-frame relevance and physical connections between nodes.The graph-based multihead attention mechanism is utilized with adjacent matrix calculation for better local-feature exploration,and short-term motion correlation modeling is completed via a temporal convolutional network.We adopted BLSTM to learn the long-termdependence and connectionist temporal classification to align the word-level sequences.The proposed method achieves competitive results regarding word error rates(1.59%)on the Chinese Sign Language dataset and the mean Jaccard Index(65.78%)on the ChaLearn LAP Continuous Gesture Dataset.
基金funded by the National Key R&D Program of China,Grant Number 2017YFB0802803Beijing Natural Science Foundation,Grant Number 4202002.
文摘Rapid advancement of intelligent transportation systems(ITS)and autonomous driving(AD)have shown the importance of accurate and efficient detection of traffic signs.However,certain drawbacks,such as balancing accuracy and real-time performance,hinder the deployment of traffic sign detection algorithms in ITS and AD domains.In this study,a novel traffic sign detection algorithm was proposed based on the bidirectional Res2Net architecture to achieve an improved balance between accuracy and speed.An enhanced backbone network module,called C2Net,which uses an upgraded bidirectional Res2Net,was introduced to mitigate information loss in the feature extraction process and to achieve information complementarity.Furthermore,a squeeze-and-excitation attention mechanism was incorporated within the channel attention of the architecture to perform channel-level feature correction on the input feature map,which effectively retains valuable features while removing non-essential features.A series of ablation experiments were conducted to validate the efficacy of the proposed methodology.The performance was evaluated using two distinct datasets:the Tsinghua-Tencent 100K and the CSUST Chinese traffic sign detection benchmark 2021.On the TT100K dataset,the method achieves precision,recall,and Map0.5 scores of 83.3%,79.3%,and 84.2%,respectively.Similarly,on the CCTSDB 2021 dataset,the method achieves precision,recall,and Map0.5 scores of 91.49%,73.79%,and 81.03%,respectively.Experimental results revealed that the proposed method had superior performance compared to conventional models,which includes the faster region-based convolutional neural network,single shot multibox detector,and you only look once version 5.
基金supported by the National Natural Science Foundation of China (Grant Nos. 51874183 and 51874182)the National Key Research and Development Program of China (Grant No. 2018YFC0809300)。
文摘The subway is the primary travel tool for urban residents in China. Due to the complex structure of the subway and high personnel density in rush hours, subway evacuation capacity is critical. The subway evacuation model is explored in this work by combining the improved social force model with the view radius using the Vicsek model. The pedestrians are divided into two categories based on different force models. The first category is sensitive pedestrians who have normal responses to emergency signs. The second category is insensitive pedestrians. By simulating different proportions of the insensitive pedestrians, we find that the escape time is directly proportional to the number of insensitive pedestrians and inversely proportional to the view radius. However, when the view radius is large enough, the escape time does not change significantly, and the evacuation of people in a small view radius environment tends to be integrated. With the improvement of view radius conditions, the escape time changes more obviously with the proportion of insensitive pedestrians. A new emergency sign layout is proposed, and the simulations show that the proposed layout can effectively reduce the escape time in a small view radius environment. However, the evacuation effect of the new escape sign layout on the large view radius environment is not apparent. In this case, the exit setting emerges as an additional factor affecting the escape time.
文摘With advancements in computing powers and the overall quality of images captured on everyday cameras,a much wider range of possibilities has opened in various scenarios.This fact has several implications for deaf and dumb people as they have a chance to communicate with a greater number of people much easier.More than ever before,there is a plethora of info about sign language usage in the real world.Sign languages,and by extension the datasets available,are of two forms,isolated sign language and continuous sign language.The main difference between the two types is that in isolated sign language,the hand signs cover individual letters of the alphabet.In continuous sign language,entire words’hand signs are used.This paper will explore a novel deep learning architecture that will use recently published large pre-trained image models to quickly and accurately recognize the alphabets in the American Sign Language(ASL).The study will focus on isolated sign language to demonstrate that it is possible to achieve a high level of classification accuracy on the data,thereby showing that interpreters can be implemented in the real world.The newly proposed Mobile-NetV2 architecture serves as the backbone of this study.It is designed to run on end devices like mobile phones and infer signals(what does it infer)from images in a relatively short amount of time.With the proposed architecture in this paper,the classification accuracy of 98.77%in the Indian Sign Language(ISL)and American Sign Language(ASL)is achieved,outperforming the existing state-of-the-art systems.
基金This work was supported by the Competitive Research Fund of The University of Aizu,Japan.
文摘Communication between people with disabilities and people who do not understand sign language is a growing social need and can be a tedious task.One of the main functions of sign language is to communicate with each other through hand gestures.Recognition of hand gestures has become an important challenge for the recognition of sign language.There are many existing models that can produce a good accuracy,but if the model test with rotated or translated images,they may face some difficulties to make good performance accuracy.To resolve these challenges of hand gesture recognition,we proposed a Rotation,Translation and Scale-invariant sign word recognition system using a convolu-tional neural network(CNN).We have followed three steps in our work:rotated,translated and scaled(RTS)version dataset generation,gesture segmentation,and sign word classification.Firstly,we have enlarged a benchmark dataset of 20 sign words by making different amounts of Rotation,Translation and Scale of the ori-ginal images to create the RTS version dataset.Then we have applied the gesture segmentation technique.The segmentation consists of three levels,i)Otsu Thresholding with YCbCr,ii)Morphological analysis:dilation through opening morphology and iii)Watershed algorithm.Finally,our designed CNN model has been trained to classify the hand gesture as well as the sign word.Our model has been evaluated using the twenty sign word dataset,five sign word dataset and the RTS version of these datasets.We achieved 99.30%accuracy from the twenty sign word dataset evaluation,99.10%accuracy from the RTS version of the twenty sign word evolution,100%accuracy from thefive sign word dataset evaluation,and 98.00%accuracy from the RTS versionfive sign word dataset evolution.Furthermore,the influence of our model exists in competitive results with state-of-the-art methods in sign word recognition.
基金Supported by NSFC(10631030) and CAS-KJCX3-SYW-S03
文摘We study the following Schrodinger-Poisson system where (Pλ){-△u+ V(x)u+λФ(x)u^p=x∈R^3,-△Ф=u^2,lim│x│→∞Ф(x) =0,u〉0,where λ≥0 is a parameter,1 〈 p 〈 +∞, V(x) and Q(x)=1 ,D.Ruiz[19] proved that(Pλ)with p∈ (2, 5) has always a positive radial solution, but (Pλ) with p E (1, 2] has solution only if λ 〉 0 small enough and no any nontrivial solution if λ≥1/4.By using sub-supersolution method,we prove that there exists λ0〉0 such that(Pλ)with p ∈(1+∞)has alaways a bound state(H^1(R^3)solution for λ∈[0,λ0)and certain functions V(x)and Q(x)in L^∞(R^3).Moreover,for every λ∈[0,λ0),the solutions uλ of (Pλ)converges,along a subsequence,to a solution of (P0)in H^1 as λ→0
基金supported by the Tianyuan Special Foundation(11526148)the second author is supported by the National Natural Science Foundation of China(11571187)
文摘In this article, we study the following nonhomogeneous Schrodinger-Poissone quations{-△u+λV(x)u+K(x)Фu=f(x,u)+g(x),x∈R^3,-△Ф=k(x)u^2, x∈R^3}where λ 〉 0 is a parameter. Under some suitable assumptions on 11, K, f and g, the existence of multiple solutions is proved by using the Ekeland's variational principle and the Mountain Pass Theorem in critical point theory. In particular, the potential V is allowed to be signchanging.