The use of hand gestures can be the most intuitive human-machine interaction medium.The early approaches for hand gesture recognition used device-based methods.These methods use mechanical or optical sensors attached ...The use of hand gestures can be the most intuitive human-machine interaction medium.The early approaches for hand gesture recognition used device-based methods.These methods use mechanical or optical sensors attached to a glove or markers,which hinder the natural human-machine communication.On the other hand,vision-based methods are less restrictive and allow for a more spontaneous communication without the need of an intermediary between human and machine.Therefore,vision gesture recognition has been a popular area of research for the past thirty years.Hand gesture recognition finds its application in many areas,particularly the automotive industry where advanced automotive human-machine interface(HMI)designers are using gesture recognition to improve driver and vehicle safety.However,technology advances go beyond active/passive safety and into convenience and comfort.In this context,one of America’s big three automakers has partnered with the Centre of Pattern Analysis and Machine Intelligence(CPAMI)at the University of Waterloo to investigate expanding their product segment through machine learning to provide an increased driver convenience and comfort with the particular application of hand gesture recognition for autonomous car parking.The present paper leverages the state-of-the-art deep learning and optimization techniques to develop a vision-based multiview dynamic hand gesture recognizer for a self-parking system.We propose a 3D-CNN gesture model architecture that we train on a publicly available hand gesture database.We apply transfer learning methods to fine-tune the pre-trained gesture model on custom-made data,which significantly improves the proposed system performance in a real world environment.We adapt the architecture of end-to-end solution to expand the state-of-the-art video classifier from a single image as input(fed by monocular camera)to a Multiview 360 feed,offered by a six cameras module.Finally,we optimize the proposed solution to work on a limited resource embedded platform(Nvidia Jetson TX2)that is used by automakers for vehicle-based features,without sacrificing the accuracy robustness and real time functionality of the system.展开更多
Hand gesture recognition (HGR) is used in a numerous applications,including medical health-care, industrial purpose and sports detection.We have developed a real-time hand gesture recognition system using inertialsens...Hand gesture recognition (HGR) is used in a numerous applications,including medical health-care, industrial purpose and sports detection.We have developed a real-time hand gesture recognition system using inertialsensors for the smart home application. Developing such a model facilitatesthe medical health field (elders or disabled ones). Home automation has alsobeen proven to be a tremendous benefit for the elderly and disabled. Residentsare admitted to smart homes for comfort, luxury, improved quality of life,and protection against intrusion and burglars. This paper proposes a novelsystem that uses principal component analysis, linear discrimination analysisfeature extraction, and random forest as a classifier to improveHGRaccuracy.We have achieved an accuracy of 94% over the publicly benchmarked HGRdataset. The proposed system can be used to detect hand gestures in thehealthcare industry as well as in the industrial and educational sectors.展开更多
Dynamic hand gesture recognition is a desired alternative means for human-computer interactions.This paper presents a hand gesture recognition system that is designed for the control of flights of unmanned aerial vehi...Dynamic hand gesture recognition is a desired alternative means for human-computer interactions.This paper presents a hand gesture recognition system that is designed for the control of flights of unmanned aerial vehicles(UAV).A data representation model that represents a dynamic gesture sequence by converting the 4-D spatiotemporal data to 2-D matrix and a 1-D array is introduced.To train the system to recognize designed gestures,skeleton data collected from a Leap Motion Controller are converted to two different data models.As many as 9124 samples of the training dataset,1938 samples of the testing dataset are created to train and test the proposed three deep learning neural networks,which are a 2-layer fully connected neural network,a 5-layer fully connected neural network and an 8-layer convolutional neural network.The static testing results show that the 2-layer fully connected neural network achieves an average accuracy of 96.7%on scaled datasets and 12.3%on non-scaled datasets.The 5-layer fully connected neural network achieves an average accuracy of 98.0%on scaled datasets and 89.1%on non-scaled datasets.The 8-layer convolutional neural network achieves an average accuracy of 89.6%on scaled datasets and 96.9%on non-scaled datasets.Testing on a drone-kit simulator and a real drone shows that this system is feasible for drone flight controls.展开更多
Individuals with cerebral palsy and muscular dystrophy often lack fine motor control of their fingers which makes it difficult to control traditional powered wheelchairs using a joystick.Studies have shown the use of ...Individuals with cerebral palsy and muscular dystrophy often lack fine motor control of their fingers which makes it difficult to control traditional powered wheelchairs using a joystick.Studies have shown the use of surface electromyography to steer powered wheelchairs or automobiles either through simulations or gaming controllers.However,these studies significantly lack issues with real world scenarios such as user’s safety,real-time control,and efficiency of the controller mechanism.The purpose of this study was to design,evaluate,and implement a hybrid human–machine interface system for a powered wheelchair that can detect human intent based on artificial neural network trained hand gesture recognition and navigate a powered wheelchair without colliding with objects around the path.Scaled Conjugate Gradient(SCG),Bayesian Regularization(BR),and Levenberg Marquart(LM)supervised artificial neural networks were trained in offline testing on eight participants without disability followed by online testing using the classifier with highest accuracy.Bayesian Regularization architecture showed highest accuracy at 98.4%across all participants and hidden layers.All participants successfully completed the path in an average of 5 min and 50 s,touching an average of 22.1%of the obstacles.The proposed hybrid system can be implemented to assist people with neuromuscular disabilities in near future.展开更多
利用Intelnet跨平台计算机视觉库OpenCV和微软Visual Studio 2010构建体感遥控器的软件开发平台。以OpenCV为基础,结合自定义算法和函数,开发利用人体手势识别来控制不同设备的体感遥控器。该设计采用基于codebook背景模型的运动检测方...利用Intelnet跨平台计算机视觉库OpenCV和微软Visual Studio 2010构建体感遥控器的软件开发平台。以OpenCV为基础,结合自定义算法和函数,开发利用人体手势识别来控制不同设备的体感遥控器。该设计采用基于codebook背景模型的运动检测方法对手势定位,以改进的CamShift算法和Kalman滤波器进行手势跟踪,进而对手势分割和识别。研究利用单目摄像头下对单手手势识别,以红外线和无线通信方式对目标区域和设备进行非接触控制。给出软件设计的基本思想,用OpenCV实现算法的具体过程和关键代码。该方案目标跟踪快、检测效率高、实时性好,为下一步在嵌入式系统中实施此方案打下一定基础。展开更多
This paper presents an anthropomorphic prosthetic hand using flexure hinges, which is controlled by the surface electromyography (sEMG) signals from 2 electrodes only. The prosthetic hand has compact structure with ...This paper presents an anthropomorphic prosthetic hand using flexure hinges, which is controlled by the surface electromyography (sEMG) signals from 2 electrodes only. The prosthetic hand has compact structure with 5 fingers and 4 Degree of Freedoms (DoFs) driven by 4 independent actuators. Helical springs are used as elastic joints and the joints of each finger are coupled by tendons. The myoelectric control system which can classify 8 prehensile hand gestures is built. Pattern recognition is employed where Mean Absolute Value (MAV), Variance (VAR), the fourth-order Autoregressive (AR) coefficient and Sample Entropy (SE) are chosen as the optimal feature set and Linear Discriminant Analysis (LDA) is utilized to reduce the dimension. A decision of hand gestures is generated by LDA classifier after the current projected feature set and the previous one are "pre-smoothed", and then the final decision is obtained when the current decision and previous decisions are "post-smoothed" from the decisions flow. The prosthetic hand can perform prehensile postures for activities of daily living and carry objects under the control of EMG signals.展开更多
文摘The use of hand gestures can be the most intuitive human-machine interaction medium.The early approaches for hand gesture recognition used device-based methods.These methods use mechanical or optical sensors attached to a glove or markers,which hinder the natural human-machine communication.On the other hand,vision-based methods are less restrictive and allow for a more spontaneous communication without the need of an intermediary between human and machine.Therefore,vision gesture recognition has been a popular area of research for the past thirty years.Hand gesture recognition finds its application in many areas,particularly the automotive industry where advanced automotive human-machine interface(HMI)designers are using gesture recognition to improve driver and vehicle safety.However,technology advances go beyond active/passive safety and into convenience and comfort.In this context,one of America’s big three automakers has partnered with the Centre of Pattern Analysis and Machine Intelligence(CPAMI)at the University of Waterloo to investigate expanding their product segment through machine learning to provide an increased driver convenience and comfort with the particular application of hand gesture recognition for autonomous car parking.The present paper leverages the state-of-the-art deep learning and optimization techniques to develop a vision-based multiview dynamic hand gesture recognizer for a self-parking system.We propose a 3D-CNN gesture model architecture that we train on a publicly available hand gesture database.We apply transfer learning methods to fine-tune the pre-trained gesture model on custom-made data,which significantly improves the proposed system performance in a real world environment.We adapt the architecture of end-to-end solution to expand the state-of-the-art video classifier from a single image as input(fed by monocular camera)to a Multiview 360 feed,offered by a six cameras module.Finally,we optimize the proposed solution to work on a limited resource embedded platform(Nvidia Jetson TX2)that is used by automakers for vehicle-based features,without sacrificing the accuracy robustness and real time functionality of the system.
基金supported by a grant (2021R1F1A1063634)of the Basic Science Research Program through the National Research Foundation (NRF)funded by the Ministry of Education,Republic of Korea.
文摘Hand gesture recognition (HGR) is used in a numerous applications,including medical health-care, industrial purpose and sports detection.We have developed a real-time hand gesture recognition system using inertialsensors for the smart home application. Developing such a model facilitatesthe medical health field (elders or disabled ones). Home automation has alsobeen proven to be a tremendous benefit for the elderly and disabled. Residentsare admitted to smart homes for comfort, luxury, improved quality of life,and protection against intrusion and burglars. This paper proposes a novelsystem that uses principal component analysis, linear discrimination analysisfeature extraction, and random forest as a classifier to improveHGRaccuracy.We have achieved an accuracy of 94% over the publicly benchmarked HGRdataset. The proposed system can be used to detect hand gestures in thehealthcare industry as well as in the industrial and educational sectors.
文摘Dynamic hand gesture recognition is a desired alternative means for human-computer interactions.This paper presents a hand gesture recognition system that is designed for the control of flights of unmanned aerial vehicles(UAV).A data representation model that represents a dynamic gesture sequence by converting the 4-D spatiotemporal data to 2-D matrix and a 1-D array is introduced.To train the system to recognize designed gestures,skeleton data collected from a Leap Motion Controller are converted to two different data models.As many as 9124 samples of the training dataset,1938 samples of the testing dataset are created to train and test the proposed three deep learning neural networks,which are a 2-layer fully connected neural network,a 5-layer fully connected neural network and an 8-layer convolutional neural network.The static testing results show that the 2-layer fully connected neural network achieves an average accuracy of 96.7%on scaled datasets and 12.3%on non-scaled datasets.The 5-layer fully connected neural network achieves an average accuracy of 98.0%on scaled datasets and 89.1%on non-scaled datasets.The 8-layer convolutional neural network achieves an average accuracy of 89.6%on scaled datasets and 96.9%on non-scaled datasets.Testing on a drone-kit simulator and a real drone shows that this system is feasible for drone flight controls.
文摘Individuals with cerebral palsy and muscular dystrophy often lack fine motor control of their fingers which makes it difficult to control traditional powered wheelchairs using a joystick.Studies have shown the use of surface electromyography to steer powered wheelchairs or automobiles either through simulations or gaming controllers.However,these studies significantly lack issues with real world scenarios such as user’s safety,real-time control,and efficiency of the controller mechanism.The purpose of this study was to design,evaluate,and implement a hybrid human–machine interface system for a powered wheelchair that can detect human intent based on artificial neural network trained hand gesture recognition and navigate a powered wheelchair without colliding with objects around the path.Scaled Conjugate Gradient(SCG),Bayesian Regularization(BR),and Levenberg Marquart(LM)supervised artificial neural networks were trained in offline testing on eight participants without disability followed by online testing using the classifier with highest accuracy.Bayesian Regularization architecture showed highest accuracy at 98.4%across all participants and hidden layers.All participants successfully completed the path in an average of 5 min and 50 s,touching an average of 22.1%of the obstacles.The proposed hybrid system can be implemented to assist people with neuromuscular disabilities in near future.
文摘利用Intelnet跨平台计算机视觉库OpenCV和微软Visual Studio 2010构建体感遥控器的软件开发平台。以OpenCV为基础,结合自定义算法和函数,开发利用人体手势识别来控制不同设备的体感遥控器。该设计采用基于codebook背景模型的运动检测方法对手势定位,以改进的CamShift算法和Kalman滤波器进行手势跟踪,进而对手势分割和识别。研究利用单目摄像头下对单手手势识别,以红外线和无线通信方式对目标区域和设备进行非接触控制。给出软件设计的基本思想,用OpenCV实现算法的具体过程和关键代码。该方案目标跟踪快、检测效率高、实时性好,为下一步在嵌入式系统中实施此方案打下一定基础。
基金This work is supported by National Natural Science Foundation of China (Grant Nos. 51575187 and 91223201), Science and Technology Program of Guangzhou (Grant No. 2014Y2-00217), Science and Technology Major Project of Huangpu District of Guang-Zhou (Grant No, 20150000661), the Fundamental Research Funds for the Central University (Grant No. 2015ZZ007) and Natural Science Foundation of Guangdong Province (Grant No. S2013030013355).
文摘This paper presents an anthropomorphic prosthetic hand using flexure hinges, which is controlled by the surface electromyography (sEMG) signals from 2 electrodes only. The prosthetic hand has compact structure with 5 fingers and 4 Degree of Freedoms (DoFs) driven by 4 independent actuators. Helical springs are used as elastic joints and the joints of each finger are coupled by tendons. The myoelectric control system which can classify 8 prehensile hand gestures is built. Pattern recognition is employed where Mean Absolute Value (MAV), Variance (VAR), the fourth-order Autoregressive (AR) coefficient and Sample Entropy (SE) are chosen as the optimal feature set and Linear Discriminant Analysis (LDA) is utilized to reduce the dimension. A decision of hand gestures is generated by LDA classifier after the current projected feature set and the previous one are "pre-smoothed", and then the final decision is obtained when the current decision and previous decisions are "post-smoothed" from the decisions flow. The prosthetic hand can perform prehensile postures for activities of daily living and carry objects under the control of EMG signals.