Artificial intelligence(AI)technology has become integral in the realm of medicine and healthcare,particularly in human activity recognition(HAR)applications such as fitness and rehabilitation tracking.This study intr...Artificial intelligence(AI)technology has become integral in the realm of medicine and healthcare,particularly in human activity recognition(HAR)applications such as fitness and rehabilitation tracking.This study introduces a robust coupling analysis framework that integrates four AI-enabled models,combining both machine learning(ML)and deep learning(DL)approaches to evaluate their effectiveness in HAR.The analytical dataset comprises 561 features sourced from the UCI-HAR database,forming the foundation for training the models.Additionally,the MHEALTH database is employed to replicate the modeling process for comparative purposes,while inclusion of the WISDM database,renowned for its challenging features,supports the framework’s resilience and adaptability.The ML-based models employ the methodologies including adaptive neuro-fuzzy inference system(ANFIS),support vector machine(SVM),and random forest(RF),for data training.In contrast,a DL-based model utilizes one-dimensional convolution neural network(1dCNN)to automate feature extraction.Furthermore,the recursive feature elimination(RFE)algorithm,which drives an ML-based estimator to eliminate low-participation features,helps identify the optimal features for enhancing model performance.The best accuracies of the ANFIS,SVM,RF,and 1dCNN models with meticulous featuring process achieve around 90%,96%,91%,and 93%,respectively.Comparative analysis using the MHEALTH dataset showcases the 1dCNN model’s remarkable perfect accuracy(100%),while the RF,SVM,and ANFIS models equipped with selected features achieve accuracies of 99.8%,99.7%,and 96.5%,respectively.Finally,when applied to the WISDM dataset,the DL-based and ML-based models attain accuracies of 91.4%and 87.3%,respectively,aligning with prior research findings.In conclusion,the proposed framework yields HAR models with commendable performance metrics,exhibiting its suitability for integration into the healthcare services system through AI-driven applications.展开更多
Virtual human is the simulation of human under the synthesis of virtual reality,artificial intelligence,and other technologies.Modern virtual human technology simulates both the external characteristics and the intern...Virtual human is the simulation of human under the synthesis of virtual reality,artificial intelligence,and other technologies.Modern virtual human technology simulates both the external characteristics and the internal emotions and personality of humans.The relationship between virtual human and human is a concrete expression of the modern symbiotic relationship between human and machine.This human-machine symbiosis can either be a fusion of the virtual human and the human or it can cause a split in the human itself.展开更多
Prompt radiation emitted during accelerator operation poses a significant health risk,necessitating a thorough search and securing of hazardous areas prior to initiation.Currently,manual sweep methods are employed.How...Prompt radiation emitted during accelerator operation poses a significant health risk,necessitating a thorough search and securing of hazardous areas prior to initiation.Currently,manual sweep methods are employed.However,the limitations of manual sweeps have become increasingly evident with the implementation of large-scale accelerators.By leveraging advancements in machine vision technology,the automatic identification of stranded personnel in controlled areas through camera imagery presents a viable solution for efficient search and security.Given the criticality of personal safety for stranded individuals,search and security processes must be sufficiently reliable.To ensure comprehensive coverage,180°camera groups were strategically positioned on both sides of the accelerator tunnel to eliminate blind spots within the monitoring range.The YOLOV8 network model was modified to enable the detection of small targets,such as hands and feet,as well as larger targets formed by individuals near the cameras.Furthermore,the system incorporates a pedestrian recognition model that detects human body parts,and an information fusion strategy is used to integrate the detected head,hands,and feet with the identified pedestrians as a cohesive unit.This strategy enhanced the capability of the model to identify pedestrians obstructed by equipment,resulting in a notable improvement in the recall rate.Specifically,recall rates of 0.915 and 0.82were obtained for Datasets 1 and 2,respectively.Although there was a slight decrease in accuracy,it aligned with the intended purpose of the search-and-secure software design.Experimental tests conducted within an accelerator tunnel demonstrated the effectiveness of this approach in achieving reliable recognition outcomes.展开更多
Efficient and flexible interactions require precisely converting human intentions into computer-recognizable signals,which is critical to the breakthrough development of metaverse.Interactive electronics face common d...Efficient and flexible interactions require precisely converting human intentions into computer-recognizable signals,which is critical to the breakthrough development of metaverse.Interactive electronics face common dilemmas,which realize highprecision and stable touch detection but are rigid,bulky,and thick or achieve high flexibility to wear but lose precision.Here,we construct highly bending-insensitive,unpixelated,and waterproof epidermal interfaces(BUW epidermal interfaces)and demonstrate their interactive applications of conformal human–machine integration.The BUW epidermal interface based on the addressable electrical contact structure exhibits high-precision and stable touch detection,high flexibility,rapid response time,excellent stability,and versatile“cut-and-paste”character.Regardless of whether being flat or bent,the BUW epidermal interface can be conformally attached to the human skin for real-time,comfortable,and unrestrained interactions.This research provides promising insight into the functional composite and structural design strategies for developing epidermal electronics,which offers a new technology route and may further broaden human–machine interactions toward metaverse.展开更多
The development of artificial intelligence(AI)and smart home technologies has driven the need for speech recognition-based solutions.This demand stems from the quest for more intuitive and natural interaction between ...The development of artificial intelligence(AI)and smart home technologies has driven the need for speech recognition-based solutions.This demand stems from the quest for more intuitive and natural interaction between users and smart devices in their homes.Speech recognition allows users to control devices and perform everyday actions through spoken commands,eliminating the need for physical interfaces or touch screens and enabling specific tasks such as turning on or off the light,heating,or lowering the blinds.The purpose of this study is to develop a speech-based classification model for recognizing human actions in the smart home.It seeks to demonstrate the effectiveness and feasibility of using machine learning techniques in predicting categories,subcategories,and actions from sentences.A dataset labeled with relevant information about categories,subcategories,and actions related to human actions in the smart home is used.The methodology uses machine learning techniques implemented in Python,extracting features using CountVectorizer to convert sentences into numerical representations.The results show that the classification model is able to accurately predict categories,subcategories,and actions based on sentences,with 82.99%accuracy for category,76.19%accuracy for subcategory,and 90.28%accuracy for action.The study concludes that using machine learning techniques is effective for recognizing and classifying human actions in the smart home,supporting its feasibility in various scenarios and opening new possibilities for advanced natural language processing systems in the field of AI and smart homes.展开更多
The inter-class face classification problem is more reasonable than the intra-class classification problem.To address this issue,we have carried out empirical research on classifying Indian people to their geographica...The inter-class face classification problem is more reasonable than the intra-class classification problem.To address this issue,we have carried out empirical research on classifying Indian people to their geographical regions.This work aimed to construct a computational classification model for classifying Indian regional face images acquired from south and east regions of India,referring to human vision.We have created an Automated Human Intelligence System(AHIS)to evaluate human visual capabilities.Analysis of AHIS response showed that face shape is a discriminative feature among the other facial features.We have developed a modified convolutional neural network to characterize the human vision response to improve face classification accuracy.The proposed model achieved mean F1 and Matthew Correlation Coefficient(MCC)of 0.92 and 0.84,respectively,on the validation set,outperforming the traditional Convolutional Neural Network(CNN).The CNN-Contoured Face(CNN-FC)model is developed to train contoured face images to investigate the influence of face shape.Finally,to cross-validate the accuracy of these models,the traditional CNN model is trained on the same dataset.With an accuracy of 92.98%,the Modified-CNN(M-CNN)model has demonstrated that the proposed method could facilitate the tangible impact in intra-classification problems.A novel Indian regional face dataset is created for supporting this supervised classification work,and it will be available to the research community.展开更多
The aim of this study is to improve the efficiency of external corrosion inspection of pipes in chemical plants.Currently,the preferred method involves manual inspection of images of corroded pipes;however,this places...The aim of this study is to improve the efficiency of external corrosion inspection of pipes in chemical plants.Currently,the preferred method involves manual inspection of images of corroded pipes;however,this places significant workload on human experts owing to the large number of required images.Furthermore,visual assessment of corrosion levels is prone to subjective errors.To address these issues,we developed an AI(artificial intelligence)-based corrosion-diagnosis system(AI corrosion-diagnosis system)and implemented it in a factory.The proposed system architecture was based on HITL(human-in-the-loop)ML(machine learning)[1].To overcome the difficulty of developing a highly accurate ML model during the PoC(proof-of-concept)stage,the system relies on cooperation between humans and the ML model,utilizing human expertise during operation.For instance,if the accuracy of the ML model was initially 60%during the development stage,a cooperative approach would be adopted during the operational stage,with humans supplementing the remaining 40%accuracy.The implemented system’s ML model achieved a recall rate of approximately 70%.The system’s implementation not only contributed to the efficiency of operations by supporting diagnosis through the ML model but also facilitated the transition to systematic data management,resulting in an overall workload reduction of approximately 50%.The operation based on HITL was demonstrated to be a crucial element for achieving efficient system operation through the collaboration of humans and ML models,even when the initial accuracy of the ML model was low.Future efforts will focus on improving the detection of corrosion at elevated locations by considering using video cameras to capture pipe images.The goal is to reduce the workload for inspectors and enhance the quality of inspections by identifying corrosion locations using ML models.展开更多
In the field of robotics and in the health sciences, transitions have been occurring in the control of robots operating with predetermined logic and rules. Robotics in health care are influencing human caring dynamics...In the field of robotics and in the health sciences, transitions have been occurring in the control of robots operating with predetermined logic and rules. Robotics in health care are influencing human caring dynamics in many ways such as enhancing dependency and surrender to machine technologies. Situations such as these are charged with possibilities of legal liabilities triggered by influences and consequences of advancing robotic technology dependency. The purpose of this paper is to identify, describe, and explain legal issues and/or dilemmas centered on robotics in healthcare while providing engaging opportunities to limit consequent legalities thus forming beneficial human health care outcomes. Laying bare these liabilities will provide critically informative data that can foster proactive encounters which can or may deter health care liabilities while ensuring quality healthcare outcomes. An attempt is made to re-conceptualize how to view agency, causality, liability responsibility, culpability, and autonomy for the new age of autonomous robots. While it is still not clear how this would turn out, a clear framing of the problem is the first step in the project.展开更多
Classification of human actions under video surveillance is gaining a lot of attention from computer vision researchers.In this paper,we have presented methodology to recognize human behavior in thin crowd which may b...Classification of human actions under video surveillance is gaining a lot of attention from computer vision researchers.In this paper,we have presented methodology to recognize human behavior in thin crowd which may be very helpful in surveillance.Research have mostly focused the problem of human detection in thin crowd,overall behavior of the crowd and actions of individuals in video sequences.Vision based Human behavior modeling is a complex task as it involves human detection,tracking,classifying normal and abnormal behavior.The proposed methodology takes input video and applies Gaussian based segmentation technique followed by post processing through presenting hole filling algorithm i.e.,fill hole inside objects algorithm.Human detection is performed by presenting human detection algorithm and then geometrical features from human skeleton are extracted using feature extraction algorithm.The classification task is achieved using binary and multi class support vector machines.The proposed technique is validated through accuracy,precision,recall and F-measure metrics.展开更多
The number and variety of applications of artificial intelligence(AI)in gastr-ointestinal(GI)endoscopy is growing rapidly.New technologies based on machine learning(ML)and convolutional neural networks(CNNs)are at var...The number and variety of applications of artificial intelligence(AI)in gastr-ointestinal(GI)endoscopy is growing rapidly.New technologies based on machine learning(ML)and convolutional neural networks(CNNs)are at various stages of development and deployment to assist patients and endoscopists in preparing for endoscopic procedures,in detection,diagnosis and classification of pathology during endoscopy and in confirmation of key performance indicators.Platforms based on ML and CNNs require regulatory approval as medical devices.Interactions between humans and the technologies we use are complex and are influenced by design,behavioural and psychological elements.Due to the substantial differences between AI and prior technologies,important differences may be expected in how we interact with advice from AI technologies.Human-AI interaction(HAII)may be optimised by developing AI algorithms to minimise false positives and designing platform interfaces to maximise usability.Human factors influencing HAII may include automation bias,alarm fatigue,algorithm aversion,learning effect and deskilling.Each of these areas merits further study in the specific setting of AI applications in GI endoscopy and professional societies should engage to ensure that sufficient emphasis is placed on human-centred design in development of new AI technologies.展开更多
Speech recognition rate will deteriorate greatly in human-machine interaction when the speaker's speech mixes with a bystander's voice. This paper proposes a time-frequency approach for Blind Source Seperation...Speech recognition rate will deteriorate greatly in human-machine interaction when the speaker's speech mixes with a bystander's voice. This paper proposes a time-frequency approach for Blind Source Seperation (BSS) for intelligent Human-Machine Interaction(HMI). Main idea of the algorithm is to simultaneously diagonalize the correlation matrix of the pre-whitened signals at different time delays for every frequency bins in time-frequency domain. The prososed method has two merits: (1) fast convergence speed; (2) high signal to interference ratio of the separated signals. Numerical evaluations are used to compare the performance of the proposed algorithm with two other deconvolution algorithms. An efficient algorithm to resolve permutation ambiguity is also proposed in this paper. The algorithm proposed saves more than 10% of computational time with properly selected parameters and achieves good performances for both simulated convolutive mixtures and real room recorded speeches.展开更多
The inculcation of bioinspiration in sensing and human–machine interface(HMI)technologies can lead to distinctive characteristics such as conformability,low power consumption,high sensitivity,and unique properties li...The inculcation of bioinspiration in sensing and human–machine interface(HMI)technologies can lead to distinctive characteristics such as conformability,low power consumption,high sensitivity,and unique properties like self-healing,self-cleaning,and adaptability.Both sensing and HMI are fields rife with opportunities for the application of bioinspired nanomaterials,particularly when it comes to wearable sensory systems where biocompatibility is an additional requirement.This review discusses recent development in bioinspired nanomaterials for wearable sensing and HMIs,with a specific focus on state-of-the-art bioinspired capacitive sensors,piezoresistive sensors,piezoelectric sensors,triboelectric sensors,magnetoelastic sensors,and electrochemical sensors.We also present a comprehensive overview of the challenges that have hindered the scientific advancement in academia and commercialization in the industry.展开更多
Automatic speaker recognition(ASR)systems are the field of Human-machine interaction and scientists have been using feature extraction and feature matching methods to analyze and synthesize these signals.One of the mo...Automatic speaker recognition(ASR)systems are the field of Human-machine interaction and scientists have been using feature extraction and feature matching methods to analyze and synthesize these signals.One of the most commonly used methods for feature extraction is Mel Frequency Cepstral Coefficients(MFCCs).Recent researches show that MFCCs are successful in processing the voice signal with high accuracies.MFCCs represents a sequence of voice signal-specific features.This experimental analysis is proposed to distinguish Turkish speakers by extracting the MFCCs from the speech recordings.Since the human perception of sound is not linear,after the filterbank step in theMFCC method,we converted the obtained log filterbanks into decibel(dB)features-based spectrograms without applying the Discrete Cosine Transform(DCT).A new dataset was created with converted spectrogram into a 2-D array.Several learning algorithms were implementedwith a 10-fold cross-validationmethod to detect the speaker.The highest accuracy of 90.2%was achieved using Multi-layer Perceptron(MLP)with tanh activation function.The most important output of this study is the inclusion of human voice as a new feature set.展开更多
In recent years,Brain-Computer Interface(BCI)system gained much popularity since it aims at establishing the communication between human brain and computer.BCI systems are applied in several research areas such as neu...In recent years,Brain-Computer Interface(BCI)system gained much popularity since it aims at establishing the communication between human brain and computer.BCI systems are applied in several research areas such as neuro-rehabilitation,robots,exoeskeletons,etc.Electroencephalography(EEG)is a technique commonly applied in capturing brain signals.It is incorporated in BCI systems since it has attractive features such as noninvasive nature,high time-resolution output,mobility and cost-effective.EEG classification process is highly essential in decision making process and it incorporates different processes namely,feature extraction,feature selection,and classification.With this motivation,the current research paper presents an Intelligent Optimal Fuzzy Support Vector Machine-based EEC recognition(IOFSVM-EEG)model for BCI system.Independent Component Analysis(ICA)technique is applied onto the proposed IOFSVM-EEG model to remove the artefacts that exist in EEG signal and to retain the meaningful EEG information.Besides,Common Spatial Pattern(CSP)-based feature extraction technique is utilized to derive a helpful set of feature vectors from the preprocessed EEG signals.Moreover,OFSVM method is applied in the classification of EEG signals,in which the parameters involved in FSVM are optimally tuned using Grasshopper Optimization Algorithm(GOA).In order to validate the enhanced EEG recognition outcomes of the proposed IOFSVM-EEG model,an extensive set of experiments was conducted.The outcomes were examined under distinct aspects.The experimental results highlighted the enhanced performance of the presented IOFSVM-EEG model over other state-of-the-art methods.展开更多
The world’s elderly population is growing every year.It is easy to say that the fall is one of the major dangers that threaten them.This paper offers a Trained Model for fall detection to help the older people live c...The world’s elderly population is growing every year.It is easy to say that the fall is one of the major dangers that threaten them.This paper offers a Trained Model for fall detection to help the older people live comfortably and alone at home.The purpose of this paper is to investigate appropriate methods for diagnosing falls by analyzing the motion and shape characteristics of the human body.Several machine learning technologies have been proposed for automatic fall detection.The proposed research reported in this paper detects a moving object by using a background subtraction algorithm with a single camera.The next step is to extract the features that are very important and generally describe the human shape and show the difference between the human falls from the daily activities.These features are based on motion,changes in human shape,and oval diameters around the human and temporal head position.The features extracted from the human mask are eventually fed in to various machine learning classifiers for fall detection.Experimental results showed the efficiency and reliability of the proposed method with a fall detection rate of 81%that have been tested with UR Fall Detection dataset.展开更多
Traditional indoor human activity recognition(HAR)is a timeseries data classification problem and needs feature extraction.Presently,considerable attention has been given to the domain ofHARdue to the enormous amount ...Traditional indoor human activity recognition(HAR)is a timeseries data classification problem and needs feature extraction.Presently,considerable attention has been given to the domain ofHARdue to the enormous amount of its real-time uses in real-time applications,namely surveillance by authorities,biometric user identification,and health monitoring of older people.The extensive usage of the Internet of Things(IoT)and wearable sensor devices has made the topic of HAR a vital subject in ubiquitous and mobile computing.The more commonly utilized inference and problemsolving technique in the HAR system have recently been deep learning(DL).The study develops aModifiedWild Horse Optimization withDLAided Symmetric Human Activity Recognition(MWHODL-SHAR)model.The major intention of the MWHODL-SHAR model lies in recognition of symmetric activities,namely jogging,walking,standing,sitting,etc.In the presented MWHODL-SHAR technique,the human activities data is pre-processed in various stages to make it compatible for further processing.A convolution neural network with an attention-based long short-term memory(CNNALSTM)model is applied for activity recognition.The MWHO algorithm is utilized as a hyperparameter tuning strategy to improve the detection rate of the CNN-ALSTM algorithm.The experimental validation of the MWHODL-SHAR technique is simulated using a benchmark dataset.An extensive comparison study revealed the betterment of theMWHODL-SHAR technique over other recent approaches.展开更多
Human recognition technology based on biometrics has become a fundamental requirement in all aspects of life due to increased concerns about security and privacy issues.Therefore,biometric systems have emerged as a te...Human recognition technology based on biometrics has become a fundamental requirement in all aspects of life due to increased concerns about security and privacy issues.Therefore,biometric systems have emerged as a technology with the capability to identify or authenticate individuals based on their physiological and behavioral characteristics.Among different viable biometric modalities,the human ear structure can offer unique and valuable discriminative characteristics for human recognition systems.In recent years,most existing traditional ear recognition systems have been designed based on computer vision models and have achieved successful results.Nevertheless,such traditional models can be sensitive to several unconstrained environmental factors.As such,some traits may be difficult to extract automatically but can still be semantically perceived as soft biometrics.This research proposes a new group of semantic features to be used as soft ear biometrics,mainly inspired by conventional descriptive traits used naturally by humans when identifying or describing each other.Hence,the research study is focused on the fusion of the soft ear biometric traits with traditional(hard)ear biometric features to investigate their validity and efficacy in augmenting human identification performance.The proposed framework has two subsystems:first,a computer vision-based subsystem,extracting traditional(hard)ear biometric traits using principal component analysis(PCA)and local binary patterns(LBP),and second,a crowdsourcing-based subsystem,deriving semantic(soft)ear biometric traits.Several feature-level fusion experiments were conducted using the AMI database to evaluate the proposed algorithm’s performance.The obtained results for both identification and verification showed that the proposed soft ear biometric information significantly improved the recognition performance of traditional ear biometrics,reaching up to 12%for LBP and 5%for PCA descriptors;when fusing all three capacities PCA,LBP,and soft traits using k-nearest neighbors(KNN)classifier.展开更多
基金funded by the National Science and Technology Council,Taiwan(Grant No.NSTC 112-2121-M-039-001)by China Medical University(Grant No.CMU112-MF-79).
文摘Artificial intelligence(AI)technology has become integral in the realm of medicine and healthcare,particularly in human activity recognition(HAR)applications such as fitness and rehabilitation tracking.This study introduces a robust coupling analysis framework that integrates four AI-enabled models,combining both machine learning(ML)and deep learning(DL)approaches to evaluate their effectiveness in HAR.The analytical dataset comprises 561 features sourced from the UCI-HAR database,forming the foundation for training the models.Additionally,the MHEALTH database is employed to replicate the modeling process for comparative purposes,while inclusion of the WISDM database,renowned for its challenging features,supports the framework’s resilience and adaptability.The ML-based models employ the methodologies including adaptive neuro-fuzzy inference system(ANFIS),support vector machine(SVM),and random forest(RF),for data training.In contrast,a DL-based model utilizes one-dimensional convolution neural network(1dCNN)to automate feature extraction.Furthermore,the recursive feature elimination(RFE)algorithm,which drives an ML-based estimator to eliminate low-participation features,helps identify the optimal features for enhancing model performance.The best accuracies of the ANFIS,SVM,RF,and 1dCNN models with meticulous featuring process achieve around 90%,96%,91%,and 93%,respectively.Comparative analysis using the MHEALTH dataset showcases the 1dCNN model’s remarkable perfect accuracy(100%),while the RF,SVM,and ANFIS models equipped with selected features achieve accuracies of 99.8%,99.7%,and 96.5%,respectively.Finally,when applied to the WISDM dataset,the DL-based and ML-based models attain accuracies of 91.4%and 87.3%,respectively,aligning with prior research findings.In conclusion,the proposed framework yields HAR models with commendable performance metrics,exhibiting its suitability for integration into the healthcare services system through AI-driven applications.
文摘Virtual human is the simulation of human under the synthesis of virtual reality,artificial intelligence,and other technologies.Modern virtual human technology simulates both the external characteristics and the internal emotions and personality of humans.The relationship between virtual human and human is a concrete expression of the modern symbiotic relationship between human and machine.This human-machine symbiosis can either be a fusion of the virtual human and the human or it can cause a split in the human itself.
文摘Prompt radiation emitted during accelerator operation poses a significant health risk,necessitating a thorough search and securing of hazardous areas prior to initiation.Currently,manual sweep methods are employed.However,the limitations of manual sweeps have become increasingly evident with the implementation of large-scale accelerators.By leveraging advancements in machine vision technology,the automatic identification of stranded personnel in controlled areas through camera imagery presents a viable solution for efficient search and security.Given the criticality of personal safety for stranded individuals,search and security processes must be sufficiently reliable.To ensure comprehensive coverage,180°camera groups were strategically positioned on both sides of the accelerator tunnel to eliminate blind spots within the monitoring range.The YOLOV8 network model was modified to enable the detection of small targets,such as hands and feet,as well as larger targets formed by individuals near the cameras.Furthermore,the system incorporates a pedestrian recognition model that detects human body parts,and an information fusion strategy is used to integrate the detected head,hands,and feet with the identified pedestrians as a cohesive unit.This strategy enhanced the capability of the model to identify pedestrians obstructed by equipment,resulting in a notable improvement in the recall rate.Specifically,recall rates of 0.915 and 0.82were obtained for Datasets 1 and 2,respectively.Although there was a slight decrease in accuracy,it aligned with the intended purpose of the search-and-secure software design.Experimental tests conducted within an accelerator tunnel demonstrated the effectiveness of this approach in achieving reliable recognition outcomes.
基金supported by National Natural Science Foundation of China(52202117,52232006,52072029,and 12102256)Collaborative Innovation Platform Project of Fu-Xia-Quan National Independent Innovation Demonstration Zone(3502ZCQXT2022005)+3 种基金Natural Science Foundation of Fujian Province of China(2022J01065)State Key Lab of Advanced Metals and Materials(2022-Z09)Fundamental Research Funds for the Central Universities(20720220075)the Ministry of Education,Singapore,under its MOE ARF Tier 2(MOE2019-T2-2-179).
文摘Efficient and flexible interactions require precisely converting human intentions into computer-recognizable signals,which is critical to the breakthrough development of metaverse.Interactive electronics face common dilemmas,which realize highprecision and stable touch detection but are rigid,bulky,and thick or achieve high flexibility to wear but lose precision.Here,we construct highly bending-insensitive,unpixelated,and waterproof epidermal interfaces(BUW epidermal interfaces)and demonstrate their interactive applications of conformal human–machine integration.The BUW epidermal interface based on the addressable electrical contact structure exhibits high-precision and stable touch detection,high flexibility,rapid response time,excellent stability,and versatile“cut-and-paste”character.Regardless of whether being flat or bent,the BUW epidermal interface can be conformally attached to the human skin for real-time,comfortable,and unrestrained interactions.This research provides promising insight into the functional composite and structural design strategies for developing epidermal electronics,which offers a new technology route and may further broaden human–machine interactions toward metaverse.
基金supported by Generalitat Valenciana with HAAS(CIAICO/2021/039)the Spanish Ministry of Science and Innovation under the Project AVANTIA PID2020-114480RB-I00.
文摘The development of artificial intelligence(AI)and smart home technologies has driven the need for speech recognition-based solutions.This demand stems from the quest for more intuitive and natural interaction between users and smart devices in their homes.Speech recognition allows users to control devices and perform everyday actions through spoken commands,eliminating the need for physical interfaces or touch screens and enabling specific tasks such as turning on or off the light,heating,or lowering the blinds.The purpose of this study is to develop a speech-based classification model for recognizing human actions in the smart home.It seeks to demonstrate the effectiveness and feasibility of using machine learning techniques in predicting categories,subcategories,and actions from sentences.A dataset labeled with relevant information about categories,subcategories,and actions related to human actions in the smart home is used.The methodology uses machine learning techniques implemented in Python,extracting features using CountVectorizer to convert sentences into numerical representations.The results show that the classification model is able to accurately predict categories,subcategories,and actions based on sentences,with 82.99%accuracy for category,76.19%accuracy for subcategory,and 90.28%accuracy for action.The study concludes that using machine learning techniques is effective for recognizing and classifying human actions in the smart home,supporting its feasibility in various scenarios and opening new possibilities for advanced natural language processing systems in the field of AI and smart homes.
文摘The inter-class face classification problem is more reasonable than the intra-class classification problem.To address this issue,we have carried out empirical research on classifying Indian people to their geographical regions.This work aimed to construct a computational classification model for classifying Indian regional face images acquired from south and east regions of India,referring to human vision.We have created an Automated Human Intelligence System(AHIS)to evaluate human visual capabilities.Analysis of AHIS response showed that face shape is a discriminative feature among the other facial features.We have developed a modified convolutional neural network to characterize the human vision response to improve face classification accuracy.The proposed model achieved mean F1 and Matthew Correlation Coefficient(MCC)of 0.92 and 0.84,respectively,on the validation set,outperforming the traditional Convolutional Neural Network(CNN).The CNN-Contoured Face(CNN-FC)model is developed to train contoured face images to investigate the influence of face shape.Finally,to cross-validate the accuracy of these models,the traditional CNN model is trained on the same dataset.With an accuracy of 92.98%,the Modified-CNN(M-CNN)model has demonstrated that the proposed method could facilitate the tangible impact in intra-classification problems.A novel Indian regional face dataset is created for supporting this supervised classification work,and it will be available to the research community.
文摘The aim of this study is to improve the efficiency of external corrosion inspection of pipes in chemical plants.Currently,the preferred method involves manual inspection of images of corroded pipes;however,this places significant workload on human experts owing to the large number of required images.Furthermore,visual assessment of corrosion levels is prone to subjective errors.To address these issues,we developed an AI(artificial intelligence)-based corrosion-diagnosis system(AI corrosion-diagnosis system)and implemented it in a factory.The proposed system architecture was based on HITL(human-in-the-loop)ML(machine learning)[1].To overcome the difficulty of developing a highly accurate ML model during the PoC(proof-of-concept)stage,the system relies on cooperation between humans and the ML model,utilizing human expertise during operation.For instance,if the accuracy of the ML model was initially 60%during the development stage,a cooperative approach would be adopted during the operational stage,with humans supplementing the remaining 40%accuracy.The implemented system’s ML model achieved a recall rate of approximately 70%.The system’s implementation not only contributed to the efficiency of operations by supporting diagnosis through the ML model but also facilitated the transition to systematic data management,resulting in an overall workload reduction of approximately 50%.The operation based on HITL was demonstrated to be a crucial element for achieving efficient system operation through the collaboration of humans and ML models,even when the initial accuracy of the ML model was low.Future efforts will focus on improving the detection of corrosion at elevated locations by considering using video cameras to capture pipe images.The goal is to reduce the workload for inspectors and enhance the quality of inspections by identifying corrosion locations using ML models.
文摘In the field of robotics and in the health sciences, transitions have been occurring in the control of robots operating with predetermined logic and rules. Robotics in health care are influencing human caring dynamics in many ways such as enhancing dependency and surrender to machine technologies. Situations such as these are charged with possibilities of legal liabilities triggered by influences and consequences of advancing robotic technology dependency. The purpose of this paper is to identify, describe, and explain legal issues and/or dilemmas centered on robotics in healthcare while providing engaging opportunities to limit consequent legalities thus forming beneficial human health care outcomes. Laying bare these liabilities will provide critically informative data that can foster proactive encounters which can or may deter health care liabilities while ensuring quality healthcare outcomes. An attempt is made to re-conceptualize how to view agency, causality, liability responsibility, culpability, and autonomy for the new age of autonomous robots. While it is still not clear how this would turn out, a clear framing of the problem is the first step in the project.
文摘Classification of human actions under video surveillance is gaining a lot of attention from computer vision researchers.In this paper,we have presented methodology to recognize human behavior in thin crowd which may be very helpful in surveillance.Research have mostly focused the problem of human detection in thin crowd,overall behavior of the crowd and actions of individuals in video sequences.Vision based Human behavior modeling is a complex task as it involves human detection,tracking,classifying normal and abnormal behavior.The proposed methodology takes input video and applies Gaussian based segmentation technique followed by post processing through presenting hole filling algorithm i.e.,fill hole inside objects algorithm.Human detection is performed by presenting human detection algorithm and then geometrical features from human skeleton are extracted using feature extraction algorithm.The classification task is achieved using binary and multi class support vector machines.The proposed technique is validated through accuracy,precision,recall and F-measure metrics.
文摘The number and variety of applications of artificial intelligence(AI)in gastr-ointestinal(GI)endoscopy is growing rapidly.New technologies based on machine learning(ML)and convolutional neural networks(CNNs)are at various stages of development and deployment to assist patients and endoscopists in preparing for endoscopic procedures,in detection,diagnosis and classification of pathology during endoscopy and in confirmation of key performance indicators.Platforms based on ML and CNNs require regulatory approval as medical devices.Interactions between humans and the technologies we use are complex and are influenced by design,behavioural and psychological elements.Due to the substantial differences between AI and prior technologies,important differences may be expected in how we interact with advice from AI technologies.Human-AI interaction(HAII)may be optimised by developing AI algorithms to minimise false positives and designing platform interfaces to maximise usability.Human factors influencing HAII may include automation bias,alarm fatigue,algorithm aversion,learning effect and deskilling.Each of these areas merits further study in the specific setting of AI applications in GI endoscopy and professional societies should engage to ensure that sufficient emphasis is placed on human-centred design in development of new AI technologies.
文摘Speech recognition rate will deteriorate greatly in human-machine interaction when the speaker's speech mixes with a bystander's voice. This paper proposes a time-frequency approach for Blind Source Seperation (BSS) for intelligent Human-Machine Interaction(HMI). Main idea of the algorithm is to simultaneously diagonalize the correlation matrix of the pre-whitened signals at different time delays for every frequency bins in time-frequency domain. The prososed method has two merits: (1) fast convergence speed; (2) high signal to interference ratio of the separated signals. Numerical evaluations are used to compare the performance of the proposed algorithm with two other deconvolution algorithms. An efficient algorithm to resolve permutation ambiguity is also proposed in this paper. The algorithm proposed saves more than 10% of computational time with properly selected parameters and achieves good performances for both simulated convolutive mixtures and real room recorded speeches.
基金support.J.C.also acknowledges the Hellman Fellows Research Grant,the UCLA Pandemic Resources Program Research Award,the Research Recovery Grant by the UCLA Academic Senate,and the Brain&Behavior Research Foundation Young Investigator Grant(Grant Number:30944)the Catalyzing Pediatric Innovation Grant(Grant Number:47744)from the West Coast Consortium for Technology&Innovation in Pediatrics,Children’s Hospital Los Angeles.
文摘The inculcation of bioinspiration in sensing and human–machine interface(HMI)technologies can lead to distinctive characteristics such as conformability,low power consumption,high sensitivity,and unique properties like self-healing,self-cleaning,and adaptability.Both sensing and HMI are fields rife with opportunities for the application of bioinspired nanomaterials,particularly when it comes to wearable sensory systems where biocompatibility is an additional requirement.This review discusses recent development in bioinspired nanomaterials for wearable sensing and HMIs,with a specific focus on state-of-the-art bioinspired capacitive sensors,piezoresistive sensors,piezoelectric sensors,triboelectric sensors,magnetoelastic sensors,and electrochemical sensors.We also present a comprehensive overview of the challenges that have hindered the scientific advancement in academia and commercialization in the industry.
基金This work was supported by the GRRC program of Gyeonggi province.[GRRC-Gachon2020(B04),Development of AI-based Healthcare Devices].
文摘Automatic speaker recognition(ASR)systems are the field of Human-machine interaction and scientists have been using feature extraction and feature matching methods to analyze and synthesize these signals.One of the most commonly used methods for feature extraction is Mel Frequency Cepstral Coefficients(MFCCs).Recent researches show that MFCCs are successful in processing the voice signal with high accuracies.MFCCs represents a sequence of voice signal-specific features.This experimental analysis is proposed to distinguish Turkish speakers by extracting the MFCCs from the speech recordings.Since the human perception of sound is not linear,after the filterbank step in theMFCC method,we converted the obtained log filterbanks into decibel(dB)features-based spectrograms without applying the Discrete Cosine Transform(DCT).A new dataset was created with converted spectrogram into a 2-D array.Several learning algorithms were implementedwith a 10-fold cross-validationmethod to detect the speaker.The highest accuracy of 90.2%was achieved using Multi-layer Perceptron(MLP)with tanh activation function.The most important output of this study is the inclusion of human voice as a new feature set.
文摘In recent years,Brain-Computer Interface(BCI)system gained much popularity since it aims at establishing the communication between human brain and computer.BCI systems are applied in several research areas such as neuro-rehabilitation,robots,exoeskeletons,etc.Electroencephalography(EEG)is a technique commonly applied in capturing brain signals.It is incorporated in BCI systems since it has attractive features such as noninvasive nature,high time-resolution output,mobility and cost-effective.EEG classification process is highly essential in decision making process and it incorporates different processes namely,feature extraction,feature selection,and classification.With this motivation,the current research paper presents an Intelligent Optimal Fuzzy Support Vector Machine-based EEC recognition(IOFSVM-EEG)model for BCI system.Independent Component Analysis(ICA)technique is applied onto the proposed IOFSVM-EEG model to remove the artefacts that exist in EEG signal and to retain the meaningful EEG information.Besides,Common Spatial Pattern(CSP)-based feature extraction technique is utilized to derive a helpful set of feature vectors from the preprocessed EEG signals.Moreover,OFSVM method is applied in the classification of EEG signals,in which the parameters involved in FSVM are optimally tuned using Grasshopper Optimization Algorithm(GOA).In order to validate the enhanced EEG recognition outcomes of the proposed IOFSVM-EEG model,an extensive set of experiments was conducted.The outcomes were examined under distinct aspects.The experimental results highlighted the enhanced performance of the presented IOFSVM-EEG model over other state-of-the-art methods.
文摘The world’s elderly population is growing every year.It is easy to say that the fall is one of the major dangers that threaten them.This paper offers a Trained Model for fall detection to help the older people live comfortably and alone at home.The purpose of this paper is to investigate appropriate methods for diagnosing falls by analyzing the motion and shape characteristics of the human body.Several machine learning technologies have been proposed for automatic fall detection.The proposed research reported in this paper detects a moving object by using a background subtraction algorithm with a single camera.The next step is to extract the features that are very important and generally describe the human shape and show the difference between the human falls from the daily activities.These features are based on motion,changes in human shape,and oval diameters around the human and temporal head position.The features extracted from the human mask are eventually fed in to various machine learning classifiers for fall detection.Experimental results showed the efficiency and reliability of the proposed method with a fall detection rate of 81%that have been tested with UR Fall Detection dataset.
文摘Traditional indoor human activity recognition(HAR)is a timeseries data classification problem and needs feature extraction.Presently,considerable attention has been given to the domain ofHARdue to the enormous amount of its real-time uses in real-time applications,namely surveillance by authorities,biometric user identification,and health monitoring of older people.The extensive usage of the Internet of Things(IoT)and wearable sensor devices has made the topic of HAR a vital subject in ubiquitous and mobile computing.The more commonly utilized inference and problemsolving technique in the HAR system have recently been deep learning(DL).The study develops aModifiedWild Horse Optimization withDLAided Symmetric Human Activity Recognition(MWHODL-SHAR)model.The major intention of the MWHODL-SHAR model lies in recognition of symmetric activities,namely jogging,walking,standing,sitting,etc.In the presented MWHODL-SHAR technique,the human activities data is pre-processed in various stages to make it compatible for further processing.A convolution neural network with an attention-based long short-term memory(CNNALSTM)model is applied for activity recognition.The MWHO algorithm is utilized as a hyperparameter tuning strategy to improve the detection rate of the CNN-ALSTM algorithm.The experimental validation of the MWHODL-SHAR technique is simulated using a benchmark dataset.An extensive comparison study revealed the betterment of theMWHODL-SHAR technique over other recent approaches.
基金supported and funded by KAU Scientific Endowment,King Abdulaziz University,Jeddah,Saudi Arabia.
文摘Human recognition technology based on biometrics has become a fundamental requirement in all aspects of life due to increased concerns about security and privacy issues.Therefore,biometric systems have emerged as a technology with the capability to identify or authenticate individuals based on their physiological and behavioral characteristics.Among different viable biometric modalities,the human ear structure can offer unique and valuable discriminative characteristics for human recognition systems.In recent years,most existing traditional ear recognition systems have been designed based on computer vision models and have achieved successful results.Nevertheless,such traditional models can be sensitive to several unconstrained environmental factors.As such,some traits may be difficult to extract automatically but can still be semantically perceived as soft biometrics.This research proposes a new group of semantic features to be used as soft ear biometrics,mainly inspired by conventional descriptive traits used naturally by humans when identifying or describing each other.Hence,the research study is focused on the fusion of the soft ear biometric traits with traditional(hard)ear biometric features to investigate their validity and efficacy in augmenting human identification performance.The proposed framework has two subsystems:first,a computer vision-based subsystem,extracting traditional(hard)ear biometric traits using principal component analysis(PCA)and local binary patterns(LBP),and second,a crowdsourcing-based subsystem,deriving semantic(soft)ear biometric traits.Several feature-level fusion experiments were conducted using the AMI database to evaluate the proposed algorithm’s performance.The obtained results for both identification and verification showed that the proposed soft ear biometric information significantly improved the recognition performance of traditional ear biometrics,reaching up to 12%for LBP and 5%for PCA descriptors;when fusing all three capacities PCA,LBP,and soft traits using k-nearest neighbors(KNN)classifier.