Diabetic Eye Disease(DED)is a fundamental cause of blindness in human beings in the medical world.Different techniques are proposed to forecast and examine the stages in Prognostication of Diabetic Retinopathy(DR).The...Diabetic Eye Disease(DED)is a fundamental cause of blindness in human beings in the medical world.Different techniques are proposed to forecast and examine the stages in Prognostication of Diabetic Retinopathy(DR).The Machine Learning(ML)and the Deep Learning(DL)algorithms are the predomi-nant techniques to project and explore the images of DR.Even though some solu-tions were adapted to challenge the cause of DR disease,still there should be an efficient and accurate DR prediction to be adapted to refine its performance.In this work,a hybrid technique was proposed for classification and prediction of DR.The proposed hybrid technique consists of Ensemble Learning(EL),2 Dimensional-Conventional Neural Network(2D-CNN),Transfer Learning(TL)and Correlation method.Initially,the Stochastic Gradient Boosting(SGB)EL method was used to predict the DR.Secondly,the boosting based EL method was used to predict the DR of images.Thirdly 2D-CNN was applied to categorize the various stages of DR images.Finally,the TL was adopted to transfer the clas-sification prediction to training datasets.When this TL was applied,a new predic-tion feature was increased.From the experiment,the proposed technique has achieved 97.8%of accuracy in prophecies of DR images and 98%accuracy in grading of images.The experiment was also extended to measure the sensitivity(99.6%)and specificity(97.3%)metrics.The predicted accuracy rate was com-pared with existing methods.展开更多
Communication is a basic need of every human being to exchange thoughts and interact with the society.Acute peoples usually confab through different spoken languages,whereas deaf people cannot do so.Therefore,the Sign...Communication is a basic need of every human being to exchange thoughts and interact with the society.Acute peoples usually confab through different spoken languages,whereas deaf people cannot do so.Therefore,the Sign Language(SL)is the communication medium of such people for their conversation and interaction with the society.The SL is expressed in terms of specific gesture for every word and a gesture is consisted in a sequence of performed signs.The acute people normally observe these signs to understand the difference between single and multiple gestures for singular and plural words respectively.The signs for singular words such as I,eat,drink,home are unalike the plural words as school,cars,players.A special training is required to gain the sufficient knowledge and practice so that people can differentiate and understand every gesture/sign appropriately.Innumerable researches have been performed to articulate the computer-based solution to understand the single gesture with the help of a single hand enumeration.The complete understanding of such communications are possible only with the help of this differentiation of gestures in computer-based solution of SL to cope with the real world environment.Hence,there is still a demand for specific environment to automate such a communication solution to interact with such type of special people.This research focuses on facilitating the deaf community by capturing the gestures in video format and then mapping and differentiating as single or multiple gestures used in words.Finally,these are converted into the respective words/sentences within a reasonable time.This provide a real time solution for the deaf people to communicate and interact with the society.展开更多
The novel coronavirus disease 2019(COVID-19)is a pandemic disease that is currently affecting over 200 countries around the world and impacting billions of people.The first step to mitigate and control its spread is t...The novel coronavirus disease 2019(COVID-19)is a pandemic disease that is currently affecting over 200 countries around the world and impacting billions of people.The first step to mitigate and control its spread is to identify and isolate the infected people.But,because of the lack of reverse transcription polymerase chain reaction(RT-CPR)tests,it is important to discover suspected COVID-19 cases as early as possible,such as by scan analysis and chest X-ray by radiologists.However,chest X-ray analysis is relatively time-consuming since it requires more than 15 minutes per case.In this paper,an automated novel detection model of COVID-19 cases is proposed to perform real-time detection of COVID-19 cases.The proposed model consists of three main stages:image segmentation using Harris Hawks optimizer,synthetic image augmentation using an enhanced Wasserstein And Auxiliary Classifier Generative Adversarial Network,and image classification using Conventional Neural Network.Raw chest X-ray images datasets are used to train and test the proposed model.Experiments demonstrate that the proposed model is very efficient in the automatic detection of COVID-19 positive cases.It achieved 99.4%accuracy,99.15%precision,99.35%recall,99.25%F-measure,and 98.5%specificity.展开更多
文摘Diabetic Eye Disease(DED)is a fundamental cause of blindness in human beings in the medical world.Different techniques are proposed to forecast and examine the stages in Prognostication of Diabetic Retinopathy(DR).The Machine Learning(ML)and the Deep Learning(DL)algorithms are the predomi-nant techniques to project and explore the images of DR.Even though some solu-tions were adapted to challenge the cause of DR disease,still there should be an efficient and accurate DR prediction to be adapted to refine its performance.In this work,a hybrid technique was proposed for classification and prediction of DR.The proposed hybrid technique consists of Ensemble Learning(EL),2 Dimensional-Conventional Neural Network(2D-CNN),Transfer Learning(TL)and Correlation method.Initially,the Stochastic Gradient Boosting(SGB)EL method was used to predict the DR.Secondly,the boosting based EL method was used to predict the DR of images.Thirdly 2D-CNN was applied to categorize the various stages of DR images.Finally,the TL was adopted to transfer the clas-sification prediction to training datasets.When this TL was applied,a new predic-tion feature was increased.From the experiment,the proposed technique has achieved 97.8%of accuracy in prophecies of DR images and 98%accuracy in grading of images.The experiment was also extended to measure the sensitivity(99.6%)and specificity(97.3%)metrics.The predicted accuracy rate was com-pared with existing methods.
基金The work presented in this paper is part of an ongoing research funded by Yayasan Universiti Teknologi PETRONAS Grant(015LC0-311 and 015LC0-029).
文摘Communication is a basic need of every human being to exchange thoughts and interact with the society.Acute peoples usually confab through different spoken languages,whereas deaf people cannot do so.Therefore,the Sign Language(SL)is the communication medium of such people for their conversation and interaction with the society.The SL is expressed in terms of specific gesture for every word and a gesture is consisted in a sequence of performed signs.The acute people normally observe these signs to understand the difference between single and multiple gestures for singular and plural words respectively.The signs for singular words such as I,eat,drink,home are unalike the plural words as school,cars,players.A special training is required to gain the sufficient knowledge and practice so that people can differentiate and understand every gesture/sign appropriately.Innumerable researches have been performed to articulate the computer-based solution to understand the single gesture with the help of a single hand enumeration.The complete understanding of such communications are possible only with the help of this differentiation of gestures in computer-based solution of SL to cope with the real world environment.Hence,there is still a demand for specific environment to automate such a communication solution to interact with such type of special people.This research focuses on facilitating the deaf community by capturing the gestures in video format and then mapping and differentiating as single or multiple gestures used in words.Finally,these are converted into the respective words/sentences within a reasonable time.This provide a real time solution for the deaf people to communicate and interact with the society.
文摘The novel coronavirus disease 2019(COVID-19)is a pandemic disease that is currently affecting over 200 countries around the world and impacting billions of people.The first step to mitigate and control its spread is to identify and isolate the infected people.But,because of the lack of reverse transcription polymerase chain reaction(RT-CPR)tests,it is important to discover suspected COVID-19 cases as early as possible,such as by scan analysis and chest X-ray by radiologists.However,chest X-ray analysis is relatively time-consuming since it requires more than 15 minutes per case.In this paper,an automated novel detection model of COVID-19 cases is proposed to perform real-time detection of COVID-19 cases.The proposed model consists of three main stages:image segmentation using Harris Hawks optimizer,synthetic image augmentation using an enhanced Wasserstein And Auxiliary Classifier Generative Adversarial Network,and image classification using Conventional Neural Network.Raw chest X-ray images datasets are used to train and test the proposed model.Experiments demonstrate that the proposed model is very efficient in the automatic detection of COVID-19 positive cases.It achieved 99.4%accuracy,99.15%precision,99.35%recall,99.25%F-measure,and 98.5%specificity.