In this research work,we proposed a medical image analysis framework with two separate releases whether or not Synovial Sarcoma(SS)is the cell structure for cancer.Within this framework the histopathology images are d...In this research work,we proposed a medical image analysis framework with two separate releases whether or not Synovial Sarcoma(SS)is the cell structure for cancer.Within this framework the histopathology images are decomposed into a third-level sub-band using a two-dimensional Discrete Wavelet Transform.Subsequently,the structure features(SFs)such as PrincipalComponentsAnalysis(PCA),Independent ComponentsAnalysis(ICA)and Linear Discriminant Analysis(LDA)were extracted from this subband image representation with the distribution of wavelet coefficients.These SFs are used as inputs of the Support Vector Machine(SVM)classifier.Also,classification of PCA+SVM,ICA+SVM,and LDA+SVM with Radial Basis Function(RBF)kernel the efficiency of the process is differentiated and compared with the best classification results.Furthermore,data collected on the internet from various histopathological centres via the Internet of Things(IoT)are stored and shared on blockchain technology across a wide range of image distribution across secure data IoT devices.Due to this,the minimum and maximum values of the kernel parameter are adjusted and updated periodically for the purpose of industrial application in device calibration.Consequently,these resolutions are presented with an excellent example of a technique for training and testing the cancer cell structure prognosis methods in spindle shaped cell(SSC)histopathological imaging databases.The performance characteristics of cross-validation are evaluated with the help of the receiver operating characteristics(ROC)curve,and significant differences in classification performance between the techniques are analyzed.The combination of LDA+SVM technique has been proven to be essential for intelligent SS cancer detection in the future,and it offers excellent classification accuracy,sensitivity,specificity.展开更多
Advancements in next-generation sequencer(NGS)platforms have improved NGS sequence data production and reduced the cost involved,which has resulted in the production of a large amount of genome data.The downstream ana...Advancements in next-generation sequencer(NGS)platforms have improved NGS sequence data production and reduced the cost involved,which has resulted in the production of a large amount of genome data.The downstream analysis of multiple associated sequences has become a bottleneck for the growing genomic data due to storage and space utilization issues in the domain of bioinformatics.The traditional string-matching algorithms are efficient for small sized data sequences and cannot process large amounts of data for downstream analysis.This study proposes a novel bit-parallelism algorithm called BitmapAligner to overcome the issues faced due to a large number of sequences and to improve the speed and quality of multiple sequence alignment(MSA).The input files(sequences)tested over BitmapAligner can be easily managed and organized using the Hadoop distributed file system.The proposed aligner converts the test file(the whole genome sequence)into binaries of an equal length of the sequence,line by line,before the sequence alignment processing.The Hadoop distributed file system splits the larger files into blocks,based on a defined block size,which is 128 MB by default.BitmapAligner can accurately process the sequence alignment using the bitmask approach on large-scale sequences after sorting the data.The experimental results indicate that BitmapAligner operates in real time,with a large number of sequences.Moreover,BitmapAligner achieves the exact start and end positions of the pattern sequence to test the MSA application in the whole genome query sequence.The MSA’s accuracy is verified by the bitmask indexing property of the bit-parallelism extended shifts(BXS)algorithm.The dynamic and exact approach of the BXS algorithm is implemented through the MapReduce function of Apache Hadoop.Conversely,the traditional seeds-and-extend approach faces the risk of errors while identifying the pattern sequences’positions.Moreover,the proposed model resolves the largescale data challenges that are covered through MapReduce in the Hadoop framework.Hive,Yarn,HBase,Cassandra,and many other pertinent flavors are to be used in the future for data structuring and annotations on the top layer of Hadoop since Hadoop is primarily used for data organization and handles text documents.展开更多
Heart disease is one of the leading causes of death in the world today.Prediction of heart disease is a prominent topic in the clinical data processing.To increase patient survival rates,early diagnosis of heart disea...Heart disease is one of the leading causes of death in the world today.Prediction of heart disease is a prominent topic in the clinical data processing.To increase patient survival rates,early diagnosis of heart disease is an important field of research in the medical field.There are many studies on the prediction of heart disease,but limited work is done on the selection of features.The selection of features is one of the best techniques for the diagnosis of heart diseases.In this research paper,we find optimal features using the brute-force algorithm,and machine learning techniques are used to improve the accuracy of heart disease prediction.For performance evaluation,accuracy,sensitivity,and specificity are used with split and cross-validation techniques.The results of the proposed technique are evaluated in three different heart disease datasets with a different number of records,and the proposed technique is found to have superior performance.The selection of optimized features generated by the brute force algorithm is used as input to machine learning algorithms such as Support Vector Machine(SVM),Random Forest(RF),K Nearest Neighbor(KNN),and Naive Bayes(NB).The proposed technique achieved 97%accuracy with Naive Bayes through split validation and 95%accuracy with Random Forest through cross-validation.Naive Bayes and Random Forest are found to outperform other classification approaches when accurately evaluated.The results of the proposed technique are compared with the results of the existing study,and the results of the proposed technique are found to be better than other state-of-the-artmethods.Therefore,our proposed approach plays an important role in the selection of important features and the automatic detection of heart disease.展开更多
In project management,effective cost estimation is one of the most cru-cial activities to efficiently manage resources by predicting the required cost to fulfill a given task.However,finding the best estimation results i...In project management,effective cost estimation is one of the most cru-cial activities to efficiently manage resources by predicting the required cost to fulfill a given task.However,finding the best estimation results in software devel-opment is challenging.Thus,accurate estimation of software development efforts is always a concern for many companies.In this paper,we proposed a novel soft-ware development effort estimation model based both on constructive cost model II(COCOMO II)and the artificial neural network(ANN).An artificial neural net-work enhances the COCOMO model,and the value of the baseline effort constant A is calibrated to use it in the proposed model equation.Three state-of-the-art publicly available datasets are used for experiments.The backpropagation feed-forward procedure used a training set by iteratively processing and training a neural network.The proposed model is tested on the test set.The estimated effort is compared with the actual effort value.Experimental results show that the effort estimated by the proposed model is very close to the real effort,thus enhanced the reliability and improving the software effort estimation accuracy.展开更多
基金This work was partly supported by the Technology development Program of MSS[No.S3033853]by Basic Science Research Program through the National Research Foundation of Korea(NRF)funded by the Ministry of Education(No.2020R1I1A3069700).
文摘In this research work,we proposed a medical image analysis framework with two separate releases whether or not Synovial Sarcoma(SS)is the cell structure for cancer.Within this framework the histopathology images are decomposed into a third-level sub-band using a two-dimensional Discrete Wavelet Transform.Subsequently,the structure features(SFs)such as PrincipalComponentsAnalysis(PCA),Independent ComponentsAnalysis(ICA)and Linear Discriminant Analysis(LDA)were extracted from this subband image representation with the distribution of wavelet coefficients.These SFs are used as inputs of the Support Vector Machine(SVM)classifier.Also,classification of PCA+SVM,ICA+SVM,and LDA+SVM with Radial Basis Function(RBF)kernel the efficiency of the process is differentiated and compared with the best classification results.Furthermore,data collected on the internet from various histopathological centres via the Internet of Things(IoT)are stored and shared on blockchain technology across a wide range of image distribution across secure data IoT devices.Due to this,the minimum and maximum values of the kernel parameter are adjusted and updated periodically for the purpose of industrial application in device calibration.Consequently,these resolutions are presented with an excellent example of a technique for training and testing the cancer cell structure prognosis methods in spindle shaped cell(SSC)histopathological imaging databases.The performance characteristics of cross-validation are evaluated with the help of the receiver operating characteristics(ROC)curve,and significant differences in classification performance between the techniques are analyzed.The combination of LDA+SVM technique has been proven to be essential for intelligent SS cancer detection in the future,and it offers excellent classification accuracy,sensitivity,specificity.
基金This work was supported in part by the National Research Foundation of Korea(NRF)grant funded by the Korea government(MSIT)(No.2018R1C1B5084424)in part by the Basic Science Research Program through the National Research Foundation of Korea(NRF)funded by the Ministry of Education(No.2019R1A6A1A03032119).
文摘Advancements in next-generation sequencer(NGS)platforms have improved NGS sequence data production and reduced the cost involved,which has resulted in the production of a large amount of genome data.The downstream analysis of multiple associated sequences has become a bottleneck for the growing genomic data due to storage and space utilization issues in the domain of bioinformatics.The traditional string-matching algorithms are efficient for small sized data sequences and cannot process large amounts of data for downstream analysis.This study proposes a novel bit-parallelism algorithm called BitmapAligner to overcome the issues faced due to a large number of sequences and to improve the speed and quality of multiple sequence alignment(MSA).The input files(sequences)tested over BitmapAligner can be easily managed and organized using the Hadoop distributed file system.The proposed aligner converts the test file(the whole genome sequence)into binaries of an equal length of the sequence,line by line,before the sequence alignment processing.The Hadoop distributed file system splits the larger files into blocks,based on a defined block size,which is 128 MB by default.BitmapAligner can accurately process the sequence alignment using the bitmask approach on large-scale sequences after sorting the data.The experimental results indicate that BitmapAligner operates in real time,with a large number of sequences.Moreover,BitmapAligner achieves the exact start and end positions of the pattern sequence to test the MSA application in the whole genome query sequence.The MSA’s accuracy is verified by the bitmask indexing property of the bit-parallelism extended shifts(BXS)algorithm.The dynamic and exact approach of the BXS algorithm is implemented through the MapReduce function of Apache Hadoop.Conversely,the traditional seeds-and-extend approach faces the risk of errors while identifying the pattern sequences’positions.Moreover,the proposed model resolves the largescale data challenges that are covered through MapReduce in the Hadoop framework.Hive,Yarn,HBase,Cassandra,and many other pertinent flavors are to be used in the future for data structuring and annotations on the top layer of Hadoop since Hadoop is primarily used for data organization and handles text documents.
基金This research was supported by Basic Science Research Program through the National Research Foundation of Korea(NRF)funded by the Ministry of Education(No.2020R1I1A3069700).
文摘Heart disease is one of the leading causes of death in the world today.Prediction of heart disease is a prominent topic in the clinical data processing.To increase patient survival rates,early diagnosis of heart disease is an important field of research in the medical field.There are many studies on the prediction of heart disease,but limited work is done on the selection of features.The selection of features is one of the best techniques for the diagnosis of heart diseases.In this research paper,we find optimal features using the brute-force algorithm,and machine learning techniques are used to improve the accuracy of heart disease prediction.For performance evaluation,accuracy,sensitivity,and specificity are used with split and cross-validation techniques.The results of the proposed technique are evaluated in three different heart disease datasets with a different number of records,and the proposed technique is found to have superior performance.The selection of optimized features generated by the brute force algorithm is used as input to machine learning algorithms such as Support Vector Machine(SVM),Random Forest(RF),K Nearest Neighbor(KNN),and Naive Bayes(NB).The proposed technique achieved 97%accuracy with Naive Bayes through split validation and 95%accuracy with Random Forest through cross-validation.Naive Bayes and Random Forest are found to outperform other classification approaches when accurately evaluated.The results of the proposed technique are compared with the results of the existing study,and the results of the proposed technique are found to be better than other state-of-the-artmethods.Therefore,our proposed approach plays an important role in the selection of important features and the automatic detection of heart disease.
基金This work was supported by the Technology development Program of MSS[No.S3033853].
文摘In project management,effective cost estimation is one of the most cru-cial activities to efficiently manage resources by predicting the required cost to fulfill a given task.However,finding the best estimation results in software devel-opment is challenging.Thus,accurate estimation of software development efforts is always a concern for many companies.In this paper,we proposed a novel soft-ware development effort estimation model based both on constructive cost model II(COCOMO II)and the artificial neural network(ANN).An artificial neural net-work enhances the COCOMO model,and the value of the baseline effort constant A is calibrated to use it in the proposed model equation.Three state-of-the-art publicly available datasets are used for experiments.The backpropagation feed-forward procedure used a training set by iteratively processing and training a neural network.The proposed model is tested on the test set.The estimated effort is compared with the actual effort value.Experimental results show that the effort estimated by the proposed model is very close to the real effort,thus enhanced the reliability and improving the software effort estimation accuracy.