期刊文献+
共找到119,413篇文章
< 1 2 250 >
每页显示 20 50 100
Quantifying the agreement and accuracy characteristics of four satellite-based LULC products for cropland classification in China
1
作者 Jie Xue Xianglin Zhang +3 位作者 Songchao Chen Bifeng Hu Nan Wang Zhou Shi 《Journal of Integrative Agriculture》 SCIE CSCD 2024年第1期283-297,共15页
Various land use and land cover(LULC)products have been produced over the past decade with the development of remote sensing technology.Despite the differences in LULC classification schemes,there is a lack of researc... Various land use and land cover(LULC)products have been produced over the past decade with the development of remote sensing technology.Despite the differences in LULC classification schemes,there is a lack of research on assessing the accuracy of their application to croplands in a unified framework.Thus,this study evaluated the spatial and area accuracies of cropland classification for four commonly used global LULC products(i.e.,MCD12Q1V6,GlobCover2009,FROM-GLC and GlobeLand30)based on the harmonised FAO criterion,and quantified the relationships between four factors(i.e.,slope,elevation,field size and crop system)and cropland classification agreement.The validation results indicated that MCD12Q1 and GlobeLand30 performed well in cropland classification regarding spatial consistency,with overall accuracies of 94.90 and 93.52%,respectively.The FROMGLC showed the worst performance,with an overall accuracy of 83.17%.Overlaying the cropland generated by the four global LULC products,we found the proportions of complete agreement and disagreement were 15.51 and 44.72% for the cropland classification,respectively.High consistency was mainly observed in the Northeast China Plain,the Huang-Huai-Hai Plain and the northern part of the Middle-lower Yangtze Plain,China.In contrast,low consistency was detected primarily on the eastern edge of the northern and semiarid region,the Yunnan-Guizhou Plateau and southern China.Field size was the most important factor for mapping cropland.For area accuracy,compared with China Statistical Yearbook data at the provincial scale,the accuracies of different products in descending order were:GlobeLand30,FROM-GLC,MCD12Q1,and GlobCover2009.The cropland classification schemes mainly caused large area deviations among the four products,and they also resulted in the different ranks of spatial accuracy and area accuracy among the four products.Our results can provide valuable suggestions for selecting cropland products at the national or provincial scale and help cropland mapping and reconstruction,which is essential for food security and crop management,so they can also contribute to achieving the Sustainable Development Goals issued by the United Nations. 展开更多
关键词 global LULC products cropland mapping accuracy evaluation food security China
下载PDF
Intrahepatic portal venous systems in adult patients with cavernous transformation of portal vein: Imaging features and a new classification 被引量:1
2
作者 Xin Huang Qian Lu +5 位作者 Yue-Wei Zhang Lin Zhang Zhi-Zhong Ren Xiao-Wei Yang Ying Liu Rui Tang 《Hepatobiliary & Pancreatic Diseases International》 SCIE CAS CSCD 2024年第5期481-486,共6页
Background: Cavernous transformation of the portal vein(CTPV) due to portal vein obstruction is a rare vascular anomaly defined as the formation of multiple collateral vessels in the hepatic hilum. This study aimed to... Background: Cavernous transformation of the portal vein(CTPV) due to portal vein obstruction is a rare vascular anomaly defined as the formation of multiple collateral vessels in the hepatic hilum. This study aimed to investigate the imaging features of intrahepatic portal vein in adult patients with CTPV and establish the relationship between the manifestations of intrahepatic portal vein and the progression of CTPV. Methods: We retrospectively analyzed 14 CTPV patients in Beijing Tsinghua Changgung Hospital. All patients underwent both direct portal venography(DPV) and computed tomography angiography(CTA) to reveal the manifestations of the portal venous system. The vessels measured included the left portal vein(LPV), right portal vein(RPV), main portal vein(MPV) and the portal vein bifurcation(PVB). Results: Nine males and 5 females, with a median age of 40.5 years, were included in the study. No significant difference was found in the diameters of the LPV or RPV measured by DPV and CTA. The visualization in terms of LPV, RPV and PVB measured by DPV was higher than that by CTA. There was a significant association between LPV/RPV and PVB/MPV in term of visibility revealed with DPV( P = 0.01), while this association was not observed with CTA. According to the imaging features of the portal vein measured by DPV, CTPV was classified into three categories to facilitate the diagnosis and treatment. Conclusions: DPV was more accurate than CTA for revealing the course of the intrahepatic portal vein in patients with CTPV. The classification of CTPV, that originated from the imaging features of the portal vein revealed by DPV, may provide a new perspective for the diagnosis and treatment of CTPV. 展开更多
关键词 Cavernous transformation of the portal vein classification Direct portal venography Intrahepatic portal venous system
下载PDF
Deep Learning Model Ensemble for the Accuracy of Classification Degenerative Arthritis
3
作者 Sang-min Lee Namgi Kim 《Computers, Materials & Continua》 SCIE EI 2023年第4期1981-1994,共14页
Artificial intelligence technologies are being studied to provide scientific evidence in the medical field and developed for use as diagnostic tools.This study focused on deep learning models to classify degenerative ... Artificial intelligence technologies are being studied to provide scientific evidence in the medical field and developed for use as diagnostic tools.This study focused on deep learning models to classify degenerative arthritis into Kellgren–Lawrence grades.Specifically,degenerative arthritis was assessed by X-ray radiographic images and classified into five classes.Subsequently,the use of various deep learning models was investigated for automating the degenerative arthritis classification process.Although research on the classification of osteoarthritis using deep learning has been conducted in previous studies,only local models have been used,and an ensemble of deep learning models has never been applied to obtain more accurate results.To address this issue,this study compared the classification performance of deep learning models,includingVGGNet,DenseNet,ResNet,TinyNet,EfficientNet,MobileNet,Xception,and ViT,on a dataset commonly used for osteoarthritis classification tasks.Our experimental results verified that even without applying a separate methodology,the performance of the ensemble was comparable to that of existing studies that only used the latest deep learning model and changed the learning method.From the trained models,two ensembles were created and evaluated:weight and specialist.The weight ensemble showed an improvement in accuracy of 1%,and the proposed specialist ensemble improved accuracy,precision,recall,and F1 score by 5%,6%,6%,and 6%,respectively,compared with the results of prior studies. 展开更多
关键词 Knee osteoarthritis deep learning convolutional neural network Kellgren–Lawrence grade classification knee X-ray
下载PDF
Performance evaluation of seven multi-label classification methods on real-world patent and publication datasets
4
作者 Shuo Xu Yuefu Zhang +1 位作者 Xin An Sainan Pi 《Journal of Data and Information Science》 CSCD 2024年第2期81-103,共23页
Purpose:Many science,technology and innovation(STI)resources are attached with several different labels.To assign automatically the resulting labels to an interested instance,many approaches with good performance on t... Purpose:Many science,technology and innovation(STI)resources are attached with several different labels.To assign automatically the resulting labels to an interested instance,many approaches with good performance on the benchmark datasets have been proposed for multi-label classification task in the literature.Furthermore,several open-source tools implementing these approaches have also been developed.However,the characteristics of real-world multi-label patent and publication datasets are not completely in line with those of benchmark ones.Therefore,the main purpose of this paper is to evaluate comprehensively seven multi-label classification methods on real-world datasets.Research limitations:Three real-world datasets differ in the following aspects:statement,data quality,and purposes.Additionally,open-source tools designed for multi-label classification also have intrinsic differences in their approaches for data processing and feature selection,which in turn impacts the performance of a multi-label classification approach.In the near future,we will enhance experimental precision and reinforce the validity of conclusions by employing more rigorous control over variables through introducing expanded parameter settings.Practical implications:The observed Macro F1 and Micro F1 scores on real-world datasets typically fall short of those achieved on benchmark datasets,underscoring the complexity of real-world multi-label classification tasks.Approaches leveraging deep learning techniques offer promising solutions by accommodating the hierarchical relationships and interdependencies among labels.With ongoing enhancements in deep learning algorithms and large-scale models,it is expected that the efficacy of multi-label classification tasks will be significantly improved,reaching a level of practical utility in the foreseeable future.Originality/value:(1)Seven multi-label classification methods are comprehensively compared on three real-world datasets.(2)The TextCNN and TextRCNN models perform better on small-scale datasets with more complex hierarchical structure of labels and more balanced document-label distribution.(3)The MLkNN method works better on the larger-scale dataset with more unbalanced document-label distribution. 展开更多
关键词 Multi-label classification Real-World datasets Hierarchical structure classification system Label correlation Machine learning
下载PDF
Classification of Sailboat Tell Tail Based on Deep Learning
5
作者 CHANG Xiaofeng YU Jintao +3 位作者 GAO Ying DING Hongchen LIU Yulong YU Huaming 《Journal of Ocean University of China》 SCIE CAS CSCD 2024年第3期710-720,共11页
The tell tail is usually placed on the triangular sail to display the running state of the air flow on the sail surface.It is of great significance to make accurate judgement on the drift of the tell tail of the sailb... The tell tail is usually placed on the triangular sail to display the running state of the air flow on the sail surface.It is of great significance to make accurate judgement on the drift of the tell tail of the sailboat during sailing for the best sailing effect.Normally it is difficult for sailors to keep an eye for a long time on the tell sail for accurate judging its changes,affected by strong sunlight and visual fatigue.In this case,we adopt computer vision technology in hope of helping the sailors judge the changes of the tell tail in ease with ease.This paper proposes for the first time a method to classify sailboat tell tails based on deep learning and an expert guidance system,supported by a sailboat tell tail classification data set on the expert guidance system of interpreting the tell tails states in different sea wind conditions,including the feature extraction performance.Considering the expression capabilities that vary with the computational features in different visual tasks,the paper focuses on five tell tail computing features,which are recoded by an automatic encoder and classified by a SVM classifier.All experimental samples were randomly divided into five groups,and four groups were selected from each group as the training set to train the classifier.The remaining one group was used as the test set for testing.The highest resolution value of the ResNet network was 80.26%.To achieve better operational results on the basis of deep computing features obtained through the ResNet network in the experiments.The method can be used to assist the sailors in making better judgement about the tell tail changes during sailing. 展开更多
关键词 tell tail sailboat classification deep learning
下载PDF
Empowering Diagnosis: Cutting-Edge Segmentation and Classification in Lung Cancer Analysis
6
作者 Iftikhar Naseer Tehreem Masood +4 位作者 Sheeraz Akram Zulfiqar Ali Awais Ahmad Shafiq Ur Rehman Arfan Jaffar 《Computers, Materials & Continua》 SCIE EI 2024年第6期4963-4977,共15页
Lung cancer is a leading cause of global mortality rates.Early detection of pulmonary tumors can significantly enhance the survival rate of patients.Recently,various Computer-Aided Diagnostic(CAD)methods have been dev... Lung cancer is a leading cause of global mortality rates.Early detection of pulmonary tumors can significantly enhance the survival rate of patients.Recently,various Computer-Aided Diagnostic(CAD)methods have been developed to enhance the detection of pulmonary nodules with high accuracy.Nevertheless,the existing method-ologies cannot obtain a high level of specificity and sensitivity.The present study introduces a novel model for Lung Cancer Segmentation and Classification(LCSC),which incorporates two improved architectures,namely the improved U-Net architecture and the improved AlexNet architecture.The LCSC model comprises two distinct stages.The first stage involves the utilization of an improved U-Net architecture to segment candidate nodules extracted from the lung lobes.Subsequently,an improved AlexNet architecture is employed to classify lung cancer.During the first stage,the proposed model demonstrates a dice accuracy of 0.855,a precision of 0.933,and a recall of 0.789 for the segmentation of candidate nodules.The suggested improved AlexNet architecture attains 97.06%accuracy,a true positive rate of 96.36%,a true negative rate of 97.77%,a positive predictive value of 97.74%,and a negative predictive value of 96.41%for classifying pulmonary cancer as either benign or malignant.The proposed LCSC model is tested and evaluated employing the publically available dataset furnished by the Lung Image Database Consortium and Image Database Resource Initiative(LIDC-IDRI).This proposed technique exhibits remarkable performance compared to the existing methods by using various evaluation parameters. 展开更多
关键词 Lung cancer SEGMENTATION AlexNet U-Net classification
下载PDF
Comprehensive understanding of glioblastoma molecular phenotypes:classification,characteristics,and transition
7
作者 Can Xu Pengyu Hou +7 位作者 Xiang Li Menglin Xiao Ziqi Zhang Ziru Li Jianglong Xu Guoming Liu Yanli Tan Chuan Fang 《Cancer Biology & Medicine》 SCIE CAS CSCD 2024年第5期363-381,共19页
Among central nervous system-associated malignancies,glioblastoma(GBM)is the most common and has the highest mortality rate.The high heterogeneity of GBM cell types and the complex tumor microenvironment frequently le... Among central nervous system-associated malignancies,glioblastoma(GBM)is the most common and has the highest mortality rate.The high heterogeneity of GBM cell types and the complex tumor microenvironment frequently lead to tumor recurrence and sudden relapse in patients treated with temozolomide.In precision medicine,research on GBM treatment is increasingly focusing on molecular subtyping to precisely characterize the cellular and molecular heterogeneity,as well as the refractory nature of GBM toward therapy.Deep understanding of the different molecular expression patterns of GBM subtypes is critical.Researchers have recently proposed tetra fractional or tripartite methods for detecting GBM molecular subtypes.The various molecular subtypes of GBM show significant differences in gene expression patterns and biological behaviors.These subtypes also exhibit high plasticity in their regulatory pathways,oncogene expression,tumor microenvironment alterations,and differential responses to standard therapy.Herein,we summarize the current molecular typing scheme of GBM and the major molecular/genetic characteristics of each subtype.Furthermore,we review the mesenchymal transition mechanisms of GBM under various regulators. 展开更多
关键词 GLIOBLASTOMA molecular phenotype classification CHARACTERISTIC mesenchymal transition
下载PDF
Urea breath test for Helicobacter pylori infection in adult dyspeptic patients: A meta-analysis of diagnostic test accuracy
8
作者 Fabian Fellipe Bueno Lemos Caroline Tianeze de Castro +9 位作者 Marcel Silva Luz Gabriel Reis Rocha Gabriel Lima Correa Santos Luís Guilherme de Oliveira Silva Mariana Santos Calmon Cláudio Lima Souza Ana Carla Zarpelon-Schutz Kádima Nayara Teixeira Dulciene Maria de Magalhães Queiroz Fabrício Freire de Melo 《World Journal of Gastroenterology》 SCIE CAS 2024年第6期579-598,共20页
BACKGROUND Helicobacter pylori(H.pylori)infection has been well-established as a significant risk factor for several gastrointestinal disorders.The urea breath test(UBT)has emerged as a leading non-invasive method for... BACKGROUND Helicobacter pylori(H.pylori)infection has been well-established as a significant risk factor for several gastrointestinal disorders.The urea breath test(UBT)has emerged as a leading non-invasive method for detecting H.pylori.Despite numerous studies confirming its substantial accuracy,the reliability of UBT results is often compromised by inherent limitations.These findings underscore the need for a rigorous statistical synthesis to clarify and reconcile the diagnostic accuracy of the UBT for the diagnosis of H.pylori infection.AIM To determine and compare the diagnostic accuracy of 13C-UBT and 14C-UBT for H.pylori infection in adult patients with dyspepsia.METHODS We conducted an independent search of the PubMed/MEDLINE,EMBASE,and Cochrane Central databases until April 2022.Our search included diagnostic accuracy studies that evaluated at least one of the index tests(^(13)C-UBT or ^(14)C-UBT)against a reference standard.We used the QUADAS-2 tool to assess the methodo-logical quality of the studies.We utilized the bivariate random-effects model to calculate sensitivity,specificity,positive and negative test likelihood ratios(LR+and LR-),as well as the diagnostic odds ratio(DOR),and their 95%confidence intervals.We conducted subgroup analyses based on urea dosing,time after urea administration,and assessment technique.To investigate a possible threshold effect,we conducted Spearman correlation analysis,and we generated summary receiver operating characteristic(SROC)curves to assess heterogeneity.Finally,we visually inspected a funnel plot and used Egger’s test to evaluate publication bias.endorsing both as reliable diagnostic tools in clinical practice.CONCLUSION In summary,our study has demonstrated that ^(13)C-UBT has been found to outperform the ^(14)C-UBT,making it the preferred diagnostic approach.Additionally,our results emphasize the significance of carefully considering urea dosage,assessment timing,and measurement techniques for both tests to enhance diagnostic precision.Nevertheless,it is crucial for researchers and clinicians to evaluate the strengths and limitations of our findings before implementing them in practice. 展开更多
关键词 Helicobacter pylori Urea breath test DIAGNOSIS Diagnostic test accuracy META-ANALYSIS
下载PDF
Curve Classification Based onMean-Variance Feature Weighting and Its Application
9
作者 Zewen Zhang Sheng Zhou Chunzheng Cao 《Computers, Materials & Continua》 SCIE EI 2024年第5期2465-2480,共16页
The classification of functional data has drawn much attention in recent years.The main challenge is representing infinite-dimensional functional data by finite-dimensional features while utilizing those features to a... The classification of functional data has drawn much attention in recent years.The main challenge is representing infinite-dimensional functional data by finite-dimensional features while utilizing those features to achieve better classification accuracy.In this paper,we propose a mean-variance-based(MV)feature weighting method for classifying functional data or functional curves.In the feature extraction stage,each sample curve is approximated by B-splines to transfer features to the coefficients of the spline basis.After that,a feature weighting approach based on statistical principles is introduced by comprehensively considering the between-class differences and within-class variations of the coefficients.We also introduce a scaling parameter to adjust the gap between the weights of features.The new feature weighting approach can adaptively enhance noteworthy local features while mitigating the impact of confusing features.The algorithms for feature weighted K-nearest neighbor and support vector machine classifiers are both provided.Moreover,the new approach can be well integrated into existing functional data classifiers,such as the generalized functional linear model and functional linear discriminant analysis,resulting in a more accurate classification.The performance of the mean-variance-based classifiers is evaluated by simulation studies and real data.The results show that the newfeatureweighting approach significantly improves the classification accuracy for complex functional data. 展开更多
关键词 Functional data analysis classification feature weighting B-SPLINES
下载PDF
A Robust Approach for Multi Classification-Based Intrusion Detection through Stacking Deep Learning Models
10
作者 Samia Allaoua Chelloug 《Computers, Materials & Continua》 SCIE EI 2024年第6期4845-4861,共17页
Intrusion detection is a predominant task that monitors and protects the network infrastructure.Therefore,many datasets have been published and investigated by researchers to analyze and understand the problem of intr... Intrusion detection is a predominant task that monitors and protects the network infrastructure.Therefore,many datasets have been published and investigated by researchers to analyze and understand the problem of intrusion prediction and detection.In particular,the Network Security Laboratory-Knowledge Discovery in Databases(NSL-KDD)is an extensively used benchmark dataset for evaluating intrusion detection systems(IDSs)as it incorporates various network traffic attacks.It is worth mentioning that a large number of studies have tackled the problem of intrusion detection using machine learning models,but the performance of these models often decreases when evaluated on new attacks.This has led to the utilization of deep learning techniques,which have showcased significant potential for processing large datasets and therefore improving detection accuracy.For that reason,this paper focuses on the role of stacking deep learning models,including convolution neural network(CNN)and deep neural network(DNN)for improving the intrusion detection rate of the NSL-KDD dataset.Each base model is trained on the NSL-KDD dataset to extract significant features.Once the base models have been trained,the stacking process proceeds to the second stage,where a simple meta-model has been trained on the predictions generated from the proposed base models.The combination of the predictions allows the meta-model to distinguish different classes of attacks and increase the detection rate.Our experimental evaluations using the NSL-KDD dataset have shown the efficacy of stacking deep learning models for intrusion detection.The performance of the ensemble of base models,combined with the meta-model,exceeds the performance of individual models.Our stacking model has attained an accuracy of 99%and an average F1-score of 93%for the multi-classification scenario.Besides,the training time of the proposed ensemble model is lower than the training time of benchmark techniques,demonstrating its efficiency and robustness. 展开更多
关键词 Intrusion detection multi classification deep learning STACKING NSL-KDD
下载PDF
Classification and Comprehension of Software Requirements Using Ensemble Learning
11
作者 Jalil Abbas Arshad Ahmad +4 位作者 Syed Muqsit Shaheed Rubia Fatima Sajid Shah Mohammad Elaffendi Gauhar Ali 《Computers, Materials & Continua》 SCIE EI 2024年第8期2839-2855,共17页
The software development process mostly depends on accurately identifying both essential and optional features.Initially,user needs are typically expressed in free-form language,requiring significant time and human re... The software development process mostly depends on accurately identifying both essential and optional features.Initially,user needs are typically expressed in free-form language,requiring significant time and human resources to translate these into clear functional and non-functional requirements.To address this challenge,various machine learning(ML)methods have been explored to automate the understanding of these requirements,aiming to reduce time and human effort.However,existing techniques often struggle with complex instructions and large-scale projects.In our study,we introduce an innovative approach known as the Functional and Non-functional Requirements Classifier(FNRC).By combining the traditional random forest algorithm with the Accuracy Sliding Window(ASW)technique,we develop optimal sub-ensembles that surpass the initial classifier’s accuracy while using fewer trees.Experimental results demonstrate that our FNRC methodology performs robustly across different datasets,achieving a balanced Precision of 75%on the PROMISE dataset and an impressive Recall of 85%on the CCHIT dataset.Both datasets consistently maintain an F-measure around 64%,highlighting FNRC’s ability to effectively balance precision and recall in diverse scenarios.These findings contribute to more accurate and efficient software development processes,increasing the probability of achieving successful project outcomes. 展开更多
关键词 Ensemble learning machine learning non-functional requirements requirement engineering accuracy sliding window
下载PDF
Depression Intensity Classification from Tweets Using Fast Text Based Weighted Soft Voting Ensemble
12
作者 Muhammad Rizwan Muhammad Faheem Mushtaq +5 位作者 Maryam Rafiq Arif Mehmood Isabel de la Torre Diez Monica Gracia Villar Helena Garay Imran Ashraf 《Computers, Materials & Continua》 SCIE EI 2024年第2期2047-2066,共20页
Predicting depression intensity from microblogs and social media posts has numerous benefits and applications,including predicting early psychological disorders and stress in individuals or the general public.A major ... Predicting depression intensity from microblogs and social media posts has numerous benefits and applications,including predicting early psychological disorders and stress in individuals or the general public.A major challenge in predicting depression using social media posts is that the existing studies do not focus on predicting the intensity of depression in social media texts but rather only perform the binary classification of depression and moreover noisy data makes it difficult to predict the true depression in the social media text.This study intends to begin by collecting relevant Tweets and generating a corpus of 210000 public tweets using Twitter public application programming interfaces(APIs).A strategy is devised to filter out only depression-related tweets by creating a list of relevant hashtags to reduce noise in the corpus.Furthermore,an algorithm is developed to annotate the data into three depression classes:‘Mild,’‘Moderate,’and‘Severe,’based on International Classification of Diseases-10(ICD-10)depression diagnostic criteria.Different baseline classifiers are applied to the annotated dataset to get a preliminary idea of classification performance on the corpus.Further FastText-based model is applied and fine-tuned with different preprocessing techniques and hyperparameter tuning to produce the tuned model,which significantly increases the depression classification performance to an 84%F1 score and 90%accuracy compared to baselines.Finally,a FastText-based weighted soft voting ensemble(WSVE)is proposed to boost the model’s performance by combining several other classifiers and assigning weights to individual models according to their individual performances.The proposed WSVE outperformed all baselines as well as FastText alone,with an F1 of 89%,5%higher than FastText alone,and an accuracy of 93%,3%higher than FastText alone.The proposed model better captures the contextual features of the relatively small sample class and aids in the detection of early depression intensity prediction from tweets with impactful performances. 展开更多
关键词 Depression classification deep learning FastText machine learning
下载PDF
Learning feature alignment and dual correlation for few‐shot image classification
13
作者 Xilang Huang Seon Han Choi 《CAAI Transactions on Intelligence Technology》 SCIE EI 2024年第2期303-318,共16页
Few‐shot image classification is the task of classifying novel classes using extremely limited labelled samples.To perform classification using the limited samples,one solution is to learn the feature alignment(FA)in... Few‐shot image classification is the task of classifying novel classes using extremely limited labelled samples.To perform classification using the limited samples,one solution is to learn the feature alignment(FA)information between the labelled and unlabelled sample features.Most FA methods use the feature mean as the class prototype and calculate the correlation between prototype and unlabelled features to learn an alignment strategy.However,mean prototypes tend to degenerate informative features because spatial features at the same position may not be equally important for the final classification,leading to inaccurate correlation calculations.Therefore,the authors propose an effective intraclass FA strategy that aggregates semantically similar spatial features from an adaptive reference prototype in low‐dimensional feature space to obtain an informative prototype feature map for precise correlation computation.Moreover,a dual correlation module to learn the hard and soft correlations was developed by the authors.This module combines the correlation information between the prototype and unlabelled features in both the original and learnable feature spaces,aiming to produce a comprehensive cross‐correlation between the prototypes and unlabelled features.Using both FA and cross‐attention modules,our model can maintain informative class features and capture important shared features for classification.Experimental results on three few‐shot classification benchmarks show that the proposed method outperformed related methods and resulted in a 3%performance boost in the 1‐shot setting by inserting the proposed module into the related methods. 展开更多
关键词 image classification machine learning metric learning
下载PDF
Diagnostic Accuracy of Computerized Bowel Sound Analysis with Non-Invasive Devices for Irritable Bowel Syndrome:A Systematic Review and Meta-Analysis
14
作者 Xia-Xiao Yan Yue-Lun Zhang +2 位作者 Yu-Pei Zhang Ying-Yun Yang Dong Wu 《Chinese Medical Sciences Journal》 CAS CSCD 2024年第2期122-130,共9页
Objective To assess the diagnostic accuracy of bowel sound analysis for irritable bowel syndrome(IBS)with a systematic review and meta-analysis.Methods We searched MEDLINE,Embase,the Cochrane Library,Web of Science,an... Objective To assess the diagnostic accuracy of bowel sound analysis for irritable bowel syndrome(IBS)with a systematic review and meta-analysis.Methods We searched MEDLINE,Embase,the Cochrane Library,Web of Science,and IEEE Xplore databases until September 2023.Cross-sectional and case-control studies on diagnostic accuracy of bowel sound analysis for IBS were identified.We estimated the pooled sensitivity,specificity,positive likelihood ratio,negative likeli-hood ratio,and diagnostic odds ratio with a 95% confidence interval(CI),and plotted a summary receiver operat-ing characteristic curve and evaluated the area under the curve.Results Four studies were included.The pooled diagnostic sensitivity,specificity,positive likelihood ratio,nega-tive likelihood ratio,and diagnostic odds ratio were 0.94(95%CI,0.87‒0.97),0.89(95%CI,0.81‒0.94),8.43(95%CI,4.81‒14.78),0.07(95%CI,0.03‒0.15),and 118.86(95%CI,44.18‒319.75),respectively,with an area under the curve of 0.97(95%CI,0.95‒0.98).Conclusions Computerized bowel sound analysis is a promising tool for IBS.However,limited high-quality data make the results'validity and applicability questionable.There is a need for more diagnostic test accuracy studies and better wearable devices for monitoring and analysis of IBS. 展开更多
关键词 irritable bowel syndrome bowel sound analysis diagnostic accuracy systematic review META-ANALYSIS
下载PDF
Research on Multi-Scale Feature Fusion Network Algorithm Based on Brain Tumor Medical Image Classification
15
作者 Yuting Zhou Xuemei Yang +1 位作者 Junping Yin Shiqi Liu 《Computers, Materials & Continua》 SCIE EI 2024年第6期5313-5333,共21页
Gliomas have the highest mortality rate of all brain tumors.Correctly classifying the glioma risk period can help doctors make reasonable treatment plans and improve patients’survival rates.This paper proposes a hier... Gliomas have the highest mortality rate of all brain tumors.Correctly classifying the glioma risk period can help doctors make reasonable treatment plans and improve patients’survival rates.This paper proposes a hierarchical multi-scale attention feature fusion medical image classification network(HMAC-Net),which effectively combines global features and local features.The network framework consists of three parallel layers:The global feature extraction layer,the local feature extraction layer,and the multi-scale feature fusion layer.A linear sparse attention mechanism is designed in the global feature extraction layer to reduce information redundancy.In the local feature extraction layer,a bilateral local attention mechanism is introduced to improve the extraction of relevant information between adjacent slices.In the multi-scale feature fusion layer,a channel fusion block combining convolutional attention mechanism and residual inverse multi-layer perceptron is proposed to prevent gradient disappearance and network degradation and improve feature representation capability.The double-branch iterative multi-scale classification block is used to improve the classification performance.On the brain glioma risk grading dataset,the results of the ablation experiment and comparison experiment show that the proposed HMAC-Net has the best performance in both qualitative analysis of heat maps and quantitative analysis of evaluation indicators.On the dataset of skin cancer classification,the generalization experiment results show that the proposed HMAC-Net has a good generalization effect. 展开更多
关键词 Medical image classification feature fusion TRANSFORMER
下载PDF
Multiscale Fusion Transformer Network for Hyperspectral Image Classification
16
作者 Yuquan Gan Hao Zhang Chen Yi 《Journal of Beijing Institute of Technology》 EI CAS 2024年第3期255-270,共16页
Convolutional neural network(CNN)has excellent ability to model locally contextual information.However,CNNs face challenges for descripting long-range semantic features,which will lead to relatively low classification... Convolutional neural network(CNN)has excellent ability to model locally contextual information.However,CNNs face challenges for descripting long-range semantic features,which will lead to relatively low classification accuracy of hyperspectral images.To address this problem,this article proposes an algorithm based on multiscale fusion and transformer network for hyperspectral image classification.Firstly,the low-level spatial-spectral features are extracted by multi-scale residual structure.Secondly,an attention module is introduced to focus on the more important spatialspectral information.Finally,high-level semantic features are represented and learned by a token learner and an improved transformer encoder.The proposed algorithm is compared with six classical hyperspectral classification algorithms on real hyperspectral images.The experimental results show that the proposed algorithm effectively improves the land cover classification accuracy of hyperspectral images. 展开更多
关键词 hyperspectral image land cover classification MULTI-SCALE TRANSFORMER
下载PDF
Classification of congenital cataracts based on multidimensional phenotypes and its association with visual outcomes
17
作者 Yuan Tan Ying-Shi Zou +8 位作者 Ying-Lin Yu Le-Yi Hu Ting Zhang Hui Chen Ling Jin Duo-Ru Lin Yi-Zhi Liu Hao-Tian Lin Zhen-Zhen Liu 《International Journal of Ophthalmology(English edition)》 SCIE CAS 2024年第3期473-479,共7页
●AIM:To establish a classification for congenital cataracts that can facilitate individualized treatment and help identify individuals with a high likelihood of different visual outcomes.●METHODS:Consecutive patient... ●AIM:To establish a classification for congenital cataracts that can facilitate individualized treatment and help identify individuals with a high likelihood of different visual outcomes.●METHODS:Consecutive patients diagnosed with congenital cataracts and undergoing surgery between January 2005 and November 2021 were recruited.Data on visual outcomes and the phenotypic characteristics of ocular biometry and the anterior and posterior segments were extracted from the patients’medical records.A hierarchical cluster analysis was performed.The main outcome measure was the identification of distinct clusters of eyes with congenital cataracts.●RESULTS:A total of 164 children(299 eyes)were divided into two clusters based on their ocular features.Cluster 1(96 eyes)had a shorter axial length(mean±SD,19.44±1.68 mm),a low prevalence of macular abnormalities(1.04%),and no retinal abnormalities or posterior cataracts.Cluster 2(203 eyes)had a greater axial length(mean±SD,20.42±2.10 mm)and a higher prevalence of macular abnormalities(8.37%),retinal abnormalities(98.52%),and posterior cataracts(4.93%).Compared with the eyes in Cluster 2(57.14%),those in Cluster 1(71.88%)had a 2.2 times higher chance of good best-corrected visual acuity[<0.7 logMAR;OR(95%CI),2.20(1.25–3.81);P=0.006].●CONCLUSION:This retrospective study categorizes congenital cataracts into two distinct clusters,each associated with a different likelihood of visual outcomes.This innovative classification may enable the personalization and prioritization of early interventions for patients who may gain the greatest benefit,thereby making strides toward precision medicine in the field of congenital cataracts. 展开更多
关键词 classification congenital cataract PHENOTYPE visual acuity cluster analysis
下载PDF
A Bitcoin Address Multi-Classification Mechanism Based on Bipartite Graph-Based Maximization Consensus
18
作者 Lejun Zhang Junjie Zhang +4 位作者 Kentaroh Toyoda Yuan Liu Jing Qiu Zhihong Tian Ran Guo 《Computer Modeling in Engineering & Sciences》 SCIE EI 2024年第4期783-800,共18页
Bitcoin is widely used as the most classic electronic currency for various electronic services such as exchanges,gambling,marketplaces,and also scams such as high-yield investment projects.Identifying the services ope... Bitcoin is widely used as the most classic electronic currency for various electronic services such as exchanges,gambling,marketplaces,and also scams such as high-yield investment projects.Identifying the services operated by a Bitcoin address can help determine the risk level of that address and build an alert model accordingly.Feature engineering can also be used to flesh out labeled addresses and to analyze the current state of Bitcoin in a small way.In this paper,we address the problem of identifying multiple classes of Bitcoin services,and for the poor classification of individual addresses that do not have significant features,we propose a Bitcoin address identification scheme based on joint multi-model prediction using the mapping relationship between addresses and entities.The innovation of the method is to(1)Extract as many valuable features as possible when an address is given to facilitate the multi-class service identification task.(2)Unlike the general supervised model approach,this paper proposes a joint prediction scheme for multiple learners based on address-entity mapping relationships.Specifically,after obtaining the overall features,the address classification and entity clustering tasks are performed separately,and the results are subjected to graph-basedmaximization consensus.The final result ismade to baseline the individual address classification results while satisfying the constraint of having similarly behaving entities as far as possible.By testing and evaluating over 26,000 Bitcoin addresses,our feature extraction method captures more useful features.In addition,the combined multi-learner model obtained results that exceeded the baseline classifier reaching an accuracy of 77.4%. 展开更多
关键词 Bitcoin multi-service classification graph maximization consensus data security
下载PDF
A Novel Locomotion Rule Rmbedding Long Short-Term Memory Network with Attention for Human Locomotor Intent Classification Using Multi-Sensors Signals
19
作者 Jiajie Shen Yan Wang Dongxu Zhang 《Computers, Materials & Continua》 SCIE EI 2024年第6期4349-4370,共22页
Locomotor intent classification has become a research hotspot due to its importance to the development of assistive robotics and wearable devices.Previous work have achieved impressive performance in classifying stead... Locomotor intent classification has become a research hotspot due to its importance to the development of assistive robotics and wearable devices.Previous work have achieved impressive performance in classifying steady locomotion states.However,it remains challenging for these methods to attain high accuracy when facing transitions between steady locomotion states.Due to the similarities between the information of the transitions and their adjacent steady states.Furthermore,most of these methods rely solely on data and overlook the objective laws between physical activities,resulting in lower accuracy,particularly when encountering complex locomotion modes such as transitions.To address the existing deficiencies,we propose the locomotion rule embedding long short-term memory(LSTM)network with Attention(LREAL)for human locomotor intent classification,with a particular focus on transitions,using data from fewer sensors(two inertial measurement units and four goniometers).The LREAL network consists of two levels:One responsible for distinguishing between steady states and transitions,and the other for the accurate identification of locomotor intent.Each classifier in these levels is composed of multiple-LSTM layers and an attention mechanism.To introduce real-world motion rules and apply constraints to the network,a prior knowledge was added to the network via a rule-modulating block.The method was tested on the ENABL3S dataset,which contains continuous locomotion date for seven steady and twelve transitions states.Experimental results showed that the LREAL network could recognize locomotor intents with an average accuracy of 99.03%and 96.52%for the steady and transitions states,respectively.It is worth noting that the LREAL network accuracy for transition-state recognition improved by 0.18%compared to other state-of-the-art network,while using data from fewer sensors. 展开更多
关键词 Lower-limb prosthetics deep neural networks motion classification
下载PDF
Stochastic seismic inversion and Bayesian facies classification applied to porosity modeling and igneous rock identification
20
作者 Fábio Júnior Damasceno Fernandes Leonardo Teixeira +1 位作者 Antonio Fernando Menezes Freire Wagner Moreira Lupinacci 《Petroleum Science》 SCIE EI CAS CSCD 2024年第2期918-935,共18页
We apply stochastic seismic inversion and Bayesian facies classification for porosity modeling and igneous rock identification in the presalt interval of the Santos Basin. This integration of seismic and well-derived ... We apply stochastic seismic inversion and Bayesian facies classification for porosity modeling and igneous rock identification in the presalt interval of the Santos Basin. This integration of seismic and well-derived information enhances reservoir characterization. Stochastic inversion and Bayesian classification are powerful tools because they permit addressing the uncertainties in the model. We used the ES-MDA algorithm to achieve the realizations equivalent to the percentiles P10, P50, and P90 of acoustic impedance, a novel method for acoustic inversion in presalt. The facies were divided into five: reservoir 1,reservoir 2, tight carbonates, clayey rocks, and igneous rocks. To deal with the overlaps in acoustic impedance values of facies, we included geological information using a priori probability, indicating that structural highs are reservoir-dominated. To illustrate our approach, we conducted porosity modeling using facies-related rock-physics models for rock-physics inversion in an area with a well drilled in a coquina bank and evaluated the thickness and extension of an igneous intrusion near the carbonate-salt interface. The modeled porosity and the classified seismic facies are in good agreement with the ones observed in the wells. Notably, the coquinas bank presents an improvement in the porosity towards the top. The a priori probability model was crucial for limiting the clayey rocks to the structural lows. In Well B, the hit rate of the igneous rock in the three scenarios is higher than 60%, showing an excellent thickness-prediction capability. 展开更多
关键词 Stochastic inversion Bayesian classification Porosity modeling Carbonate reservoirs Igneous rocks
下载PDF
上一页 1 2 250 下一页 到第
使用帮助 返回顶部