Modern medicine is reliant on various medical imaging technologies for non-invasively observing patients’anatomy.However,the interpretation of medical images can be highly subjective and dependent on the expertise of...Modern medicine is reliant on various medical imaging technologies for non-invasively observing patients’anatomy.However,the interpretation of medical images can be highly subjective and dependent on the expertise of clinicians.Moreover,some potentially useful quantitative information in medical images,especially that which is not visible to the naked eye,is often ignored during clinical practice.In contrast,radiomics performs high-throughput feature extraction from medical images,which enables quantitative analysis of medical images and prediction of various clinical endpoints.Studies have reported that radiomics exhibits promising performance in diagnosis and predicting treatment responses and prognosis,demonstrating its potential to be a non-invasive auxiliary tool for personalized medicine.However,radiomics remains in a developmental phase as numerous technical challenges have yet to be solved,especially in feature engineering and statistical modeling.In this review,we introduce the current utility of radiomics by summarizing research on its application in the diagnosis,prognosis,and prediction of treatment responses in patients with cancer.We focus on machine learning approaches,for feature extraction and selection during feature engineering and for imbalanced datasets and multi-modality fusion during statistical modeling.Furthermore,we introduce the stability,reproducibility,and interpretability of features,and the generalizability and interpretability of models.Finally,we offer possible solutions to current challenges in radiomics research.展开更多
BACKGROUND Gastric cystica profunda(GCP)represents a rare condition characterized by cystic dilation of gastric glands within the mucosal and/or submucosal layers.GCP is often linked to,or may progress into,early gast...BACKGROUND Gastric cystica profunda(GCP)represents a rare condition characterized by cystic dilation of gastric glands within the mucosal and/or submucosal layers.GCP is often linked to,or may progress into,early gastric cancer(EGC).AIM To provide a comprehensive evaluation of the endoscopic features of GCP while assessing the efficacy of endoscopic treatment,thereby offering guidance for diagnosis and treatment.METHODS This retrospective study involved 104 patients with GCP who underwent endoscopic resection.Alongside demographic and clinical data,regular patient followups were conducted to assess local recurrence.RESULTS Among the 104 patients diagnosed with GCP who underwent endoscopic resection,12.5%had a history of previous gastric procedures.The primary site predominantly affected was the cardia(38.5%,n=40).GCP commonly exhibited intraluminal growth(99%),regular presentation(74.0%),and ulcerative mucosa(61.5%).The leading endoscopic feature was the mucosal lesion type(59.6%,n=62).The average maximum diameter was 20.9±15.3 mm,with mucosal involvement in 60.6%(n=63).Procedures lasted 73.9±57.5 min,achieving complete resection in 91.3%(n=95).Recurrence(4.8%)was managed via either surgical intervention(n=1)or through endoscopic resection(n=4).Final pathology confirmed that 59.6%of GCP cases were associated with EGC.Univariate analysis indicated that elderly males were more susceptible to GCP associated with EGC.Conversely,multivariate analysis identified lesion morphology and endoscopic features as significant risk factors.Survival analysis demonstrated no statistically significant difference in recurrence between GCP with and without EGC(P=0.72).CONCLUSION The findings suggested that endoscopic resection might serve as an effective and minimally invasive treatment for GCP with or without EGC.展开更多
Traumatic spinal cord injury is potentially catastrophic and can lead to permanent disability or even death.China has the largest population of patients with traumatic spinal cord injury.Previous studies of traumatic ...Traumatic spinal cord injury is potentially catastrophic and can lead to permanent disability or even death.China has the largest population of patients with traumatic spinal cord injury.Previous studies of traumatic spinal cord injury in China have mostly been regional in scope;national-level studies have been rare.To the best of our knowledge,no national-level study of treatment status and economic burden has been performed.This retrospective study aimed to examine the epidemiological and clinical features,treatment status,and economic burden of traumatic spinal cord injury in China at the national level.We included 13,465 traumatic spinal cord injury patients who were injured between January 2013 and December 2018 and treated in 30 hospitals in 11 provinces/municipalities representing all geographical divisions of China.Patient epidemiological and clinical features,treatment status,and total and daily costs were recorded.Trends in the percentage of traumatic spinal cord injuries among all hospitalized patients and among patients hospitalized in the orthopedic department and cost of care were assessed by annual percentage change using the Joinpoint Regression Program.The percentage of traumatic spinal cord injuries among all hospitalized patients and among patients hospitalized in the orthopedic department did not significantly change overall(annual percentage change,-0.5%and 2.1%,respectively).A total of 10,053(74.7%)patients underwent surgery.Only 2.8%of patients who underwent surgery did so within 24 hours of injury.A total of 2005(14.9%)patients were treated with high-dose(≥500 mg)methylprednisolone sodium succinate/methylprednisolone(MPSS/MP);615(4.6%)received it within 8 hours.The total cost for acute traumatic spinal cord injury decreased over the study period(-4.7%),while daily cost did not significantly change(1.0%increase).Our findings indicate that public health initiatives should aim at improving hospitals’ability to complete early surgery within 24 hours,which is associated with improved sensorimotor recovery,increasing the awareness rate of clinical guidelines related to high-dose MPSS/MP to reduce the use of the treatment with insufficient evidence.展开更多
Cultural relics line graphic serves as a crucial form of traditional artifact information documentation,which is a simple and intuitive product with low cost of displaying compared with 3D models.Dimensionality reduct...Cultural relics line graphic serves as a crucial form of traditional artifact information documentation,which is a simple and intuitive product with low cost of displaying compared with 3D models.Dimensionality reduction is undoubtedly necessary for line drawings.However,most existing methods for artifact drawing rely on the principles of orthographic projection that always cannot avoid angle occlusion and data overlapping while the surface of cultural relics is complex.Therefore,conformal mapping was introduced as a dimensionality reduction way to compensate for the limitation of orthographic projection.Based on the given criteria for assessing surface complexity,this paper proposed a three-dimensional feature guideline extraction method for complex cultural relic surfaces.A 2D and 3D combined factor that measured the importance of points on describing surface features,vertex weight,was designed.Then the selection threshold for feature guideline extraction was determined based on the differences between vertex weight and shape index distributions.The feasibility and stability were verified through experiments conducted on real cultural relic surface data.Results demonstrated the ability of the method to address the challenges associated with the automatic generation of line drawings for complex surfaces.The extraction method and the obtained results will be useful for line graphic drawing,displaying and propaganda of cultural relics.展开更多
Sign language recognition is vital for enhancing communication accessibility among the Deaf and hard-of-hearing communities.In Japan,approximately 360,000 individualswith hearing and speech disabilities rely on Japane...Sign language recognition is vital for enhancing communication accessibility among the Deaf and hard-of-hearing communities.In Japan,approximately 360,000 individualswith hearing and speech disabilities rely on Japanese Sign Language(JSL)for communication.However,existing JSL recognition systems have faced significant performance limitations due to inherent complexities.In response to these challenges,we present a novel JSL recognition system that employs a strategic fusion approach,combining joint skeleton-based handcrafted features and pixel-based deep learning features.Our system incorporates two distinct streams:the first stream extracts crucial handcrafted features,emphasizing the capture of hand and body movements within JSL gestures.Simultaneously,a deep learning-based transfer learning stream captures hierarchical representations of JSL gestures in the second stream.Then,we concatenated the critical information of the first stream and the hierarchy of the second stream features to produce the multiple levels of the fusion features,aiming to create a comprehensive representation of the JSL gestures.After reducing the dimensionality of the feature,a feature selection approach and a kernel-based support vector machine(SVM)were used for the classification.To assess the effectiveness of our approach,we conducted extensive experiments on our Lab JSL dataset and a publicly available Arabic sign language(ArSL)dataset.Our results unequivocally demonstrate that our fusion approach significantly enhances JSL recognition accuracy and robustness compared to individual feature sets or traditional recognition methods.展开更多
Time series anomaly detection is crucial in various industrial applications to identify unusual behaviors within the time series data.Due to the challenges associated with annotating anomaly events,time series reconst...Time series anomaly detection is crucial in various industrial applications to identify unusual behaviors within the time series data.Due to the challenges associated with annotating anomaly events,time series reconstruction has become a prevalent approach for unsupervised anomaly detection.However,effectively learning representations and achieving accurate detection results remain challenging due to the intricate temporal patterns and dependencies in real-world time series.In this paper,we propose a cross-dimension attentive feature fusion network for time series anomaly detection,referred to as CAFFN.Specifically,a series and feature mixing block is introduced to learn representations in 1D space.Additionally,a fast Fourier transform is employed to convert the time series into 2D space,providing the capability for 2D feature extraction.Finally,a cross-dimension attentive feature fusion mechanism is designed that adaptively integrates features across different dimensions for anomaly detection.Experimental results on real-world time series datasets demonstrate that CAFFN performs better than other competing methods in time series anomaly detection.展开更多
With the rapid development of the mobile communication and the Internet,the previous web anomaly detectionand identificationmodels were built relying on security experts’empirical knowledge and attack features.Althou...With the rapid development of the mobile communication and the Internet,the previous web anomaly detectionand identificationmodels were built relying on security experts’empirical knowledge and attack features.Althoughthis approach can achieve higher detection performance,it requires huge human labor and resources to maintainthe feature library.In contrast,semantic feature engineering can dynamically discover new semantic featuresand optimize feature selection by automatically analyzing the semantic information contained in the data itself,thus reducing dependence on prior knowledge.However,current semantic features still have the problem ofsemantic expression singularity,as they are extracted from a single semantic mode such as word segmentation,character segmentation,or arbitrary semantic feature extraction.This paper extracts features of web requestsfrom dual semantic granularity,and proposes a semantic feature fusion method to solve the above problems.Themethod first preprocesses web requests,and extracts word-level and character-level semantic features of URLs viaconvolutional neural network(CNN),respectively.By constructing three loss functions to reduce losses betweenfeatures,labels and categories.Experiments on the HTTP CSIC 2010,Malicious URLs and HttpParams datasetsverify the proposedmethod.Results show that compared withmachine learning,deep learningmethods and BERTmodel,the proposed method has better detection performance.And it achieved the best detection rate of 99.16%in the dataset HttpParams.展开更多
Deception detection plays a crucial role in criminal investigation.Videos contain a wealth of information regarding apparent and physiological changes in individuals,and thus can serve as an effective means of decepti...Deception detection plays a crucial role in criminal investigation.Videos contain a wealth of information regarding apparent and physiological changes in individuals,and thus can serve as an effective means of deception detection.In this paper,we investigate video-based deception detection considering both apparent visual features such as eye gaze,head pose and facial action unit(AU),and non-contact heart rate detected by remote photoplethysmography(rPPG)technique.Multiple wrapper-based feature selection methods combined with the K-nearest neighbor(KNN)and support vector machine(SVM)classifiers are employed to screen the most effective features for deception detection.We evaluate the performance of the proposed method on both a self-collected physiological-assisted visual deception detection(PV3D)dataset and a public bag-oflies(BOL)dataset.Experimental results demonstrate that the SVM classifier with symbiotic organisms search(SOS)feature selection yields the best overall performance,with an area under the curve(AUC)of 83.27%and accuracy(ACC)of 83.33%for PV3D,and an AUC of 71.18%and ACC of 70.33%for BOL.This demonstrates the stability and effectiveness of the proposed method in video-based deception detection tasks.展开更多
Radiomics is a non-invasive method for extracting quantitative and higher-dimensional features from medical images for diagnosis.It has received great attention due to its huge application prospects in recent years.We...Radiomics is a non-invasive method for extracting quantitative and higher-dimensional features from medical images for diagnosis.It has received great attention due to its huge application prospects in recent years.We can know that the number of features selected by the existing radiomics feature selectionmethods is basically about ten.In this paper,a heuristic feature selection method based on frequency iteration and multiple supervised training mode is proposed.Based on the combination between features,it decomposes all features layer by layer to select the optimal features for each layer,then fuses the optimal features to form a local optimal group layer by layer and iterates to the global optimal combination finally.Compared with the currentmethod with the best prediction performance in the three data sets,thismethod proposed in this paper can reduce the number of features fromabout ten to about three without losing classification accuracy and even significantly improving classification accuracy.The proposed method has better interpretability and generalization ability,which gives it great potential in the feature selection of radiomics.展开更多
Biometric recognition is a widely used technology for user authentication.In the application of this technology,biometric security and recognition accuracy are two important issues that should be considered.In terms o...Biometric recognition is a widely used technology for user authentication.In the application of this technology,biometric security and recognition accuracy are two important issues that should be considered.In terms of biometric security,cancellable biometrics is an effective technique for protecting biometric data.Regarding recognition accuracy,feature representation plays a significant role in the performance and reliability of cancellable biometric systems.How to design good feature representations for cancellable biometrics is a challenging topic that has attracted a great deal of attention from the computer vision community,especially from researchers of cancellable biometrics.Feature extraction and learning in cancellable biometrics is to find suitable feature representations with a view to achieving satisfactory recognition performance,while the privacy of biometric data is protected.This survey informs the progress,trend and challenges of feature extraction and learning for cancellable biometrics,thus shedding light on the latest developments and future research of this area.展开更多
The classification of functional data has drawn much attention in recent years.The main challenge is representing infinite-dimensional functional data by finite-dimensional features while utilizing those features to a...The classification of functional data has drawn much attention in recent years.The main challenge is representing infinite-dimensional functional data by finite-dimensional features while utilizing those features to achieve better classification accuracy.In this paper,we propose a mean-variance-based(MV)feature weighting method for classifying functional data or functional curves.In the feature extraction stage,each sample curve is approximated by B-splines to transfer features to the coefficients of the spline basis.After that,a feature weighting approach based on statistical principles is introduced by comprehensively considering the between-class differences and within-class variations of the coefficients.We also introduce a scaling parameter to adjust the gap between the weights of features.The new feature weighting approach can adaptively enhance noteworthy local features while mitigating the impact of confusing features.The algorithms for feature weighted K-nearest neighbor and support vector machine classifiers are both provided.Moreover,the new approach can be well integrated into existing functional data classifiers,such as the generalized functional linear model and functional linear discriminant analysis,resulting in a more accurate classification.The performance of the mean-variance-based classifiers is evaluated by simulation studies and real data.The results show that the newfeatureweighting approach significantly improves the classification accuracy for complex functional data.展开更多
While single-modal visible light images or infrared images provide limited information,infrared light captures significant thermal radiation data,whereas visible light excels in presenting detailed texture information...While single-modal visible light images or infrared images provide limited information,infrared light captures significant thermal radiation data,whereas visible light excels in presenting detailed texture information.Com-bining images obtained from both modalities allows for leveraging their respective strengths and mitigating individual limitations,resulting in high-quality images with enhanced contrast and rich texture details.Such capabilities hold promising applications in advanced visual tasks including target detection,instance segmentation,military surveillance,pedestrian detection,among others.This paper introduces a novel approach,a dual-branch decomposition fusion network based on AutoEncoder(AE),which decomposes multi-modal features into intensity and texture information for enhanced fusion.Local contrast enhancement module(CEM)and texture detail enhancement module(DEM)are devised to process the decomposed images,followed by image fusion through the decoder.The proposed loss function ensures effective retention of key information from the source images of both modalities.Extensive comparisons and generalization experiments demonstrate the superior performance of our network in preserving pixel intensity distribution and retaining texture details.From the qualitative results,we can see the advantages of fusion details and local contrast.In the quantitative experiments,entropy(EN),mutual information(MI),structural similarity(SSIM)and other results have improved and exceeded the SOTA(State of the Art)model as a whole.展开更多
In the era of the Internet,widely used web applications have become the target of hacker attacks because they contain a large amount of personal information.Among these vulnerabilities,stealing private data through cr...In the era of the Internet,widely used web applications have become the target of hacker attacks because they contain a large amount of personal information.Among these vulnerabilities,stealing private data through crosssite scripting(XSS)attacks is one of the most commonly used attacks by hackers.Currently,deep learning-based XSS attack detection methods have good application prospects;however,they suffer from problems such as being prone to overfitting,a high false alarm rate,and low accuracy.To address these issues,we propose a multi-stage feature extraction and fusion model for XSS detection based on Random Forest feature enhancement.The model utilizes RandomForests to capture the intrinsic structure and patterns of the data by extracting leaf node indices as features,which are subsequentlymergedwith the original data features to forma feature setwith richer information content.Further feature extraction is conducted through three parallel channels.Channel I utilizes parallel onedimensional convolutional layers(1Dconvolutional layers)with different convolutional kernel sizes to extract local features at different scales and performmulti-scale feature fusion;Channel II employsmaximum one-dimensional pooling layers(max 1D pooling layers)of various sizes to extract key features from the data;and Channel III extracts global information bi-directionally using a Bi-Directional Long-Short TermMemory Network(Bi-LSTM)and incorporates a multi-head attention mechanism to enhance global features.Finally,effective classification and prediction of XSS are performed by fusing the features of the three channels.To test the effectiveness of the model,we conduct experiments on six datasets.We achieve an accuracy of 100%on the UNSW-NB15 dataset and 99.99%on the CICIDS2017 dataset,which is higher than that of the existing models.展开更多
The safety and stability of high-speed maglev trains traveling on viaducts in crosswinds critically depend on their aerodynamic characteristics.Therefore,this paper uses an improved delayed detached eddy simulation(ID...The safety and stability of high-speed maglev trains traveling on viaducts in crosswinds critically depend on their aerodynamic characteristics.Therefore,this paper uses an improved delayed detached eddy simulation(IDDES)method to investigate the aerodynamic features of high-speed maglev trains with different marshaling lengths under crosswinds.The effects of marshaling lengths(varying from 3-car to 8-car groups)on the train’s aerodynamic performance,surface pressure,and the flow field surrounding the train were investigated using the three-dimensional unsteady compressible Navier-Stokes(N-S)equations.The results showed that the marshaling lengths had minimal influence on the aerodynamic performance of the head and middle cars.Conversely,the marshaling lengths are negatively correlated with the time-average side force coefficient(CS)and time-average lift force coefficient(Cl)of the tail car.Compared to the tail car of the 3-car groups,the CS and Cl fell by 27.77%and 18.29%,respectively,for the tail car of the 8-car groups.It is essential to pay more attention to the operational safety of the head car,as it exhibits the highest time average CS.Additionally,the mean pressure difference between the two sides of the tail car body increased with the marshaling lengths,and the side force direction on the tail car was opposite to that of the head and middle cars.Furthermore,the turbulent kinetic energy of the wake structure on the windward side quickly decreased as marshaling lengths increased.展开更多
Feature matching plays a key role in computer vision. However, due to the limitations of the descriptors, the putative matches are inevitably contaminated by massive outliers.This paper attempts to tackle the outlier ...Feature matching plays a key role in computer vision. However, due to the limitations of the descriptors, the putative matches are inevitably contaminated by massive outliers.This paper attempts to tackle the outlier filtering problem from two aspects. First, a robust and efficient graph interaction model,is proposed, with the assumption that matches are correlated with each other rather than independently distributed. To this end, we construct a graph based on the local relationships of matches and formulate the outlier filtering task as a binary labeling energy minimization problem, where the pairwise term encodes the interaction between matches. We further show that this formulation can be solved globally by graph cut algorithm. Our new formulation always improves the performance of previous localitybased method without noticeable deterioration in processing time,adding a few milliseconds. Second, to construct a better graph structure, a robust and geometrically meaningful topology-aware relationship is developed to capture the topology relationship between matches. The two components in sum lead to topology interaction matching(TIM), an effective and efficient method for outlier filtering. Extensive experiments on several large and diverse datasets for multiple vision tasks including general feature matching, as well as relative pose estimation, homography and fundamental matrix estimation, loop-closure detection, and multi-modal image matching, demonstrate that our TIM is more competitive than current state-of-the-art methods, in terms of generality, efficiency, and effectiveness. The source code is publicly available at http://github.com/YifanLu2000/TIM.展开更多
In classification problems,datasets often contain a large amount of features,but not all of them are relevant for accurate classification.In fact,irrelevant features may even hinder classification accuracy.Feature sel...In classification problems,datasets often contain a large amount of features,but not all of them are relevant for accurate classification.In fact,irrelevant features may even hinder classification accuracy.Feature selection aims to alleviate this issue by minimizing the number of features in the subset while simultaneously minimizing the classification error rate.Single-objective optimization approaches employ an evaluation function designed as an aggregate function with a parameter,but the results obtained depend on the value of the parameter.To eliminate this parameter’s influence,the problem can be reformulated as a multi-objective optimization problem.The Whale Optimization Algorithm(WOA)is widely used in optimization problems because of its simplicity and easy implementation.In this paper,we propose a multi-strategy assisted multi-objective WOA(MSMOWOA)to address feature selection.To enhance the algorithm’s search ability,we integrate multiple strategies such as Levy flight,Grey Wolf Optimizer,and adaptive mutation into it.Additionally,we utilize an external repository to store non-dominant solution sets and grid technology is used to maintain diversity.Results on fourteen University of California Irvine(UCI)datasets demonstrate that our proposed method effectively removes redundant features and improves classification performance.The source code can be accessed from the website:https://github.com/zc0315/MSMOWOA.展开更多
BACKGROUND Duodenal neuroendocrine tumours(DNETs)are rare neoplasms.However,the incidence of DNETs has been increasing in recent years,especially as an incidental finding during endoscopic studies.Regrettably,there is...BACKGROUND Duodenal neuroendocrine tumours(DNETs)are rare neoplasms.However,the incidence of DNETs has been increasing in recent years,especially as an incidental finding during endoscopic studies.Regrettably,there is no consensus regarding the ideal treatment of DNETs.Even there are few studies on the clinical features and survival analysis of DNETs.AIM To analyze the clinical characteristics and prognostic factors of patients with duodenal neuroendocrine tumours.METHODS The clinical data of DNETs diagnosed in the First Affiliated Hospital of Air Force Military Medical University from June 2011 to July 2022 were collected.Neuroen-docrine tumours located in the ampulla area of the duodenum were divided into the ampullary region group;neuroendocrine tumours in any part of the duo-denum outside the ampullary area were divided into the nonampullary region group.Using a retrospective study,the clinical characteristics of the two groups and risk factors affecting the survival of DNET patients were analysed.RESULTS Twenty-nine DNET patients were screened.The male to female ratio was 1:1.9,and females comprised the majority.The ampullary region group accounted for 24.1%(7/29),while the nonampullary region group accounted for 75.9%(22/29).When diagnosed,the clinical symptoms of the ampullary region group were mainly abdominal pain(85.7%),while those of the nonampullary region groups were mainly abdominal distension(59.1%).There were differences in the composition of staging of tumours between the two groups(Fisher's exact probability method,P=0.001),with nonampullary stage II tumours(68.2%)being the main stage(P<0.05).After the diagnosis of DNETs,the survival rate of the ampullary region group was 14.3%(1/7),which was lower than that of 72.7%(16/22)in the nonampullary region group(Fisher's exact probability method,P=0.011).The survival time of the ampullary region group was shorter than that of the nonampullary region group(P<0.000).The median survival time of the ampullary region group was 10.0 months and that of the nonampullary region group was 451.0 months.Multivariate analysis showed that tumours in the ampulla region and no surgical treatment after diagnosis were independent risk factors for the survival of DNET patients(HR=0.029,95%CI 0.004-0.199,P<0.000;HR=12.609,95%CI:2.889-55.037,P=0.001).Further analysis of nonampullary DNET patients showed that the survival time of patients with a tumour diameter<2 cm was longer than that of patients with a tumour diameter≥2 cm(t=7.243,P=0.048).As of follow-up,6 patients who died of nonampullary DNETs had a tumour diameter that was≥2 cm,and 3 patients in stage IV had liver metastasis.Patients with a tumour diameter<2 cm underwent surgical treatment,and all survived after surgery.CONCLUSION Surgical treatment is a protective factor for prolonging the survival of DNET patients.Compared to DNETs in the ampullary region,patients in the nonampullary region group had a longer survival period.The liver is the organ most susceptible to distant metastasis of nonampullary DNETs.展开更多
At present,super-resolution algorithms are employed to tackle the challenge of low image resolution,but it is difficult to extract differentiated feature details based on various inputs,resulting in poor generalizatio...At present,super-resolution algorithms are employed to tackle the challenge of low image resolution,but it is difficult to extract differentiated feature details based on various inputs,resulting in poor generalization ability.Given this situation,this study first analyzes the features of some feature extraction modules of the current super-resolution algorithm and then proposes an adaptive feature fusion block(AFB)for feature extraction.This module mainly comprises dynamic convolution,attention mechanism,and pixel-based gating mechanism.Combined with dynamic convolution with scale information,the network can extract more differentiated feature information.The introduction of a channel spatial attention mechanism combined with multi-feature fusion further enables the network to retain more important feature information.Dynamic convolution and pixel-based gating mechanisms enhance the module’s adaptability.Finally,a comparative experiment of a super-resolution algorithm based on the AFB module is designed to substantiate the efficiency of the AFB module.The results revealed that the network combined with the AFB module has stronger generalization ability and expression ability.展开更多
The degradation of optical remote sensing images due to atmospheric haze poses a significant obstacle,profoundly impeding their effective utilization across various domains.Dehazing methodologies have emerged as pivot...The degradation of optical remote sensing images due to atmospheric haze poses a significant obstacle,profoundly impeding their effective utilization across various domains.Dehazing methodologies have emerged as pivotal components of image preprocessing,fostering an improvement in the quality of remote sensing imagery.This enhancement renders remote sensing data more indispensable,thereby enhancing the accuracy of target iden-tification.Conventional defogging techniques based on simplistic atmospheric degradation models have proven inadequate for mitigating non-uniform haze within remotely sensed images.In response to this challenge,a novel UNet Residual Attention Network(URA-Net)is proposed.This paradigmatic approach materializes as an end-to-end convolutional neural network distinguished by its utilization of multi-scale dense feature fusion clusters and gated jump connections.The essence of our methodology lies in local feature fusion within dense residual clusters,enabling the extraction of pertinent features from both preceding and current local data,depending on contextual demands.The intelligently orchestrated gated structures facilitate the propagation of these features to the decoder,resulting in superior outcomes in haze removal.Empirical validation through a plethora of experiments substantiates the efficacy of URA-Net,demonstrating its superior performance compared to existing methods when applied to established datasets for remote sensing image defogging.On the RICE-1 dataset,URA-Net achieves a Peak Signal-to-Noise Ratio(PSNR)of 29.07 dB,surpassing the Dark Channel Prior(DCP)by 11.17 dB,the All-in-One Network for Dehazing(AOD)by 7.82 dB,the Optimal Transmission Map and Adaptive Atmospheric Light For Dehazing(OTM-AAL)by 5.37 dB,the Unsupervised Single Image Dehazing(USID)by 8.0 dB,and the Superpixel-based Remote Sensing Image Dehazing(SRD)by 8.5 dB.Particularly noteworthy,on the SateHaze1k dataset,URA-Net attains preeminence in overall performance,yielding defogged images characterized by consistent visual quality.This underscores the contribution of the research to the advancement of remote sensing technology,providing a robust and efficient solution for alleviating the adverse effects of haze on image quality.展开更多
Photoacoustic imaging(PAI)is a noninvasive emerging imaging method based on the photoacoustic effect,which provides necessary assistance for medical diagnosis.It has the characteristics of large imaging depth and high...Photoacoustic imaging(PAI)is a noninvasive emerging imaging method based on the photoacoustic effect,which provides necessary assistance for medical diagnosis.It has the characteristics of large imaging depth and high contrast.However,limited by the equipment cost and reconstruction time requirements,the existing PAI systems distributed with annular array transducers are difficult to take into account both the image quality and the imaging speed.In this paper,a triple-path feature transform network(TFT-Net)for ring-array photoacoustic tomography is proposed to enhance the imaging quality from limited-view and sparse measurement data.Specifically,the network combines the raw photoacoustic pressure signals and conventional linear reconstruction images as input data,and takes the photoacoustic physical model as a prior information to guide the reconstruction process.In addition,to enhance the ability of extracting signal features,the residual block and squeeze and excitation block are introduced into the TFT-Net.For further efficient reconstruction,the final output of photoacoustic signals uses‘filter-then-upsample’operation with a pixel-shuffle multiplexer and a max out module.Experiment results on simulated and in-vivo data demonstrate that the constructed TFT-Net can restore the target boundary clearly,reduce background noise,and realize fast and high-quality photoacoustic image reconstruction of limited view with sparse sampling.展开更多
基金supported in part by the National Natural Science Foundation of China(82072019)the Shenzhen Basic Research Program(JCYJ20210324130209023)+5 种基金the Shenzhen-Hong Kong-Macao S&T Program(Category C)(SGDX20201103095002019)the Mainland-Hong Kong Joint Funding Scheme(MHKJFS)(MHP/005/20),the Project of Strategic Importance Fund(P0035421)the Projects of RISA(P0043001)from the Hong Kong Polytechnic University,the Natural Science Foundation of Jiangsu Province(BK20201441)the Provincial and Ministry Co-constructed Project of Henan Province Medical Science and Technology Research(SBGJ202103038,SBGJ202102056)the Henan Province Key R&D and Promotion Project(Science and Technology Research)(222102310015)the Natural Science Foundation of Henan Province(222300420575),and the Henan Province Science and Technology Research(222102310322).
文摘Modern medicine is reliant on various medical imaging technologies for non-invasively observing patients’anatomy.However,the interpretation of medical images can be highly subjective and dependent on the expertise of clinicians.Moreover,some potentially useful quantitative information in medical images,especially that which is not visible to the naked eye,is often ignored during clinical practice.In contrast,radiomics performs high-throughput feature extraction from medical images,which enables quantitative analysis of medical images and prediction of various clinical endpoints.Studies have reported that radiomics exhibits promising performance in diagnosis and predicting treatment responses and prognosis,demonstrating its potential to be a non-invasive auxiliary tool for personalized medicine.However,radiomics remains in a developmental phase as numerous technical challenges have yet to be solved,especially in feature engineering and statistical modeling.In this review,we introduce the current utility of radiomics by summarizing research on its application in the diagnosis,prognosis,and prediction of treatment responses in patients with cancer.We focus on machine learning approaches,for feature extraction and selection during feature engineering and for imbalanced datasets and multi-modality fusion during statistical modeling.Furthermore,we introduce the stability,reproducibility,and interpretability of features,and the generalizability and interpretability of models.Finally,we offer possible solutions to current challenges in radiomics research.
基金Supported by the 74th General Support of China Postdoctoral Science Foundation,No.2023M740675the National Natural Science Foundation of China,No.82170555+2 种基金Shanghai Academic/Technology Research Leader,No.22XD1422400Shuguang Program of Shanghai Education Development Foundation and Shanghai Municipal Education Commission,No.2022SG06Shanghai"Rising Stars of Medical Talent"Youth Development Program,No.20224Z0005.
文摘BACKGROUND Gastric cystica profunda(GCP)represents a rare condition characterized by cystic dilation of gastric glands within the mucosal and/or submucosal layers.GCP is often linked to,or may progress into,early gastric cancer(EGC).AIM To provide a comprehensive evaluation of the endoscopic features of GCP while assessing the efficacy of endoscopic treatment,thereby offering guidance for diagnosis and treatment.METHODS This retrospective study involved 104 patients with GCP who underwent endoscopic resection.Alongside demographic and clinical data,regular patient followups were conducted to assess local recurrence.RESULTS Among the 104 patients diagnosed with GCP who underwent endoscopic resection,12.5%had a history of previous gastric procedures.The primary site predominantly affected was the cardia(38.5%,n=40).GCP commonly exhibited intraluminal growth(99%),regular presentation(74.0%),and ulcerative mucosa(61.5%).The leading endoscopic feature was the mucosal lesion type(59.6%,n=62).The average maximum diameter was 20.9±15.3 mm,with mucosal involvement in 60.6%(n=63).Procedures lasted 73.9±57.5 min,achieving complete resection in 91.3%(n=95).Recurrence(4.8%)was managed via either surgical intervention(n=1)or through endoscopic resection(n=4).Final pathology confirmed that 59.6%of GCP cases were associated with EGC.Univariate analysis indicated that elderly males were more susceptible to GCP associated with EGC.Conversely,multivariate analysis identified lesion morphology and endoscopic features as significant risk factors.Survival analysis demonstrated no statistically significant difference in recurrence between GCP with and without EGC(P=0.72).CONCLUSION The findings suggested that endoscopic resection might serve as an effective and minimally invasive treatment for GCP with or without EGC.
基金supported by the National Key Research and Development Project,No.2019YFA0112100(to SF).
文摘Traumatic spinal cord injury is potentially catastrophic and can lead to permanent disability or even death.China has the largest population of patients with traumatic spinal cord injury.Previous studies of traumatic spinal cord injury in China have mostly been regional in scope;national-level studies have been rare.To the best of our knowledge,no national-level study of treatment status and economic burden has been performed.This retrospective study aimed to examine the epidemiological and clinical features,treatment status,and economic burden of traumatic spinal cord injury in China at the national level.We included 13,465 traumatic spinal cord injury patients who were injured between January 2013 and December 2018 and treated in 30 hospitals in 11 provinces/municipalities representing all geographical divisions of China.Patient epidemiological and clinical features,treatment status,and total and daily costs were recorded.Trends in the percentage of traumatic spinal cord injuries among all hospitalized patients and among patients hospitalized in the orthopedic department and cost of care were assessed by annual percentage change using the Joinpoint Regression Program.The percentage of traumatic spinal cord injuries among all hospitalized patients and among patients hospitalized in the orthopedic department did not significantly change overall(annual percentage change,-0.5%and 2.1%,respectively).A total of 10,053(74.7%)patients underwent surgery.Only 2.8%of patients who underwent surgery did so within 24 hours of injury.A total of 2005(14.9%)patients were treated with high-dose(≥500 mg)methylprednisolone sodium succinate/methylprednisolone(MPSS/MP);615(4.6%)received it within 8 hours.The total cost for acute traumatic spinal cord injury decreased over the study period(-4.7%),while daily cost did not significantly change(1.0%increase).Our findings indicate that public health initiatives should aim at improving hospitals’ability to complete early surgery within 24 hours,which is associated with improved sensorimotor recovery,increasing the awareness rate of clinical guidelines related to high-dose MPSS/MP to reduce the use of the treatment with insufficient evidence.
基金National Natural Science Foundation of China(Nos.42071444,42101444)。
文摘Cultural relics line graphic serves as a crucial form of traditional artifact information documentation,which is a simple and intuitive product with low cost of displaying compared with 3D models.Dimensionality reduction is undoubtedly necessary for line drawings.However,most existing methods for artifact drawing rely on the principles of orthographic projection that always cannot avoid angle occlusion and data overlapping while the surface of cultural relics is complex.Therefore,conformal mapping was introduced as a dimensionality reduction way to compensate for the limitation of orthographic projection.Based on the given criteria for assessing surface complexity,this paper proposed a three-dimensional feature guideline extraction method for complex cultural relic surfaces.A 2D and 3D combined factor that measured the importance of points on describing surface features,vertex weight,was designed.Then the selection threshold for feature guideline extraction was determined based on the differences between vertex weight and shape index distributions.The feasibility and stability were verified through experiments conducted on real cultural relic surface data.Results demonstrated the ability of the method to address the challenges associated with the automatic generation of line drawings for complex surfaces.The extraction method and the obtained results will be useful for line graphic drawing,displaying and propaganda of cultural relics.
基金supported by the Competitive Research Fund of the University of Aizu,Japan.
文摘Sign language recognition is vital for enhancing communication accessibility among the Deaf and hard-of-hearing communities.In Japan,approximately 360,000 individualswith hearing and speech disabilities rely on Japanese Sign Language(JSL)for communication.However,existing JSL recognition systems have faced significant performance limitations due to inherent complexities.In response to these challenges,we present a novel JSL recognition system that employs a strategic fusion approach,combining joint skeleton-based handcrafted features and pixel-based deep learning features.Our system incorporates two distinct streams:the first stream extracts crucial handcrafted features,emphasizing the capture of hand and body movements within JSL gestures.Simultaneously,a deep learning-based transfer learning stream captures hierarchical representations of JSL gestures in the second stream.Then,we concatenated the critical information of the first stream and the hierarchy of the second stream features to produce the multiple levels of the fusion features,aiming to create a comprehensive representation of the JSL gestures.After reducing the dimensionality of the feature,a feature selection approach and a kernel-based support vector machine(SVM)were used for the classification.To assess the effectiveness of our approach,we conducted extensive experiments on our Lab JSL dataset and a publicly available Arabic sign language(ArSL)dataset.Our results unequivocally demonstrate that our fusion approach significantly enhances JSL recognition accuracy and robustness compared to individual feature sets or traditional recognition methods.
基金supported in part by the National Natural Science Foundation of China(Grants 62376172,62006163,62376043)in part by the National Postdoctoral Program for Innovative Talents(Grant BX20200226)in part by Sichuan Science and Technology Planning Project(Grants 2022YFSY0047,2022YFQ0014,2023ZYD0143,2022YFH0021,2023YFQ0020,24QYCX0354,24NSFTD0025).
文摘Time series anomaly detection is crucial in various industrial applications to identify unusual behaviors within the time series data.Due to the challenges associated with annotating anomaly events,time series reconstruction has become a prevalent approach for unsupervised anomaly detection.However,effectively learning representations and achieving accurate detection results remain challenging due to the intricate temporal patterns and dependencies in real-world time series.In this paper,we propose a cross-dimension attentive feature fusion network for time series anomaly detection,referred to as CAFFN.Specifically,a series and feature mixing block is introduced to learn representations in 1D space.Additionally,a fast Fourier transform is employed to convert the time series into 2D space,providing the capability for 2D feature extraction.Finally,a cross-dimension attentive feature fusion mechanism is designed that adaptively integrates features across different dimensions for anomaly detection.Experimental results on real-world time series datasets demonstrate that CAFFN performs better than other competing methods in time series anomaly detection.
基金a grant from the National Natural Science Foundation of China(Nos.11905239,12005248 and 12105303).
文摘With the rapid development of the mobile communication and the Internet,the previous web anomaly detectionand identificationmodels were built relying on security experts’empirical knowledge and attack features.Althoughthis approach can achieve higher detection performance,it requires huge human labor and resources to maintainthe feature library.In contrast,semantic feature engineering can dynamically discover new semantic featuresand optimize feature selection by automatically analyzing the semantic information contained in the data itself,thus reducing dependence on prior knowledge.However,current semantic features still have the problem ofsemantic expression singularity,as they are extracted from a single semantic mode such as word segmentation,character segmentation,or arbitrary semantic feature extraction.This paper extracts features of web requestsfrom dual semantic granularity,and proposes a semantic feature fusion method to solve the above problems.Themethod first preprocesses web requests,and extracts word-level and character-level semantic features of URLs viaconvolutional neural network(CNN),respectively.By constructing three loss functions to reduce losses betweenfeatures,labels and categories.Experiments on the HTTP CSIC 2010,Malicious URLs and HttpParams datasetsverify the proposedmethod.Results show that compared withmachine learning,deep learningmethods and BERTmodel,the proposed method has better detection performance.And it achieved the best detection rate of 99.16%in the dataset HttpParams.
基金National Natural Science Foundation of China(No.62271186)Anhui Key Project of Research and Development Plan(No.202104d07020005)。
文摘Deception detection plays a crucial role in criminal investigation.Videos contain a wealth of information regarding apparent and physiological changes in individuals,and thus can serve as an effective means of deception detection.In this paper,we investigate video-based deception detection considering both apparent visual features such as eye gaze,head pose and facial action unit(AU),and non-contact heart rate detected by remote photoplethysmography(rPPG)technique.Multiple wrapper-based feature selection methods combined with the K-nearest neighbor(KNN)and support vector machine(SVM)classifiers are employed to screen the most effective features for deception detection.We evaluate the performance of the proposed method on both a self-collected physiological-assisted visual deception detection(PV3D)dataset and a public bag-oflies(BOL)dataset.Experimental results demonstrate that the SVM classifier with symbiotic organisms search(SOS)feature selection yields the best overall performance,with an area under the curve(AUC)of 83.27%and accuracy(ACC)of 83.33%for PV3D,and an AUC of 71.18%and ACC of 70.33%for BOL.This demonstrates the stability and effectiveness of the proposed method in video-based deception detection tasks.
基金Major Project for New Generation of AI Grant No.2018AAA0100400)the Scientific Research Fund of Hunan Provincial Education Department,China(Grant Nos.21A0350,21C0439,22A0408,22A0414,2022JJ30231,22B0559)the National Natural Science Foundation of Hunan Province,China(Grant No.2022JJ50051).
文摘Radiomics is a non-invasive method for extracting quantitative and higher-dimensional features from medical images for diagnosis.It has received great attention due to its huge application prospects in recent years.We can know that the number of features selected by the existing radiomics feature selectionmethods is basically about ten.In this paper,a heuristic feature selection method based on frequency iteration and multiple supervised training mode is proposed.Based on the combination between features,it decomposes all features layer by layer to select the optimal features for each layer,then fuses the optimal features to form a local optimal group layer by layer and iterates to the global optimal combination finally.Compared with the currentmethod with the best prediction performance in the three data sets,thismethod proposed in this paper can reduce the number of features fromabout ten to about three without losing classification accuracy and even significantly improving classification accuracy.The proposed method has better interpretability and generalization ability,which gives it great potential in the feature selection of radiomics.
基金Australian Research Council,Grant/Award Numbers:DP190103660,DP200103207,LP180100663UniSQ Capacity Building Grants,Grant/Award Number:1008313。
文摘Biometric recognition is a widely used technology for user authentication.In the application of this technology,biometric security and recognition accuracy are two important issues that should be considered.In terms of biometric security,cancellable biometrics is an effective technique for protecting biometric data.Regarding recognition accuracy,feature representation plays a significant role in the performance and reliability of cancellable biometric systems.How to design good feature representations for cancellable biometrics is a challenging topic that has attracted a great deal of attention from the computer vision community,especially from researchers of cancellable biometrics.Feature extraction and learning in cancellable biometrics is to find suitable feature representations with a view to achieving satisfactory recognition performance,while the privacy of biometric data is protected.This survey informs the progress,trend and challenges of feature extraction and learning for cancellable biometrics,thus shedding light on the latest developments and future research of this area.
基金the National Social Science Foundation of China(Grant No.22BTJ035).
文摘The classification of functional data has drawn much attention in recent years.The main challenge is representing infinite-dimensional functional data by finite-dimensional features while utilizing those features to achieve better classification accuracy.In this paper,we propose a mean-variance-based(MV)feature weighting method for classifying functional data or functional curves.In the feature extraction stage,each sample curve is approximated by B-splines to transfer features to the coefficients of the spline basis.After that,a feature weighting approach based on statistical principles is introduced by comprehensively considering the between-class differences and within-class variations of the coefficients.We also introduce a scaling parameter to adjust the gap between the weights of features.The new feature weighting approach can adaptively enhance noteworthy local features while mitigating the impact of confusing features.The algorithms for feature weighted K-nearest neighbor and support vector machine classifiers are both provided.Moreover,the new approach can be well integrated into existing functional data classifiers,such as the generalized functional linear model and functional linear discriminant analysis,resulting in a more accurate classification.The performance of the mean-variance-based classifiers is evaluated by simulation studies and real data.The results show that the newfeatureweighting approach significantly improves the classification accuracy for complex functional data.
基金supported in part by the National Natural Science Foundation of China(Grant No.61971078)Chongqing Education Commission Science and Technology Major Project(No.KJZD-M202301901).
文摘While single-modal visible light images or infrared images provide limited information,infrared light captures significant thermal radiation data,whereas visible light excels in presenting detailed texture information.Com-bining images obtained from both modalities allows for leveraging their respective strengths and mitigating individual limitations,resulting in high-quality images with enhanced contrast and rich texture details.Such capabilities hold promising applications in advanced visual tasks including target detection,instance segmentation,military surveillance,pedestrian detection,among others.This paper introduces a novel approach,a dual-branch decomposition fusion network based on AutoEncoder(AE),which decomposes multi-modal features into intensity and texture information for enhanced fusion.Local contrast enhancement module(CEM)and texture detail enhancement module(DEM)are devised to process the decomposed images,followed by image fusion through the decoder.The proposed loss function ensures effective retention of key information from the source images of both modalities.Extensive comparisons and generalization experiments demonstrate the superior performance of our network in preserving pixel intensity distribution and retaining texture details.From the qualitative results,we can see the advantages of fusion details and local contrast.In the quantitative experiments,entropy(EN),mutual information(MI),structural similarity(SSIM)and other results have improved and exceeded the SOTA(State of the Art)model as a whole.
文摘In the era of the Internet,widely used web applications have become the target of hacker attacks because they contain a large amount of personal information.Among these vulnerabilities,stealing private data through crosssite scripting(XSS)attacks is one of the most commonly used attacks by hackers.Currently,deep learning-based XSS attack detection methods have good application prospects;however,they suffer from problems such as being prone to overfitting,a high false alarm rate,and low accuracy.To address these issues,we propose a multi-stage feature extraction and fusion model for XSS detection based on Random Forest feature enhancement.The model utilizes RandomForests to capture the intrinsic structure and patterns of the data by extracting leaf node indices as features,which are subsequentlymergedwith the original data features to forma feature setwith richer information content.Further feature extraction is conducted through three parallel channels.Channel I utilizes parallel onedimensional convolutional layers(1Dconvolutional layers)with different convolutional kernel sizes to extract local features at different scales and performmulti-scale feature fusion;Channel II employsmaximum one-dimensional pooling layers(max 1D pooling layers)of various sizes to extract key features from the data;and Channel III extracts global information bi-directionally using a Bi-Directional Long-Short TermMemory Network(Bi-LSTM)and incorporates a multi-head attention mechanism to enhance global features.Finally,effective classification and prediction of XSS are performed by fusing the features of the three channels.To test the effectiveness of the model,we conduct experiments on six datasets.We achieve an accuracy of 100%on the UNSW-NB15 dataset and 99.99%on the CICIDS2017 dataset,which is higher than that of the existing models.
基金supported by Wuyi University Hong Kong and Macao Joint Research and Development Fund(GrantsNos.2021WGALH15,2019WGALH17,2019WGALH15)the National Natural Science Foundation of China-Guangdong Joint Fund(GrantsNo.2019A1515111052)+2 种基金the National Natural Science Foundation of China(Grant No.52202426)a grant from the Research Grants Council(RGC)of the Hong Kong Special Administrative Region(SAR),China(Grants No.15205723)a grant from the Hong Kong Polytechnic University(Grant No.P0045325).
文摘The safety and stability of high-speed maglev trains traveling on viaducts in crosswinds critically depend on their aerodynamic characteristics.Therefore,this paper uses an improved delayed detached eddy simulation(IDDES)method to investigate the aerodynamic features of high-speed maglev trains with different marshaling lengths under crosswinds.The effects of marshaling lengths(varying from 3-car to 8-car groups)on the train’s aerodynamic performance,surface pressure,and the flow field surrounding the train were investigated using the three-dimensional unsteady compressible Navier-Stokes(N-S)equations.The results showed that the marshaling lengths had minimal influence on the aerodynamic performance of the head and middle cars.Conversely,the marshaling lengths are negatively correlated with the time-average side force coefficient(CS)and time-average lift force coefficient(Cl)of the tail car.Compared to the tail car of the 3-car groups,the CS and Cl fell by 27.77%and 18.29%,respectively,for the tail car of the 8-car groups.It is essential to pay more attention to the operational safety of the head car,as it exhibits the highest time average CS.Additionally,the mean pressure difference between the two sides of the tail car body increased with the marshaling lengths,and the side force direction on the tail car was opposite to that of the head and middle cars.Furthermore,the turbulent kinetic energy of the wake structure on the windward side quickly decreased as marshaling lengths increased.
基金supported by the National Natural Science Foundation of China (62276192)。
文摘Feature matching plays a key role in computer vision. However, due to the limitations of the descriptors, the putative matches are inevitably contaminated by massive outliers.This paper attempts to tackle the outlier filtering problem from two aspects. First, a robust and efficient graph interaction model,is proposed, with the assumption that matches are correlated with each other rather than independently distributed. To this end, we construct a graph based on the local relationships of matches and formulate the outlier filtering task as a binary labeling energy minimization problem, where the pairwise term encodes the interaction between matches. We further show that this formulation can be solved globally by graph cut algorithm. Our new formulation always improves the performance of previous localitybased method without noticeable deterioration in processing time,adding a few milliseconds. Second, to construct a better graph structure, a robust and geometrically meaningful topology-aware relationship is developed to capture the topology relationship between matches. The two components in sum lead to topology interaction matching(TIM), an effective and efficient method for outlier filtering. Extensive experiments on several large and diverse datasets for multiple vision tasks including general feature matching, as well as relative pose estimation, homography and fundamental matrix estimation, loop-closure detection, and multi-modal image matching, demonstrate that our TIM is more competitive than current state-of-the-art methods, in terms of generality, efficiency, and effectiveness. The source code is publicly available at http://github.com/YifanLu2000/TIM.
基金supported in part by the Natural Science Youth Foundation of Hebei Province under Grant F2019403207in part by the PhD Research Startup Foundation of Hebei GEO University under Grant BQ2019055+3 种基金in part by the Open Research Project of the Hubei Key Laboratory of Intelligent Geo-Information Processing under Grant KLIGIP-2021A06in part by the Fundamental Research Funds for the Universities in Hebei Province under Grant QN202220in part by the Science and Technology Research Project for Universities of Hebei under Grant ZD2020344in part by the Guangxi Natural Science Fund General Project under Grant 2021GXNSFAA075029.
文摘In classification problems,datasets often contain a large amount of features,but not all of them are relevant for accurate classification.In fact,irrelevant features may even hinder classification accuracy.Feature selection aims to alleviate this issue by minimizing the number of features in the subset while simultaneously minimizing the classification error rate.Single-objective optimization approaches employ an evaluation function designed as an aggregate function with a parameter,but the results obtained depend on the value of the parameter.To eliminate this parameter’s influence,the problem can be reformulated as a multi-objective optimization problem.The Whale Optimization Algorithm(WOA)is widely used in optimization problems because of its simplicity and easy implementation.In this paper,we propose a multi-strategy assisted multi-objective WOA(MSMOWOA)to address feature selection.To enhance the algorithm’s search ability,we integrate multiple strategies such as Levy flight,Grey Wolf Optimizer,and adaptive mutation into it.Additionally,we utilize an external repository to store non-dominant solution sets and grid technology is used to maintain diversity.Results on fourteen University of California Irvine(UCI)datasets demonstrate that our proposed method effectively removes redundant features and improves classification performance.The source code can be accessed from the website:https://github.com/zc0315/MSMOWOA.
基金The study protocol was approved by the Clinical Research Ethics Committee of Honghui Hospital,Xi’an Jiaotong University(No.202401004).
文摘BACKGROUND Duodenal neuroendocrine tumours(DNETs)are rare neoplasms.However,the incidence of DNETs has been increasing in recent years,especially as an incidental finding during endoscopic studies.Regrettably,there is no consensus regarding the ideal treatment of DNETs.Even there are few studies on the clinical features and survival analysis of DNETs.AIM To analyze the clinical characteristics and prognostic factors of patients with duodenal neuroendocrine tumours.METHODS The clinical data of DNETs diagnosed in the First Affiliated Hospital of Air Force Military Medical University from June 2011 to July 2022 were collected.Neuroen-docrine tumours located in the ampulla area of the duodenum were divided into the ampullary region group;neuroendocrine tumours in any part of the duo-denum outside the ampullary area were divided into the nonampullary region group.Using a retrospective study,the clinical characteristics of the two groups and risk factors affecting the survival of DNET patients were analysed.RESULTS Twenty-nine DNET patients were screened.The male to female ratio was 1:1.9,and females comprised the majority.The ampullary region group accounted for 24.1%(7/29),while the nonampullary region group accounted for 75.9%(22/29).When diagnosed,the clinical symptoms of the ampullary region group were mainly abdominal pain(85.7%),while those of the nonampullary region groups were mainly abdominal distension(59.1%).There were differences in the composition of staging of tumours between the two groups(Fisher's exact probability method,P=0.001),with nonampullary stage II tumours(68.2%)being the main stage(P<0.05).After the diagnosis of DNETs,the survival rate of the ampullary region group was 14.3%(1/7),which was lower than that of 72.7%(16/22)in the nonampullary region group(Fisher's exact probability method,P=0.011).The survival time of the ampullary region group was shorter than that of the nonampullary region group(P<0.000).The median survival time of the ampullary region group was 10.0 months and that of the nonampullary region group was 451.0 months.Multivariate analysis showed that tumours in the ampulla region and no surgical treatment after diagnosis were independent risk factors for the survival of DNET patients(HR=0.029,95%CI 0.004-0.199,P<0.000;HR=12.609,95%CI:2.889-55.037,P=0.001).Further analysis of nonampullary DNET patients showed that the survival time of patients with a tumour diameter<2 cm was longer than that of patients with a tumour diameter≥2 cm(t=7.243,P=0.048).As of follow-up,6 patients who died of nonampullary DNETs had a tumour diameter that was≥2 cm,and 3 patients in stage IV had liver metastasis.Patients with a tumour diameter<2 cm underwent surgical treatment,and all survived after surgery.CONCLUSION Surgical treatment is a protective factor for prolonging the survival of DNET patients.Compared to DNETs in the ampullary region,patients in the nonampullary region group had a longer survival period.The liver is the organ most susceptible to distant metastasis of nonampullary DNETs.
基金Supported by Sichuan Science and Technology Program(2021YFQ0003,2023YFSY0026,2023YFH0004).
文摘At present,super-resolution algorithms are employed to tackle the challenge of low image resolution,but it is difficult to extract differentiated feature details based on various inputs,resulting in poor generalization ability.Given this situation,this study first analyzes the features of some feature extraction modules of the current super-resolution algorithm and then proposes an adaptive feature fusion block(AFB)for feature extraction.This module mainly comprises dynamic convolution,attention mechanism,and pixel-based gating mechanism.Combined with dynamic convolution with scale information,the network can extract more differentiated feature information.The introduction of a channel spatial attention mechanism combined with multi-feature fusion further enables the network to retain more important feature information.Dynamic convolution and pixel-based gating mechanisms enhance the module’s adaptability.Finally,a comparative experiment of a super-resolution algorithm based on the AFB module is designed to substantiate the efficiency of the AFB module.The results revealed that the network combined with the AFB module has stronger generalization ability and expression ability.
基金This project is supported by the National Natural Science Foundation of China(NSFC)(No.61902158).
文摘The degradation of optical remote sensing images due to atmospheric haze poses a significant obstacle,profoundly impeding their effective utilization across various domains.Dehazing methodologies have emerged as pivotal components of image preprocessing,fostering an improvement in the quality of remote sensing imagery.This enhancement renders remote sensing data more indispensable,thereby enhancing the accuracy of target iden-tification.Conventional defogging techniques based on simplistic atmospheric degradation models have proven inadequate for mitigating non-uniform haze within remotely sensed images.In response to this challenge,a novel UNet Residual Attention Network(URA-Net)is proposed.This paradigmatic approach materializes as an end-to-end convolutional neural network distinguished by its utilization of multi-scale dense feature fusion clusters and gated jump connections.The essence of our methodology lies in local feature fusion within dense residual clusters,enabling the extraction of pertinent features from both preceding and current local data,depending on contextual demands.The intelligently orchestrated gated structures facilitate the propagation of these features to the decoder,resulting in superior outcomes in haze removal.Empirical validation through a plethora of experiments substantiates the efficacy of URA-Net,demonstrating its superior performance compared to existing methods when applied to established datasets for remote sensing image defogging.On the RICE-1 dataset,URA-Net achieves a Peak Signal-to-Noise Ratio(PSNR)of 29.07 dB,surpassing the Dark Channel Prior(DCP)by 11.17 dB,the All-in-One Network for Dehazing(AOD)by 7.82 dB,the Optimal Transmission Map and Adaptive Atmospheric Light For Dehazing(OTM-AAL)by 5.37 dB,the Unsupervised Single Image Dehazing(USID)by 8.0 dB,and the Superpixel-based Remote Sensing Image Dehazing(SRD)by 8.5 dB.Particularly noteworthy,on the SateHaze1k dataset,URA-Net attains preeminence in overall performance,yielding defogged images characterized by consistent visual quality.This underscores the contribution of the research to the advancement of remote sensing technology,providing a robust and efficient solution for alleviating the adverse effects of haze on image quality.
基金supported by National Key R&D Program of China[2022YFC2402400]the National Natural Science Foundation of China[Grant No.62275062]Guangdong Provincial Key Laboratory of Biomedical Optical Imaging Technology[Grant No.2020B121201010-4].
文摘Photoacoustic imaging(PAI)is a noninvasive emerging imaging method based on the photoacoustic effect,which provides necessary assistance for medical diagnosis.It has the characteristics of large imaging depth and high contrast.However,limited by the equipment cost and reconstruction time requirements,the existing PAI systems distributed with annular array transducers are difficult to take into account both the image quality and the imaging speed.In this paper,a triple-path feature transform network(TFT-Net)for ring-array photoacoustic tomography is proposed to enhance the imaging quality from limited-view and sparse measurement data.Specifically,the network combines the raw photoacoustic pressure signals and conventional linear reconstruction images as input data,and takes the photoacoustic physical model as a prior information to guide the reconstruction process.In addition,to enhance the ability of extracting signal features,the residual block and squeeze and excitation block are introduced into the TFT-Net.For further efficient reconstruction,the final output of photoacoustic signals uses‘filter-then-upsample’operation with a pixel-shuffle multiplexer and a max out module.Experiment results on simulated and in-vivo data demonstrate that the constructed TFT-Net can restore the target boundary clearly,reduce background noise,and realize fast and high-quality photoacoustic image reconstruction of limited view with sparse sampling.