Pattern matching method is one of the classic classifications of existing online portfolio selection strategies. This article aims to study the key aspects of this method—measurement of similarity and selection of si...Pattern matching method is one of the classic classifications of existing online portfolio selection strategies. This article aims to study the key aspects of this method—measurement of similarity and selection of similarity sets, and proposes a Portfolio Selection Method based on Pattern Matching with Dual Information of Direction and Distance (PMDI). By studying different combination methods of indicators such as Euclidean distance, Chebyshev distance, and correlation coefficient, important information such as direction and distance in stock historical price information is extracted, thereby filtering out the similarity set required for pattern matching based investment portfolio selection algorithms. A large number of experiments conducted on two datasets of real stock markets have shown that PMDI outperforms other algorithms in balancing income and risk. Therefore, it is suitable for the financial environment in the real world.展开更多
Feature matching plays a key role in computer vision. However, due to the limitations of the descriptors, the putative matches are inevitably contaminated by massive outliers.This paper attempts to tackle the outlier ...Feature matching plays a key role in computer vision. However, due to the limitations of the descriptors, the putative matches are inevitably contaminated by massive outliers.This paper attempts to tackle the outlier filtering problem from two aspects. First, a robust and efficient graph interaction model,is proposed, with the assumption that matches are correlated with each other rather than independently distributed. To this end, we construct a graph based on the local relationships of matches and formulate the outlier filtering task as a binary labeling energy minimization problem, where the pairwise term encodes the interaction between matches. We further show that this formulation can be solved globally by graph cut algorithm. Our new formulation always improves the performance of previous localitybased method without noticeable deterioration in processing time,adding a few milliseconds. Second, to construct a better graph structure, a robust and geometrically meaningful topology-aware relationship is developed to capture the topology relationship between matches. The two components in sum lead to topology interaction matching(TIM), an effective and efficient method for outlier filtering. Extensive experiments on several large and diverse datasets for multiple vision tasks including general feature matching, as well as relative pose estimation, homography and fundamental matrix estimation, loop-closure detection, and multi-modal image matching, demonstrate that our TIM is more competitive than current state-of-the-art methods, in terms of generality, efficiency, and effectiveness. The source code is publicly available at http://github.com/YifanLu2000/TIM.展开更多
The paper develops a multiple matching attenuation method based on extended filtering in the curvelet domain,which combines the traditional Wiener filtering method with the matching attenuation method in curvelet doma...The paper develops a multiple matching attenuation method based on extended filtering in the curvelet domain,which combines the traditional Wiener filtering method with the matching attenuation method in curvelet domain.Firstly,the method uses the predicted multiple data to generate the Hilbert transform records,time derivative records and time derivative records of Hilbert transform.Then,the above records are transformed into the curvelet domain and multiple matching attenuation based on least squares extended filtering is performed.Finally,the attenuation results are transformed back into the time-space domain.Tests on the model data and field data show that the method proposed in the paper effectively suppress the multiples while preserving the primaries well.Furthermore,it has higher accuracy in eliminating multiple reflections,which is more suitable for the multiple attenuation tasks in the areas with complex structures compared to the time-space domain extended filtering method and the conventional curvelet transform method.展开更多
For accurately identifying the distribution charac-teristic of Gaussian-like noises in unmanned aerial vehicle(UAV)state estimation,this paper proposes a non-parametric scheme based on curve similarity matching.In the...For accurately identifying the distribution charac-teristic of Gaussian-like noises in unmanned aerial vehicle(UAV)state estimation,this paper proposes a non-parametric scheme based on curve similarity matching.In the framework of the pro-posed scheme,a Parzen window(kernel density estimation,KDE)method on sliding window technology is applied for roughly esti-mating the sample probability density,a precise data probability density function(PDF)model is constructed with the least square method on K-fold cross validation,and the testing result based on evaluation method is obtained based on some data characteristic analyses of curve shape,abruptness and symmetry.Some com-parison simulations with classical methods and UAV flight exper-iment shows that the proposed scheme has higher recognition accuracy than classical methods for some kinds of Gaussian-like data,which provides better reference for the design of Kalman filter(KF)in complex water environment.展开更多
Many efforts have been devoted to efficient task scheduling in Multi-Unmanned Aerial Vehicle(UAV)edge computing.However,the heterogeneity of UAV computation resource,and the task re-allocating between UAVs have not be...Many efforts have been devoted to efficient task scheduling in Multi-Unmanned Aerial Vehicle(UAV)edge computing.However,the heterogeneity of UAV computation resource,and the task re-allocating between UAVs have not been fully considered yet.Moreover,most existing works neglect the fact that a task can only be executed on the UAV equipped with its desired service function(SF).In this backdrop,this paper formulates the task scheduling problem as a multi-objective task scheduling problem,which aims at maximizing the task execution success ratio while minimizing the average weighted sum of all tasks’completion time and energy consumption.Optimizing three coupled goals in a realtime manner with the dynamic arrival of tasks hinders us from adopting existing methods,like machine learning-based solutions that require a long training time and tremendous pre-knowledge about the task arrival process,or heuristic-based ones that usually incur a long decision-making time.To tackle this problem in a distributed manner,we establish a matching theory framework,in which three conflicting goals are treated as the preferences of tasks,SFs and UAVs.Then,a Distributed Matching Theory-based Re-allocating(DiMaToRe)algorithm is put forward.We formally proved that a stable matching can be achieved by our proposal.Extensive simulation results show that Di Ma To Re algorithm outperforms benchmark algorithms under diverse parameter settings and has good robustness.展开更多
Accurate forecasting of time series is crucial across various domains.Many prediction tasks rely on effectively segmenting,matching,and time series data alignment.For instance,regardless of time series with the same g...Accurate forecasting of time series is crucial across various domains.Many prediction tasks rely on effectively segmenting,matching,and time series data alignment.For instance,regardless of time series with the same granularity,segmenting them into different granularity events can effectively mitigate the impact of varying time scales on prediction accuracy.However,these events of varying granularity frequently intersect with each other,which may possess unequal durations.Even minor differences can result in significant errors when matching time series with future trends.Besides,directly using matched events but unaligned events as state vectors in machine learning-based prediction models can lead to insufficient prediction accuracy.Therefore,this paper proposes a short-term forecasting method for time series based on a multi-granularity event,MGE-SP(multi-granularity event-based short-termprediction).First,amethodological framework for MGE-SP established guides the implementation steps.The framework consists of three key steps,including multi-granularity event matching based on the LTF(latest time first)strategy,multi-granularity event alignment using a piecewise aggregate approximation based on the compression ratio,and a short-term prediction model based on XGBoost.The data from a nationwide online car-hailing service in China ensures the method’s reliability.The average RMSE(root mean square error)and MAE(mean absolute error)of the proposed method are 3.204 and 2.360,lower than the respective values of 4.056 and 3.101 obtained using theARIMA(autoregressive integratedmoving average)method,as well as the values of 4.278 and 2.994 obtained using k-means-SVR(support vector regression)method.The other experiment is conducted on stock data froma public data set.The proposed method achieved an average RMSE and MAE of 0.836 and 0.696,lower than the respective values of 1.019 and 0.844 obtained using the ARIMA method,as well as the values of 1.350 and 1.172 obtained using the k-means-SVR method.展开更多
BACKGROUND:Patients with diabetes mellitus(DM)are vulnerable to community-acquired pneumonia(CAP),which have a high mortality rate.We aimed to investigate the value of heparin-binding protein(HBP)as a prognostic marke...BACKGROUND:Patients with diabetes mellitus(DM)are vulnerable to community-acquired pneumonia(CAP),which have a high mortality rate.We aimed to investigate the value of heparin-binding protein(HBP)as a prognostic marker of mortality in patients with DM and CAP.METHODS:This retrospective study included CAP patients who were tested for HBP at intensive care unit(ICU)admission from January 2019 to April 2020.Patients were allocated to the DM or non-DM group and paired with propensity score matching.Baseline characteristics and clinical outcomes up to 90 days were evaluated.The primary outcome was the 10-day mortality.Receiver operating characteristic(ROC)curves,Kaplan-Meier analysis,and Cox regression were used for statistical analysis.RESULTS:Among 152 enrolled patients,60 pairs were successfully matched.There was no significant difference in 10-day mortality,while more patients in the DM group died within 28 d(P=0.024)and 90 d(P=0.008).In the DM group,HBP levels at ICU admission were higher in 10-day non-survivors than in 10-day survivors(median 182.21[IQR:55.43-300]ng/ml vs.median 66.40[IQR:34.13-107.85]ng/mL,P=0.019),and HBP levels could predict the 10-day mortality with an area under the ROC curve of 0.747.The cut-off value,sensitivity,and specificity were 160.6 ng/mL,66.7%,and 90.2%,respectively.Multivariate Cox regression analysis indicated that HBP was an independent prognostic factor for 10-day(HR 7.196,95%CI:1.596-32.455,P=0.01),28-day(HR 4.381,95%CI:1.449-13.245,P=0.009),and 90-day mortality(HR 4.581,95%CI:1.637-12.819,P=0.004)in patients with DM.CONCLUSION:Plasma HBP at ICU admission was associated with the 10-day,28-day,and 90-day mortality,and might be a prognostic factor in patients with DM and CAP.展开更多
In minimally invasive surgery,endoscopes or laparoscopes equipped with miniature cameras and tools are used to enter the human body for therapeutic purposes through small incisions or natural cavities.However,in clini...In minimally invasive surgery,endoscopes or laparoscopes equipped with miniature cameras and tools are used to enter the human body for therapeutic purposes through small incisions or natural cavities.However,in clinical operating environments,endoscopic images often suffer from challenges such as low texture,uneven illumination,and non-rigid structures,which affect feature observation and extraction.This can severely impact surgical navigation or clinical diagnosis due to missing feature points in endoscopic images,leading to treatment and postoperative recovery issues for patients.To address these challenges,this paper introduces,for the first time,a Cross-Channel Multi-Modal Adaptive Spatial Feature Fusion(ASFF)module based on the lightweight architecture of EfficientViT.Additionally,a novel lightweight feature extraction and matching network based on attention mechanism is proposed.This network dynamically adjusts attention weights for cross-modal information from grayscale images and optical flow images through a dual-branch Siamese network.It extracts static and dynamic information features ranging from low-level to high-level,and from local to global,ensuring robust feature extraction across different widths,noise levels,and blur scenarios.Global and local matching are performed through a multi-level cascaded attention mechanism,with cross-channel attention introduced to simultaneously extract low-level and high-level features.Extensive ablation experiments and comparative studies are conducted on the HyperKvasir,EAD,M2caiSeg,CVC-ClinicDB,and UCL synthetic datasets.Experimental results demonstrate that the proposed network improves upon the baseline EfficientViT-B3 model by 75.4%in accuracy(Acc),while also enhancing runtime performance and storage efficiency.When compared with the complex DenseDescriptor feature extraction network,the difference in Acc is less than 7.22%,and IoU calculation results on specific datasets outperform complex dense models.Furthermore,this method increases the F1 score by 33.2%and accelerates runtime by 70.2%.It is noteworthy that the speed of CMMCAN surpasses that of comparative lightweight models,with feature extraction and matching performance comparable to existing complex models but with faster speed and higher cost-effectiveness.展开更多
Artificial immune detection can be used to detect network intrusions in an adaptive approach and proper matching methods can improve the accuracy of immune detection methods.This paper proposes an artificial immune de...Artificial immune detection can be used to detect network intrusions in an adaptive approach and proper matching methods can improve the accuracy of immune detection methods.This paper proposes an artificial immune detection model for network intrusion data based on a quantitative matching method.The proposed model defines the detection process by using network data and decimal values to express features and artificial immune mechanisms are simulated to define immune elements.Then,to improve the accuracy of similarity calculation,a quantitative matching method is proposed.The model uses mathematical methods to train and evolve immune elements,increasing the diversity of immune recognition and allowing for the successful detection of unknown intrusions.The proposed model’s objective is to accurately identify known intrusions and expand the identification of unknown intrusions through signature detection and immune detection,overcoming the disadvantages of traditional methods.The experiment results show that the proposed model can detect intrusions effectively.It has a detection rate of more than 99.6%on average and a false alarm rate of 0.0264%.It outperforms existing immune intrusion detection methods in terms of comprehensive detection performance.展开更多
Road traffic monitoring is an imperative topic widely discussed among researchers.Systems used to monitor traffic frequently rely on cameras mounted on bridges or roadsides.However,aerial images provide the flexibilit...Road traffic monitoring is an imperative topic widely discussed among researchers.Systems used to monitor traffic frequently rely on cameras mounted on bridges or roadsides.However,aerial images provide the flexibility to use mobile platforms to detect the location and motion of the vehicle over a larger area.To this end,different models have shown the ability to recognize and track vehicles.However,these methods are not mature enough to produce accurate results in complex road scenes.Therefore,this paper presents an algorithm that combines state-of-the-art techniques for identifying and tracking vehicles in conjunction with image bursts.The extracted frames were converted to grayscale,followed by the application of a georeferencing algorithm to embed coordinate information into the images.The masking technique eliminated irrelevant data and reduced the computational cost of the overall monitoring system.Next,Sobel edge detection combined with Canny edge detection and Hough line transform has been applied for noise reduction.After preprocessing,the blob detection algorithm helped detect the vehicles.Vehicles of varying sizes have been detected by implementing a dynamic thresholding scheme.Detection was done on the first image of every burst.Then,to track vehicles,the model of each vehicle was made to find its matches in the succeeding images using the template matching algorithm.To further improve the tracking accuracy by incorporating motion information,Scale Invariant Feature Transform(SIFT)features have been used to find the best possible match among multiple matches.An accuracy rate of 87%for detection and 80%accuracy for tracking in the A1 Motorway Netherland dataset has been achieved.For the Vehicle Aerial Imaging from Drone(VAID)dataset,an accuracy rate of 86%for detection and 78%accuracy for tracking has been achieved.展开更多
Randomized controlled trials(RCTs)have long been recognized as the gold standard for establishing causal relationships in clinical research.Despite that,various limitations of RCTs prevent its widespread implementatio...Randomized controlled trials(RCTs)have long been recognized as the gold standard for establishing causal relationships in clinical research.Despite that,various limitations of RCTs prevent its widespread implementation,ranging from the ethicality of withholding potentially-lifesaving treatment from a group to relatively poor external validity due to stringent inclusion criteria,amongst others.However,with the introduction of propensity score matching(PSM)as a retrospective statistical tool,new frontiers in establishing causation in clinical research were opened up.PSM predicts treatment effects using observational data from existing sources such as registries or electronic health records,to create a matched sample of participants who received or did not receive the intervention based on their propensity scores,which takes into account characteristics such as age,gender and comorbidities.Given its retrospective nature and its use of observational data from existing sources,PSM circumvents the aforementioned ethical issues faced by RCTs.Majority of RCTs exclude elderly,pregnant women and young children;thus,evidence of therapy efficacy is rarely proven by robust clinical research for this population.On the other hand,by matching study patient characteristics to that of the population of interest,including the elderly,pregnant women and young children,PSM allows for generalization of results to the wider population and hence greatly increases the external validity.Instead of replacing RCTs with PSM,the synergistic integration of PSM into RCTs stands to provide better research outcomes with both methods complementing each other.For example,in an RCT investigating the impact of mannitol on outcomes among participants of the Intensive Blood Pressure Reduction in Acute Cerebral Hemorrhage Trial,the baseline characteristics of comorbidities and current medications between treatment and control arms were significantly different despite the randomization protocol.Therefore,PSM was incorporated in its analysis to create samples from the treatment and control arms that were matched in terms of these baseline characteristics,thus providing a fairer comparison for the impact of mannitol.This literature review reports the applications,advantages,and considerations of using PSM with RCTs,illustrating its utility in refining randomization,improving external validity,and accounting for non-compliance to protocol.Future research should consider integrating the use of PSM in RCTs to better generalize outcomes to target populations for clinical practice and thereby benefit a wider range of patients,while maintaining the robustness of randomization offered by RCTs.展开更多
The flexibility in radiotherapy can be improved if patients can be moved between any one of the department’s medical linear accelerators (LINACs) without the need to change anything in the patient’s treatment plan. ...The flexibility in radiotherapy can be improved if patients can be moved between any one of the department’s medical linear accelerators (LINACs) without the need to change anything in the patient’s treatment plan. For this to be possible, the dosimetric characteristics of the various accelerators must be the same, or nearly the same. The purpose of this work is to describe further and compare measurements and parameters after the initial vendor-recommended beam matching of the five LINACs. Deviations related to dose calculations and to beam matched accelerators may compromise treatment accuracy. The safest and most practical way to ensure that all accelerators are within clinical acceptable accuracy is to include TPS calculations in the LINACs matching evaluation. Treatment planning system (TPS) was used to create three photons plans with different field sizes 3 × 3 cm, 10 × 10 cm and 25 × 25 cm at a depth of 4.5 cm in Perspex. Calculated TPS plans were sent to Mosaiq to be delivered by five LINACs. TPS plans were compared with five LINACs measurements data using Gamma analyses of 2% and 2 mm. The results suggest that for four out of the five LINACs, there was generally good agreement, less than a 2% deviation between the planned dose distribution and the measured dose distribution. However, one specific LINAC named “Asterix” exhibited a deviation of 2.121% from the planned dose. The results show that all of the LINACs’ performance were within the acceptable deviation and delivering radiation dose consistently and accurately.展开更多
A critical component of visual simultaneous localization and mapping is loop closure detection(LCD),an operation judging whether a robot has come to a pre-visited area.Concretely,given a query image(i.e.,the latest vi...A critical component of visual simultaneous localization and mapping is loop closure detection(LCD),an operation judging whether a robot has come to a pre-visited area.Concretely,given a query image(i.e.,the latest view observed by the robot),it proceeds by first exploring images with similar semantic information,followed by solving the relative relationship between candidate pairs in the 3D space.In this work,a novel appearance-based LCD system is proposed.Specifically,candidate frame selection is conducted via the combination of Superfeatures and aggregated selective match kernel(ASMK).We incorporate an incremental strategy into the vanilla ASMK to make it applied in the LCD task.It is demonstrated that this setting is memory-wise efficient and can achieve remarkable performance.To dig up consistent geometry between image pairs during loop closure verification,we propose a simple yet surprisingly effective feature matching algorithm,termed locality preserving matching with global consensus(LPM-GC).The major objective of LPM-GC is to retain the local neighborhood information of true feature correspondences between candidate pairs,where a global constraint is further designed to effectively remove false correspondences in challenging sceneries,e.g.,containing numerous repetitive structures.Meanwhile,we derive a closed-form solution that enables our approach to provide reliable correspondences within only a few milliseconds.The performance of the proposed approach has been experimentally evaluated on ten publicly available and challenging datasets.Results show that our method can achieve better performance over the state-of-the-art in both feature matching and LCD tasks.We have released our code of LPM-GC at https://github.com/jiayi-ma/LPM-GC.展开更多
Background: Minimally invasive surgery is becoming increasingly popular in the field of pancreatic surgery. However, there are few studies of robotic distal pancreatectomy(RDP) for pancreatic ductal adenocarcinoma(PDA...Background: Minimally invasive surgery is becoming increasingly popular in the field of pancreatic surgery. However, there are few studies of robotic distal pancreatectomy(RDP) for pancreatic ductal adenocarcinoma(PDAC). This study aimed to investigate the efficacy and feasibility of RDP for PDAC. Methods: Patients who underwent RDP or laparoscopic distal pancreatectomy(LDP) for PDAC between January 2015 and September 2020 were reviewed. Propensity score matching analyses were performed. Results: Of the 335 patients included in the study, 24 underwent RDP and 311 underwent LDP. A total of 21 RDP patients were matched 1:1 with LDP patients. RDP was associated with longer operative time(209.7 vs. 163.2 min;P = 0.003), lower open conversion rate(0% vs. 4.8%;P < 0.001), higher cost(15 722 vs. 12 699 dollars;P = 0.003), and a higher rate of achievement of an R0 resection margin(90.5% vs. 61.9%;P = 0.042). However, postoperative pancreatic fistula grade B or C showed no significant intergroup difference(9.5% vs. 9.5%). The median disease-free survival(34.5 vs. 17.3 months;P = 0.588) and overall survival(37.7 vs. 21.9 months;P = 0.171) were comparable between the groups. Conclusions: RDP is associated with longer operative time, a higher cost of surgery, and a higher likelihood of achieving R0 margins than LDP.展开更多
Background: Significant portal hypertension(SPH) is a relative contraindication for patients with resectable hepatocellular carcinoma(HCC). However, increasing evidence indicates that liver resection is feasible for H...Background: Significant portal hypertension(SPH) is a relative contraindication for patients with resectable hepatocellular carcinoma(HCC). However, increasing evidence indicates that liver resection is feasible for HCC patients with SPH. Methods: HCC patients with cirrhosis who underwent laparoscopic liver resection(LLR) in two centers from January 2013 to April 2018 were included. Surgical and survival outcomes were analyzed to explore potential prognostic factors. Propensity score matching(PSM) analysis was performed to minimize bias. Results: A total of 165 patients were divided into two groups based on the presence(SPH, n = 76) or absence(non-SPH, n = 89) of SPH. Patients in the SPH group had longer operative time, more blood loss, and more advanced TNM stage than patients in the non-SPH group( P < 0.05). However, there were no significant differences in the postoperative 90-day mortality rate( n = 0), overall postoperative complications(47.4% vs. 41.6%, P = 0.455), Clavien-Dindo classification( P = 0.347), conversion to open surgery(9.2% vs. 6.7%, P = 0.557), or length of hospitalization(16 vs. 15 days, P = 0.203) between the SPH and non-SPH groups before PSM. Similar results were obtained after PSM. The 1-, 3-, and 5-year overall survival(OS) and recurrence-free survival rates in the SPH group were not significantly different from those in the non-SPH group both before and after PSM(log-rank P > 0.05). After PSM, alpha-fetoprotein(AFP) ≥ 400 μg/L [hazard ratio(HR) = 4.71, 95% confidence interval(CI): 2.69-8.25], ascites(HR = 2.18, 95% CI: 1.30-3.66), American Society of Anesthesiologists(ASA) classification(Ⅲ vs. Ⅱ)(HR = 2.13, 95% CI: 1.11-4.07) and tumor diameter > 5 cm(HR = 3.91, 95% CI: 2.02-7.56) independently predicted worse OS. Conclusions: LLR for patients with HCC complicated with SPH appears feasible at the price of increasing operative time and blood loss. AFP, ascites, ASA classification and tumor diameter may predict the prognosis of HCC complicated with SPH after LLR.展开更多
The Li-ion capacitors(LICs)develop rapidly due to their double-high features of high-energy density and high-power density.However,the relative low capacity of cathode and sluggish kinetics of anode seriously impede t...The Li-ion capacitors(LICs)develop rapidly due to their double-high features of high-energy density and high-power density.However,the relative low capacity of cathode and sluggish kinetics of anode seriously impede the development of LICs.Herein,the precisely pore-engineered and heteroatomtailored defective hierarchical porous carbons(DHPCs)as large-capacity cathode and high-rate anode to construct high-performance dual-carbon LICs have been developed.The DHPCs are prepared based on triple-activation mechanisms by direct pyrolysis of sustainable lignin with urea to generate the interconnected hierarchical porous structure and plentiful heteroatominduced defects.Benefiting from these advanced merits,DHPCs show the well-matched high capacity and fast kinetics of both cathode and anode,exhibiting large capacities,superior rate capability and long-term lifespan.Both experimental and computational results demonstrate the strong synergistic effect of pore and dopants for Li storage.Consequently,the assembled dual-carbon LIC exhibits high voltage of 4.5 V,high-energy density of 208 Wh kg^(−1),ultrahigh power density of 53.4 kW kg^(−1)and almost zerodecrement cycling lifetime.Impressively,the full device with high mass loading of 9.4 mg cm^(−2)on cathode still outputs high-energy density of 187 Wh kg^(−1),demonstrative of their potential as electrode materials for high-performance electrochemical devices.展开更多
In this paper,a stable two-sided matching(TSM)method considering the matching intention of agents under a hesitant fuzzy environment is proposed.The method uses a hesitant fuzzy element(HFE)as its basis.First,the HFE ...In this paper,a stable two-sided matching(TSM)method considering the matching intention of agents under a hesitant fuzzy environment is proposed.The method uses a hesitant fuzzy element(HFE)as its basis.First,the HFE preference matrix is transformed into the normalized HFE preference matrix.On this basis,the distance and the projection of the normalized HFEs on positive and negative ideal solutions are calculated.Then,the normalized HFEs are transformed into agent satisfactions.Considering the stable matching constraints,a multiobjective programming model with the objective of maximizing the satisfactions of two-sided agents is constructed.Based on the agent satisfaction matrix,the matching intention matrix of two-sided agents is built.According to the agent satisfaction matrix and matching intention matrix,the comprehensive satisfaction matrix is set up.Furthermore,the multiobjective programming model based on satisfactions is transformed into a multiobjective programming model based on comprehensive satisfactions.Using the G-S algorithm,the multiobjective programming model based on comprehensive satisfactions is solved,and then the best TSM scheme is obtained.Finally,a terminal distribution example is used to verify the feasibility and effectiveness of the proposed method.展开更多
As the fundamental problem in the computer vision area,image matching has wide applications in pose estimation,3D reconstruction,image retrieval,etc.Suffering from the influence of external factors,the process of imag...As the fundamental problem in the computer vision area,image matching has wide applications in pose estimation,3D reconstruction,image retrieval,etc.Suffering from the influence of external factors,the process of image matching using classical local detectors,e.g.,scale-invariant feature transform(SIFT),and the outlier filtering approaches,e.g.,Random sample consensus(RANSAC),show high computation speed and pool robustness under changing illumination and viewpoints conditions,while image matching approaches with deep learning strategy(such as HardNet,OANet)display reliable achievements in large-scale datasets with challenging scenes.However,the past learning-based approaches are limited to the distinction and quality of the dataset and the training strategy in the image-matching approaches.As an extension of the previous conference paper,this paper proposes an accurate and robust image matching approach using fewer training data in an end-to-end manner,which could be used to estimate the pose error This research first proposes a novel dataset cleaning and construction strategy to eliminate the noise and improve the training efficiency;Secondly,a novel loss named quadratic hinge triplet loss(QHT)is proposed to gather more effective and stable feature matching;Thirdly,in the outlier filtering process,the stricter OANet and bundle adjustment are applied for judging samples by adding the epipolar distance constraint and triangulation constraint to generate more outstanding matches;Finally,to recall the matching pairs,dynamic guided matching is used and then submit the inliers after the PyRANSAC process.Multiple evaluation metrics are used and reported in the 1st place in the Track1 of CVPR Image-Matching Challenge Workshop.The results show that the proposed method has advanced performance in large-scale and challenging Phototourism benchmark.展开更多
BACKGROUND Patients with atrial fibrillation(AF)and prior stroke history have a high risk of cardiovascular events despite anticoagulation therapy.It is unclear whether catheter ablation(CA)has further benefits in the...BACKGROUND Patients with atrial fibrillation(AF)and prior stroke history have a high risk of cardiovascular events despite anticoagulation therapy.It is unclear whether catheter ablation(CA)has further benefits in these patients.METHODS AF patients with a previous history of stroke or systemic embolism(SE)from the prospective Chinese Atrial Fibrillation Registry study between August 2011 and December 2020 were included in the analysis.Patients were matched in a 1:1 ratio to CA or medical treatment(MT)based on propensity score.The primary outcome was a composite of all-cause death or ischemic stroke(IS)/SE.RESULTS During a total of 4.1±2.3 years of follow-up,the primary outcome occurred in 111 patients in the CA group(3.3 per 100 person-years)and in 229 patients in the MT group(5.7 per 100 person-years).The CA group had a lower risk of the primary outcome compared to the MT group[hazard ratio(HR)=0.59,95%CI:0.47–0.74,P<0.001].There was a significant decreasing risk of all-cause mortality(HR=0.43,95%CI:0.31–0.61,P<0.001),IS/SE(HR=0.73,95%CI:0.54–0.97,P=0.033),cardiovascular mortality(HR=0.32,95%CI:0.19–0.54,P<0.001)and AF recurrence(HR=0.33,95%CI:0.30–0.37,P<0.001)in the CA group compared to that in the MT group.Sensitivity analysis generated consistent results when adjusting for time-dependent usage of anticoagulants.CONCLUSIONS In AF patients with a prior stroke history,CA was associated with a lower combined risk of all-cause death or IS/SE.Further clinical trials are warranted to confirm the benefits of CA in these patients.展开更多
Graph pattern matching(GPM)can be used to mine the key information in graphs.Exact GPM is one of the most commonly used methods among all the GPM-related methods,which aims to exactly find all subgraphs for a given qu...Graph pattern matching(GPM)can be used to mine the key information in graphs.Exact GPM is one of the most commonly used methods among all the GPM-related methods,which aims to exactly find all subgraphs for a given query graph in a data graph.The exact GPM has been widely used in biological data analyses,social network analyses and other fields.In this paper,the applications of the exact GPM were first introduced,and the research progress of the exact GPM was summarized.Then,the related algorithms were introduced in detail,and the experiments on the state-of-the-art exact GPM algorithms were conducted to compare their performance.Based on the experimental results,the applicable scenarios of the algorithms were pointed out.New research opportunities in this area were proposed.展开更多
文摘Pattern matching method is one of the classic classifications of existing online portfolio selection strategies. This article aims to study the key aspects of this method—measurement of similarity and selection of similarity sets, and proposes a Portfolio Selection Method based on Pattern Matching with Dual Information of Direction and Distance (PMDI). By studying different combination methods of indicators such as Euclidean distance, Chebyshev distance, and correlation coefficient, important information such as direction and distance in stock historical price information is extracted, thereby filtering out the similarity set required for pattern matching based investment portfolio selection algorithms. A large number of experiments conducted on two datasets of real stock markets have shown that PMDI outperforms other algorithms in balancing income and risk. Therefore, it is suitable for the financial environment in the real world.
基金supported by the National Natural Science Foundation of China (62276192)。
文摘Feature matching plays a key role in computer vision. However, due to the limitations of the descriptors, the putative matches are inevitably contaminated by massive outliers.This paper attempts to tackle the outlier filtering problem from two aspects. First, a robust and efficient graph interaction model,is proposed, with the assumption that matches are correlated with each other rather than independently distributed. To this end, we construct a graph based on the local relationships of matches and formulate the outlier filtering task as a binary labeling energy minimization problem, where the pairwise term encodes the interaction between matches. We further show that this formulation can be solved globally by graph cut algorithm. Our new formulation always improves the performance of previous localitybased method without noticeable deterioration in processing time,adding a few milliseconds. Second, to construct a better graph structure, a robust and geometrically meaningful topology-aware relationship is developed to capture the topology relationship between matches. The two components in sum lead to topology interaction matching(TIM), an effective and efficient method for outlier filtering. Extensive experiments on several large and diverse datasets for multiple vision tasks including general feature matching, as well as relative pose estimation, homography and fundamental matrix estimation, loop-closure detection, and multi-modal image matching, demonstrate that our TIM is more competitive than current state-of-the-art methods, in terms of generality, efficiency, and effectiveness. The source code is publicly available at http://github.com/YifanLu2000/TIM.
基金funded by the Wenhai Program of the ST Fund of Laoshan Laboratory (No.202204803)the National Natural Science Foundation of China (Nos.42074138,42206195)+1 种基金the National Key R&D Program of China (No.2022YFC2803501)the Research Project of the China National Petroleum Corporation (No.2021ZG02)。
文摘The paper develops a multiple matching attenuation method based on extended filtering in the curvelet domain,which combines the traditional Wiener filtering method with the matching attenuation method in curvelet domain.Firstly,the method uses the predicted multiple data to generate the Hilbert transform records,time derivative records and time derivative records of Hilbert transform.Then,the above records are transformed into the curvelet domain and multiple matching attenuation based on least squares extended filtering is performed.Finally,the attenuation results are transformed back into the time-space domain.Tests on the model data and field data show that the method proposed in the paper effectively suppress the multiples while preserving the primaries well.Furthermore,it has higher accuracy in eliminating multiple reflections,which is more suitable for the multiple attenuation tasks in the areas with complex structures compared to the time-space domain extended filtering method and the conventional curvelet transform method.
基金supported by the National Natural Science Foundation of China(62033010)Qing Lan Project of Jiangsu Province(R2023Q07)。
文摘For accurately identifying the distribution charac-teristic of Gaussian-like noises in unmanned aerial vehicle(UAV)state estimation,this paper proposes a non-parametric scheme based on curve similarity matching.In the framework of the pro-posed scheme,a Parzen window(kernel density estimation,KDE)method on sliding window technology is applied for roughly esti-mating the sample probability density,a precise data probability density function(PDF)model is constructed with the least square method on K-fold cross validation,and the testing result based on evaluation method is obtained based on some data characteristic analyses of curve shape,abruptness and symmetry.Some com-parison simulations with classical methods and UAV flight exper-iment shows that the proposed scheme has higher recognition accuracy than classical methods for some kinds of Gaussian-like data,which provides better reference for the design of Kalman filter(KF)in complex water environment.
基金supported by the National Natural Science Foundation of China under Grant 62171465。
文摘Many efforts have been devoted to efficient task scheduling in Multi-Unmanned Aerial Vehicle(UAV)edge computing.However,the heterogeneity of UAV computation resource,and the task re-allocating between UAVs have not been fully considered yet.Moreover,most existing works neglect the fact that a task can only be executed on the UAV equipped with its desired service function(SF).In this backdrop,this paper formulates the task scheduling problem as a multi-objective task scheduling problem,which aims at maximizing the task execution success ratio while minimizing the average weighted sum of all tasks’completion time and energy consumption.Optimizing three coupled goals in a realtime manner with the dynamic arrival of tasks hinders us from adopting existing methods,like machine learning-based solutions that require a long training time and tremendous pre-knowledge about the task arrival process,or heuristic-based ones that usually incur a long decision-making time.To tackle this problem in a distributed manner,we establish a matching theory framework,in which three conflicting goals are treated as the preferences of tasks,SFs and UAVs.Then,a Distributed Matching Theory-based Re-allocating(DiMaToRe)algorithm is put forward.We formally proved that a stable matching can be achieved by our proposal.Extensive simulation results show that Di Ma To Re algorithm outperforms benchmark algorithms under diverse parameter settings and has good robustness.
基金funded by the Fujian Province Science and Technology Plan,China(Grant Number 2019H0017).
文摘Accurate forecasting of time series is crucial across various domains.Many prediction tasks rely on effectively segmenting,matching,and time series data alignment.For instance,regardless of time series with the same granularity,segmenting them into different granularity events can effectively mitigate the impact of varying time scales on prediction accuracy.However,these events of varying granularity frequently intersect with each other,which may possess unequal durations.Even minor differences can result in significant errors when matching time series with future trends.Besides,directly using matched events but unaligned events as state vectors in machine learning-based prediction models can lead to insufficient prediction accuracy.Therefore,this paper proposes a short-term forecasting method for time series based on a multi-granularity event,MGE-SP(multi-granularity event-based short-termprediction).First,amethodological framework for MGE-SP established guides the implementation steps.The framework consists of three key steps,including multi-granularity event matching based on the LTF(latest time first)strategy,multi-granularity event alignment using a piecewise aggregate approximation based on the compression ratio,and a short-term prediction model based on XGBoost.The data from a nationwide online car-hailing service in China ensures the method’s reliability.The average RMSE(root mean square error)and MAE(mean absolute error)of the proposed method are 3.204 and 2.360,lower than the respective values of 4.056 and 3.101 obtained using theARIMA(autoregressive integratedmoving average)method,as well as the values of 4.278 and 2.994 obtained using k-means-SVR(support vector regression)method.The other experiment is conducted on stock data froma public data set.The proposed method achieved an average RMSE and MAE of 0.836 and 0.696,lower than the respective values of 1.019 and 0.844 obtained using the ARIMA method,as well as the values of 1.350 and 1.172 obtained using the k-means-SVR method.
基金supported by the National Key Research and Development Program of China(2021YFC2501800)Leader Project of Henan Province Health Young and Middle-aged Professor(HNSWJW2020013).
文摘BACKGROUND:Patients with diabetes mellitus(DM)are vulnerable to community-acquired pneumonia(CAP),which have a high mortality rate.We aimed to investigate the value of heparin-binding protein(HBP)as a prognostic marker of mortality in patients with DM and CAP.METHODS:This retrospective study included CAP patients who were tested for HBP at intensive care unit(ICU)admission from January 2019 to April 2020.Patients were allocated to the DM or non-DM group and paired with propensity score matching.Baseline characteristics and clinical outcomes up to 90 days were evaluated.The primary outcome was the 10-day mortality.Receiver operating characteristic(ROC)curves,Kaplan-Meier analysis,and Cox regression were used for statistical analysis.RESULTS:Among 152 enrolled patients,60 pairs were successfully matched.There was no significant difference in 10-day mortality,while more patients in the DM group died within 28 d(P=0.024)and 90 d(P=0.008).In the DM group,HBP levels at ICU admission were higher in 10-day non-survivors than in 10-day survivors(median 182.21[IQR:55.43-300]ng/ml vs.median 66.40[IQR:34.13-107.85]ng/mL,P=0.019),and HBP levels could predict the 10-day mortality with an area under the ROC curve of 0.747.The cut-off value,sensitivity,and specificity were 160.6 ng/mL,66.7%,and 90.2%,respectively.Multivariate Cox regression analysis indicated that HBP was an independent prognostic factor for 10-day(HR 7.196,95%CI:1.596-32.455,P=0.01),28-day(HR 4.381,95%CI:1.449-13.245,P=0.009),and 90-day mortality(HR 4.581,95%CI:1.637-12.819,P=0.004)in patients with DM.CONCLUSION:Plasma HBP at ICU admission was associated with the 10-day,28-day,and 90-day mortality,and might be a prognostic factor in patients with DM and CAP.
基金This work was supported by Science and Technology Cooperation Special Project of Shijiazhuang(SJZZXA23005).
文摘In minimally invasive surgery,endoscopes or laparoscopes equipped with miniature cameras and tools are used to enter the human body for therapeutic purposes through small incisions or natural cavities.However,in clinical operating environments,endoscopic images often suffer from challenges such as low texture,uneven illumination,and non-rigid structures,which affect feature observation and extraction.This can severely impact surgical navigation or clinical diagnosis due to missing feature points in endoscopic images,leading to treatment and postoperative recovery issues for patients.To address these challenges,this paper introduces,for the first time,a Cross-Channel Multi-Modal Adaptive Spatial Feature Fusion(ASFF)module based on the lightweight architecture of EfficientViT.Additionally,a novel lightweight feature extraction and matching network based on attention mechanism is proposed.This network dynamically adjusts attention weights for cross-modal information from grayscale images and optical flow images through a dual-branch Siamese network.It extracts static and dynamic information features ranging from low-level to high-level,and from local to global,ensuring robust feature extraction across different widths,noise levels,and blur scenarios.Global and local matching are performed through a multi-level cascaded attention mechanism,with cross-channel attention introduced to simultaneously extract low-level and high-level features.Extensive ablation experiments and comparative studies are conducted on the HyperKvasir,EAD,M2caiSeg,CVC-ClinicDB,and UCL synthetic datasets.Experimental results demonstrate that the proposed network improves upon the baseline EfficientViT-B3 model by 75.4%in accuracy(Acc),while also enhancing runtime performance and storage efficiency.When compared with the complex DenseDescriptor feature extraction network,the difference in Acc is less than 7.22%,and IoU calculation results on specific datasets outperform complex dense models.Furthermore,this method increases the F1 score by 33.2%and accelerates runtime by 70.2%.It is noteworthy that the speed of CMMCAN surpasses that of comparative lightweight models,with feature extraction and matching performance comparable to existing complex models but with faster speed and higher cost-effectiveness.
基金This research was funded by the Scientific Research Project of Leshan Normal University(No.2022SSDX002)the Scientific Plan Project of Leshan(No.22NZD012).
文摘Artificial immune detection can be used to detect network intrusions in an adaptive approach and proper matching methods can improve the accuracy of immune detection methods.This paper proposes an artificial immune detection model for network intrusion data based on a quantitative matching method.The proposed model defines the detection process by using network data and decimal values to express features and artificial immune mechanisms are simulated to define immune elements.Then,to improve the accuracy of similarity calculation,a quantitative matching method is proposed.The model uses mathematical methods to train and evolve immune elements,increasing the diversity of immune recognition and allowing for the successful detection of unknown intrusions.The proposed model’s objective is to accurately identify known intrusions and expand the identification of unknown intrusions through signature detection and immune detection,overcoming the disadvantages of traditional methods.The experiment results show that the proposed model can detect intrusions effectively.It has a detection rate of more than 99.6%on average and a false alarm rate of 0.0264%.It outperforms existing immune intrusion detection methods in terms of comprehensive detection performance.
基金supported by a grant from the Basic Science Research Program through the National Research Foundation(NRF)(2021R1F1A1063634)funded by the Ministry of Science and ICT(MSIT),Republic of KoreaThe authors are thankful to the Deanship of Scientific Research at Najran University for funding this work under the Research Group Funding Program Grant Code(NU/RG/SERC/13/40)+2 种基金Also,the authors are thankful to Prince Satam bin Abdulaziz University for supporting this study via funding from Prince Satam bin Abdulaziz University project number(PSAU/2024/R/1445)This work was also supported by Princess Nourah bint Abdulrahman University Researchers Supporting Project Number(PNURSP2023R54)Princess Nourah bint Abdulrahman University,Riyadh,Saudi Arabia.
文摘Road traffic monitoring is an imperative topic widely discussed among researchers.Systems used to monitor traffic frequently rely on cameras mounted on bridges or roadsides.However,aerial images provide the flexibility to use mobile platforms to detect the location and motion of the vehicle over a larger area.To this end,different models have shown the ability to recognize and track vehicles.However,these methods are not mature enough to produce accurate results in complex road scenes.Therefore,this paper presents an algorithm that combines state-of-the-art techniques for identifying and tracking vehicles in conjunction with image bursts.The extracted frames were converted to grayscale,followed by the application of a georeferencing algorithm to embed coordinate information into the images.The masking technique eliminated irrelevant data and reduced the computational cost of the overall monitoring system.Next,Sobel edge detection combined with Canny edge detection and Hough line transform has been applied for noise reduction.After preprocessing,the blob detection algorithm helped detect the vehicles.Vehicles of varying sizes have been detected by implementing a dynamic thresholding scheme.Detection was done on the first image of every burst.Then,to track vehicles,the model of each vehicle was made to find its matches in the succeeding images using the template matching algorithm.To further improve the tracking accuracy by incorporating motion information,Scale Invariant Feature Transform(SIFT)features have been used to find the best possible match among multiple matches.An accuracy rate of 87%for detection and 80%accuracy for tracking in the A1 Motorway Netherland dataset has been achieved.For the Vehicle Aerial Imaging from Drone(VAID)dataset,an accuracy rate of 86%for detection and 78%accuracy for tracking has been achieved.
文摘Randomized controlled trials(RCTs)have long been recognized as the gold standard for establishing causal relationships in clinical research.Despite that,various limitations of RCTs prevent its widespread implementation,ranging from the ethicality of withholding potentially-lifesaving treatment from a group to relatively poor external validity due to stringent inclusion criteria,amongst others.However,with the introduction of propensity score matching(PSM)as a retrospective statistical tool,new frontiers in establishing causation in clinical research were opened up.PSM predicts treatment effects using observational data from existing sources such as registries or electronic health records,to create a matched sample of participants who received or did not receive the intervention based on their propensity scores,which takes into account characteristics such as age,gender and comorbidities.Given its retrospective nature and its use of observational data from existing sources,PSM circumvents the aforementioned ethical issues faced by RCTs.Majority of RCTs exclude elderly,pregnant women and young children;thus,evidence of therapy efficacy is rarely proven by robust clinical research for this population.On the other hand,by matching study patient characteristics to that of the population of interest,including the elderly,pregnant women and young children,PSM allows for generalization of results to the wider population and hence greatly increases the external validity.Instead of replacing RCTs with PSM,the synergistic integration of PSM into RCTs stands to provide better research outcomes with both methods complementing each other.For example,in an RCT investigating the impact of mannitol on outcomes among participants of the Intensive Blood Pressure Reduction in Acute Cerebral Hemorrhage Trial,the baseline characteristics of comorbidities and current medications between treatment and control arms were significantly different despite the randomization protocol.Therefore,PSM was incorporated in its analysis to create samples from the treatment and control arms that were matched in terms of these baseline characteristics,thus providing a fairer comparison for the impact of mannitol.This literature review reports the applications,advantages,and considerations of using PSM with RCTs,illustrating its utility in refining randomization,improving external validity,and accounting for non-compliance to protocol.Future research should consider integrating the use of PSM in RCTs to better generalize outcomes to target populations for clinical practice and thereby benefit a wider range of patients,while maintaining the robustness of randomization offered by RCTs.
文摘The flexibility in radiotherapy can be improved if patients can be moved between any one of the department’s medical linear accelerators (LINACs) without the need to change anything in the patient’s treatment plan. For this to be possible, the dosimetric characteristics of the various accelerators must be the same, or nearly the same. The purpose of this work is to describe further and compare measurements and parameters after the initial vendor-recommended beam matching of the five LINACs. Deviations related to dose calculations and to beam matched accelerators may compromise treatment accuracy. The safest and most practical way to ensure that all accelerators are within clinical acceptable accuracy is to include TPS calculations in the LINACs matching evaluation. Treatment planning system (TPS) was used to create three photons plans with different field sizes 3 × 3 cm, 10 × 10 cm and 25 × 25 cm at a depth of 4.5 cm in Perspex. Calculated TPS plans were sent to Mosaiq to be delivered by five LINACs. TPS plans were compared with five LINACs measurements data using Gamma analyses of 2% and 2 mm. The results suggest that for four out of the five LINACs, there was generally good agreement, less than a 2% deviation between the planned dose distribution and the measured dose distribution. However, one specific LINAC named “Asterix” exhibited a deviation of 2.121% from the planned dose. The results show that all of the LINACs’ performance were within the acceptable deviation and delivering radiation dose consistently and accurately.
基金supported by the Key Research and Development Program of Hubei Province(2020BAB113)。
文摘A critical component of visual simultaneous localization and mapping is loop closure detection(LCD),an operation judging whether a robot has come to a pre-visited area.Concretely,given a query image(i.e.,the latest view observed by the robot),it proceeds by first exploring images with similar semantic information,followed by solving the relative relationship between candidate pairs in the 3D space.In this work,a novel appearance-based LCD system is proposed.Specifically,candidate frame selection is conducted via the combination of Superfeatures and aggregated selective match kernel(ASMK).We incorporate an incremental strategy into the vanilla ASMK to make it applied in the LCD task.It is demonstrated that this setting is memory-wise efficient and can achieve remarkable performance.To dig up consistent geometry between image pairs during loop closure verification,we propose a simple yet surprisingly effective feature matching algorithm,termed locality preserving matching with global consensus(LPM-GC).The major objective of LPM-GC is to retain the local neighborhood information of true feature correspondences between candidate pairs,where a global constraint is further designed to effectively remove false correspondences in challenging sceneries,e.g.,containing numerous repetitive structures.Meanwhile,we derive a closed-form solution that enables our approach to provide reliable correspondences within only a few milliseconds.The performance of the proposed approach has been experimentally evaluated on ten publicly available and challenging datasets.Results show that our method can achieve better performance over the state-of-the-art in both feature matching and LCD tasks.We have released our code of LPM-GC at https://github.com/jiayi-ma/LPM-GC.
文摘Background: Minimally invasive surgery is becoming increasingly popular in the field of pancreatic surgery. However, there are few studies of robotic distal pancreatectomy(RDP) for pancreatic ductal adenocarcinoma(PDAC). This study aimed to investigate the efficacy and feasibility of RDP for PDAC. Methods: Patients who underwent RDP or laparoscopic distal pancreatectomy(LDP) for PDAC between January 2015 and September 2020 were reviewed. Propensity score matching analyses were performed. Results: Of the 335 patients included in the study, 24 underwent RDP and 311 underwent LDP. A total of 21 RDP patients were matched 1:1 with LDP patients. RDP was associated with longer operative time(209.7 vs. 163.2 min;P = 0.003), lower open conversion rate(0% vs. 4.8%;P < 0.001), higher cost(15 722 vs. 12 699 dollars;P = 0.003), and a higher rate of achievement of an R0 resection margin(90.5% vs. 61.9%;P = 0.042). However, postoperative pancreatic fistula grade B or C showed no significant intergroup difference(9.5% vs. 9.5%). The median disease-free survival(34.5 vs. 17.3 months;P = 0.588) and overall survival(37.7 vs. 21.9 months;P = 0.171) were comparable between the groups. Conclusions: RDP is associated with longer operative time, a higher cost of surgery, and a higher likelihood of achieving R0 margins than LDP.
基金supported by grants from the National Natu-ral Science Foundation of China(81701950 and 82172135)Medi-cal Research Projects of Chongqing for staffagainst the epidemic(2020FYYX248)the Kuanren Talents Program of the Second Affiliated Hospital,Chongqing Medical University(KY2019Y002).
文摘Background: Significant portal hypertension(SPH) is a relative contraindication for patients with resectable hepatocellular carcinoma(HCC). However, increasing evidence indicates that liver resection is feasible for HCC patients with SPH. Methods: HCC patients with cirrhosis who underwent laparoscopic liver resection(LLR) in two centers from January 2013 to April 2018 were included. Surgical and survival outcomes were analyzed to explore potential prognostic factors. Propensity score matching(PSM) analysis was performed to minimize bias. Results: A total of 165 patients were divided into two groups based on the presence(SPH, n = 76) or absence(non-SPH, n = 89) of SPH. Patients in the SPH group had longer operative time, more blood loss, and more advanced TNM stage than patients in the non-SPH group( P < 0.05). However, there were no significant differences in the postoperative 90-day mortality rate( n = 0), overall postoperative complications(47.4% vs. 41.6%, P = 0.455), Clavien-Dindo classification( P = 0.347), conversion to open surgery(9.2% vs. 6.7%, P = 0.557), or length of hospitalization(16 vs. 15 days, P = 0.203) between the SPH and non-SPH groups before PSM. Similar results were obtained after PSM. The 1-, 3-, and 5-year overall survival(OS) and recurrence-free survival rates in the SPH group were not significantly different from those in the non-SPH group both before and after PSM(log-rank P > 0.05). After PSM, alpha-fetoprotein(AFP) ≥ 400 μg/L [hazard ratio(HR) = 4.71, 95% confidence interval(CI): 2.69-8.25], ascites(HR = 2.18, 95% CI: 1.30-3.66), American Society of Anesthesiologists(ASA) classification(Ⅲ vs. Ⅱ)(HR = 2.13, 95% CI: 1.11-4.07) and tumor diameter > 5 cm(HR = 3.91, 95% CI: 2.02-7.56) independently predicted worse OS. Conclusions: LLR for patients with HCC complicated with SPH appears feasible at the price of increasing operative time and blood loss. AFP, ascites, ASA classification and tumor diameter may predict the prognosis of HCC complicated with SPH after LLR.
基金financialy supported by National Natural Science Foundation of China(Grants 22005298,22125903,51872283,22075279,22279137)Dalian Innovation Support Plan for High Level Talents(2019RT09)+3 种基金Dalian National Laboratory For Clean Energy(DNL),CAS,DNL Cooperation Fund,CAS(DNL201912,DNL201915,DNL202016,DNL202019),DICP(DICP I2020032)The Joint Fund of the Yulin University and the Dalian National Laboratory for Clean Energy(YLUDNL Fund 2021002,YLU-DNL Fund 2021009)Suzhou University Scientific Research Platform(2021XJPT07)China Postdoctoral Science Foundation(2019 M661141)
文摘The Li-ion capacitors(LICs)develop rapidly due to their double-high features of high-energy density and high-power density.However,the relative low capacity of cathode and sluggish kinetics of anode seriously impede the development of LICs.Herein,the precisely pore-engineered and heteroatomtailored defective hierarchical porous carbons(DHPCs)as large-capacity cathode and high-rate anode to construct high-performance dual-carbon LICs have been developed.The DHPCs are prepared based on triple-activation mechanisms by direct pyrolysis of sustainable lignin with urea to generate the interconnected hierarchical porous structure and plentiful heteroatominduced defects.Benefiting from these advanced merits,DHPCs show the well-matched high capacity and fast kinetics of both cathode and anode,exhibiting large capacities,superior rate capability and long-term lifespan.Both experimental and computational results demonstrate the strong synergistic effect of pore and dopants for Li storage.Consequently,the assembled dual-carbon LIC exhibits high voltage of 4.5 V,high-energy density of 208 Wh kg^(−1),ultrahigh power density of 53.4 kW kg^(−1)and almost zerodecrement cycling lifetime.Impressively,the full device with high mass loading of 9.4 mg cm^(−2)on cathode still outputs high-energy density of 187 Wh kg^(−1),demonstrative of their potential as electrode materials for high-performance electrochemical devices.
基金supported by the National Natural Science Foundation of China (Grant No.71861015)the Humanities and Social Science Foundation of the Ministry of Education of China (Grant No.18YJA630047)the Distinguished Young Scholar Talent of Jiangxi Province (Grant No.20192BCBL23008).
文摘In this paper,a stable two-sided matching(TSM)method considering the matching intention of agents under a hesitant fuzzy environment is proposed.The method uses a hesitant fuzzy element(HFE)as its basis.First,the HFE preference matrix is transformed into the normalized HFE preference matrix.On this basis,the distance and the projection of the normalized HFEs on positive and negative ideal solutions are calculated.Then,the normalized HFEs are transformed into agent satisfactions.Considering the stable matching constraints,a multiobjective programming model with the objective of maximizing the satisfactions of two-sided agents is constructed.Based on the agent satisfaction matrix,the matching intention matrix of two-sided agents is built.According to the agent satisfaction matrix and matching intention matrix,the comprehensive satisfaction matrix is set up.Furthermore,the multiobjective programming model based on satisfactions is transformed into a multiobjective programming model based on comprehensive satisfactions.Using the G-S algorithm,the multiobjective programming model based on comprehensive satisfactions is solved,and then the best TSM scheme is obtained.Finally,a terminal distribution example is used to verify the feasibility and effectiveness of the proposed method.
文摘As the fundamental problem in the computer vision area,image matching has wide applications in pose estimation,3D reconstruction,image retrieval,etc.Suffering from the influence of external factors,the process of image matching using classical local detectors,e.g.,scale-invariant feature transform(SIFT),and the outlier filtering approaches,e.g.,Random sample consensus(RANSAC),show high computation speed and pool robustness under changing illumination and viewpoints conditions,while image matching approaches with deep learning strategy(such as HardNet,OANet)display reliable achievements in large-scale datasets with challenging scenes.However,the past learning-based approaches are limited to the distinction and quality of the dataset and the training strategy in the image-matching approaches.As an extension of the previous conference paper,this paper proposes an accurate and robust image matching approach using fewer training data in an end-to-end manner,which could be used to estimate the pose error This research first proposes a novel dataset cleaning and construction strategy to eliminate the noise and improve the training efficiency;Secondly,a novel loss named quadratic hinge triplet loss(QHT)is proposed to gather more effective and stable feature matching;Thirdly,in the outlier filtering process,the stricter OANet and bundle adjustment are applied for judging samples by adding the epipolar distance constraint and triangulation constraint to generate more outstanding matches;Finally,to recall the matching pairs,dynamic guided matching is used and then submit the inliers after the PyRANSAC process.Multiple evaluation metrics are used and reported in the 1st place in the Track1 of CVPR Image-Matching Challenge Workshop.The results show that the proposed method has advanced performance in large-scale and challenging Phototourism benchmark.
文摘BACKGROUND Patients with atrial fibrillation(AF)and prior stroke history have a high risk of cardiovascular events despite anticoagulation therapy.It is unclear whether catheter ablation(CA)has further benefits in these patients.METHODS AF patients with a previous history of stroke or systemic embolism(SE)from the prospective Chinese Atrial Fibrillation Registry study between August 2011 and December 2020 were included in the analysis.Patients were matched in a 1:1 ratio to CA or medical treatment(MT)based on propensity score.The primary outcome was a composite of all-cause death or ischemic stroke(IS)/SE.RESULTS During a total of 4.1±2.3 years of follow-up,the primary outcome occurred in 111 patients in the CA group(3.3 per 100 person-years)and in 229 patients in the MT group(5.7 per 100 person-years).The CA group had a lower risk of the primary outcome compared to the MT group[hazard ratio(HR)=0.59,95%CI:0.47–0.74,P<0.001].There was a significant decreasing risk of all-cause mortality(HR=0.43,95%CI:0.31–0.61,P<0.001),IS/SE(HR=0.73,95%CI:0.54–0.97,P=0.033),cardiovascular mortality(HR=0.32,95%CI:0.19–0.54,P<0.001)and AF recurrence(HR=0.33,95%CI:0.30–0.37,P<0.001)in the CA group compared to that in the MT group.Sensitivity analysis generated consistent results when adjusting for time-dependent usage of anticoagulants.CONCLUSIONS In AF patients with a prior stroke history,CA was associated with a lower combined risk of all-cause death or IS/SE.Further clinical trials are warranted to confirm the benefits of CA in these patients.
文摘Graph pattern matching(GPM)can be used to mine the key information in graphs.Exact GPM is one of the most commonly used methods among all the GPM-related methods,which aims to exactly find all subgraphs for a given query graph in a data graph.The exact GPM has been widely used in biological data analyses,social network analyses and other fields.In this paper,the applications of the exact GPM were first introduced,and the research progress of the exact GPM was summarized.Then,the related algorithms were introduced in detail,and the experiments on the state-of-the-art exact GPM algorithms were conducted to compare their performance.Based on the experimental results,the applicable scenarios of the algorithms were pointed out.New research opportunities in this area were proposed.