Purpose:Many science,technology and innovation(STI)resources are attached with several different labels.To assign automatically the resulting labels to an interested instance,many approaches with good performance on t...Purpose:Many science,technology and innovation(STI)resources are attached with several different labels.To assign automatically the resulting labels to an interested instance,many approaches with good performance on the benchmark datasets have been proposed for multi-label classification task in the literature.Furthermore,several open-source tools implementing these approaches have also been developed.However,the characteristics of real-world multi-label patent and publication datasets are not completely in line with those of benchmark ones.Therefore,the main purpose of this paper is to evaluate comprehensively seven multi-label classification methods on real-world datasets.Research limitations:Three real-world datasets differ in the following aspects:statement,data quality,and purposes.Additionally,open-source tools designed for multi-label classification also have intrinsic differences in their approaches for data processing and feature selection,which in turn impacts the performance of a multi-label classification approach.In the near future,we will enhance experimental precision and reinforce the validity of conclusions by employing more rigorous control over variables through introducing expanded parameter settings.Practical implications:The observed Macro F1 and Micro F1 scores on real-world datasets typically fall short of those achieved on benchmark datasets,underscoring the complexity of real-world multi-label classification tasks.Approaches leveraging deep learning techniques offer promising solutions by accommodating the hierarchical relationships and interdependencies among labels.With ongoing enhancements in deep learning algorithms and large-scale models,it is expected that the efficacy of multi-label classification tasks will be significantly improved,reaching a level of practical utility in the foreseeable future.Originality/value:(1)Seven multi-label classification methods are comprehensively compared on three real-world datasets.(2)The TextCNN and TextRCNN models perform better on small-scale datasets with more complex hierarchical structure of labels and more balanced document-label distribution.(3)The MLkNN method works better on the larger-scale dataset with more unbalanced document-label distribution.展开更多
This paper proposed a method to generate semi-experimental biomedical datasets based on full-wave simulation software.The system noise such as antenna port couplings is fully considered in the proposed datasets,which ...This paper proposed a method to generate semi-experimental biomedical datasets based on full-wave simulation software.The system noise such as antenna port couplings is fully considered in the proposed datasets,which is more realistic than synthetical datasets.In this paper,datasets containing different shapes are constructed based on the relative permittivities of human tissues.Then,a back-propagation scheme is used to obtain the rough reconstructions,which will be fed into a U-net convolutional neural network(CNN)to recover the high-resolution images.Numerical results show that the network trained on the datasets generated by the proposed method can obtain satisfying reconstruction results and is promising to be applied in real-time biomedical imaging.展开更多
This article delves into the analysis of performance and utilization of Support Vector Machines (SVMs) for the critical task of forest fire detection using image datasets. With the increasing threat of forest fires to...This article delves into the analysis of performance and utilization of Support Vector Machines (SVMs) for the critical task of forest fire detection using image datasets. With the increasing threat of forest fires to ecosystems and human settlements, the need for rapid and accurate detection systems is of utmost importance. SVMs, renowned for their strong classification capabilities, exhibit proficiency in recognizing patterns associated with fire within images. By training on labeled data, SVMs acquire the ability to identify distinctive attributes associated with fire, such as flames, smoke, or alterations in the visual characteristics of the forest area. The document thoroughly examines the use of SVMs, covering crucial elements like data preprocessing, feature extraction, and model training. It rigorously evaluates parameters such as accuracy, efficiency, and practical applicability. The knowledge gained from this study aids in the development of efficient forest fire detection systems, enabling prompt responses and improving disaster management. Moreover, the correlation between SVM accuracy and the difficulties presented by high-dimensional datasets is carefully investigated, demonstrated through a revealing case study. The relationship between accuracy scores and the different resolutions used for resizing the training datasets has also been discussed in this article. These comprehensive studies result in a definitive overview of the difficulties faced and the potential sectors requiring further improvement and focus.展开更多
BACKGROUND Real-world data on tofacitinib(TOF)covering a period of more than 1 year for a sufficient number of Asian patients with ulcerative colitis(UC)are scarce.AIM To investigate the long-term efficacy and safety ...BACKGROUND Real-world data on tofacitinib(TOF)covering a period of more than 1 year for a sufficient number of Asian patients with ulcerative colitis(UC)are scarce.AIM To investigate the long-term efficacy and safety of TOF treatment for UC,including clinical issues.METHODS We performed a retrospective single-center observational analysis of 111 UC patients administered TOF at Hyogo Medical University as a tertiary inflammatory bowel disease center.All consecutive UC patients who received TOF between May 2018 and February 2020 were enrolled.Patients were followed up until August 2020.The primary outcome was the clinical response rate at week 8.Secondary outcomes included clinical remission at week 8,cumulative persistence rate of TOF administration,colectomy-free survival,relapse after tapering of TOF and predictors of clinical response at week 8 and week 48.RESULTS The clinical response and remission rates were 66.3%and 50.5%at week 8,and 47.1%and 43.5%at week 48,respectively.The overall cumulative clinical remission rate was 61.7%at week 48 and history of anti-tumor necrosis factor-alpha(TNF-α)agents use had no influence(P=0.25).The cumulative TOF persistence rate at week 48 was significantly lower in patients without clinical remission than in those with remission at week 8(30.9%vs 88.1%;P<0.001).Baseline partial Mayo Score was significantly lower in responders vs non-responders at week 8(odds ratio:0.61,95%confidence interval:0.45-0.82,P=0.001).Relapse occurred in 45.7%of patients after TOF tapering,and 85.7%of patients responded within 4 wk after re-increase.All 6 patients with herpes zoster(HZ)developed the infection after achieving remission by TOF.CONCLUSION TOF was more effective in UC patients with mild activity at baseline and its efficacy was not affected by previous treatment with anti-TNF-αagents.Most relapsed patients responded again after re-increase of TOF and nearly half relapsed after tapering off TOF.Special attention is needed for tapering and HZ.展开更多
BACKGROUND Although chronic erosive gastritis(CEG)is common,its clinical characteristics have not been fully elucidated.The lack of consensus regarding its treatment has resulted in varied treatment regimens.AIM To ex...BACKGROUND Although chronic erosive gastritis(CEG)is common,its clinical characteristics have not been fully elucidated.The lack of consensus regarding its treatment has resulted in varied treatment regimens.AIM To explore the clinical characteristics,treatment patterns,and short-term outcomes in CEG patients in China.METHODS We recruited patients with chronic non-atrophic or mild-to-moderate atrophic gastritis with erosion based on endoscopy and pathology.Patients and treating physicians completed a questionnaire regarding history,endoscopic findings,and treatment plans as well as a follow-up questionnaire to investigate changes in symptoms after 4 wk of treatment.RESULTS Three thousand five hundred sixty-three patients from 42 centers across 24 cities in China were included.Epigastric pain(68.0%),abdominal distension(62.6%),and postprandial fullness(47.5%)were the most common presenting symptoms.Gastritis was classified as chronic non-atrophic in 69.9%of patients.Among those with erosive lesions,72.1%of patients had lesions in the antrum,51.0%had multiple lesions,and 67.3%had superficial flat lesions.In patients with epigastric pain,the combination of a mucosal protective agent(MPA)and proton pump inhibitor was more effective.For those with postprandial fullness,acid regurgitation,early satiety,or nausea,a MPA appeared more promising.CONCLUSION CEG is a multifactorial disease which is common in Asian patients and has non-specific symptoms.Gastroscopy may play a major role in its detection and diagnosis.Treatment should be individualized based on symptom profile.展开更多
Rock mass quality serves as a vital index for predicting the stability and safety status of rock tunnel faces.In tunneling practice,the rock mass quality is often assessed via a combination of qualitative and quantita...Rock mass quality serves as a vital index for predicting the stability and safety status of rock tunnel faces.In tunneling practice,the rock mass quality is often assessed via a combination of qualitative and quantitative parameters.However,due to the harsh on-site construction conditions,it is rather difficult to obtain some of the evaluation parameters which are essential for the rock mass quality prediction.In this study,a novel improved Swin Transformer is proposed to detect,segment,and quantify rock mass characteristic parameters such as water leakage,fractures,weak interlayers.The site experiment results demonstrate that the improved Swin Transformer achieves optimal segmentation results and achieving accuracies of 92%,81%,and 86%for water leakage,fractures,and weak interlayers,respectively.A multisource rock tunnel face characteristic(RTFC)dataset includes 11 parameters for predicting rock mass quality is established.Considering the limitations in predictive performance of incomplete evaluation parameters exist in this dataset,a novel tree-augmented naive Bayesian network(BN)is proposed to address the challenge of the incomplete dataset and achieved a prediction accuracy of 88%.In comparison with other commonly used Machine Learning models the proposed BN-based approach proved an improved performance on predicting the rock mass quality with the incomplete dataset.By utilizing the established BN,a further sensitivity analysis is conducted to quantitatively evaluate the importance of the various parameters,results indicate that the rock strength and fractures parameter exert the most significant influence on rock mass quality.展开更多
Recently,automotive intrusion detection systems(IDSs)have emerged as promising defense approaches to counter attacks on in-vehicle networks(IVNs).However,the effectiveness of IDSs relies heavily on the quality of the ...Recently,automotive intrusion detection systems(IDSs)have emerged as promising defense approaches to counter attacks on in-vehicle networks(IVNs).However,the effectiveness of IDSs relies heavily on the quality of the datasets used for training and evaluation.Despite the availability of several datasets for automotive IDSs,there has been a lack of comprehensive analysis focusing on assessing these datasets.This paper aims to address the need for dataset assessment in the context of automotive IDSs.It proposes qualitative and quantitative metrics that are independent of specific automotive IDSs,to evaluate the quality of datasets.These metrics take into consideration various aspects such as dataset description,collection environment,and attack complexity.This paper evaluates eight commonly used datasets for automotive IDSs using the proposed metrics.The evaluation reveals biases in the datasets,particularly in terms of limited contexts and lack of diversity.Additionally,it highlights that the attacks in the datasets were mostly injected without considering normal behaviors,which poses challenges for training and evaluating machine learning-based IDSs.This paper emphasizes the importance of addressing the identified limitations in existing datasets to improve the performance and adaptability of automotive IDSs.The proposed metrics can serve as valuable guidelines for researchers and practitioners in selecting and constructing high-quality datasets for automotive security applications.Finally,this paper presents the requirements for high-quality datasets,including the need for representativeness,diversity,and balance.展开更多
Phishing attacks pose a significant security threat by masquerading as trustworthy entities to steal sensitive information,a problem that persists despite user awareness.This study addresses the pressing issue of phis...Phishing attacks pose a significant security threat by masquerading as trustworthy entities to steal sensitive information,a problem that persists despite user awareness.This study addresses the pressing issue of phishing attacks on websites and assesses the performance of three prominent Machine Learning(ML)models—Artificial Neural Networks(ANN),Convolutional Neural Networks(CNN),and Long Short-Term Memory(LSTM)—utilizing authentic datasets sourced from Kaggle and Mendeley repositories.Extensive experimentation and analysis reveal that the CNN model achieves a better accuracy of 98%.On the other hand,LSTM shows the lowest accuracy of 96%.These findings underscore the potential of ML techniques in enhancing phishing detection systems and bolstering cybersecurity measures against evolving phishing tactics,offering a promising avenue for safeguarding sensitive information and online security.展开更多
Objective:Little progress has been made in recent years using first-line chemotherapy,including gemcitabine combined with nab-paclitaxel,FOLFIRINOX,and NALIRIFOX,for advanced pancreatic adenocarcinoma(APC).In addition...Objective:Little progress has been made in recent years using first-line chemotherapy,including gemcitabine combined with nab-paclitaxel,FOLFIRINOX,and NALIRIFOX,for advanced pancreatic adenocarcinoma(APC).In addition,the optimal second-line chemotherapy regimen has not been determined.This study aimed to compare the effectiveness of different types of second-line chemotherapy for APC.Methods:Patients with APC who received first-line treatment from January 2008 to January 2021 were considered eligible for this retrospective analysis.The primary and secondary endpoints were overall survival(OS)and progression-free survival(PFS),respectively.Results:Four hundred and thirty-seven and 617 patients were treated with 5-fluorouracil-and gemcitabine-based chemotherapy as first-line treatment,respectively.Demographic and clinical features,except age and liver metastasis,were comparable between the two groups(P<0.05).The median OS was 8.8 and 7.8 months in patients who received a 5-fluorouracil-and gemcitabine-based combined regimen for first-line therapy,respectively(HR=1.244,95%CI=1.090–1.419;P<0.001).The median OS was 5.6 and 1.9 months in patients who received second-line chemotherapy and supportive care,respectively(HR=0.766,95%CI=0.677–0.867;P<0.001).The median PFS was not significantly differently between gemcitabine or 5-fluorouracil monotherapy and combination therapy.Conclusions:A 5-fluorouracil-or gemcitabine-based combined regimen was shown to be as effective as a single 5-fluorouracil or gemcitabine regimen as second-line therapy for patients with APC.展开更多
To address the shortage of public datasets for customs X-ray images of contraband and the difficulties in deploying trained models in engineering applications,a method has been proposed that employs the Extract-Transf...To address the shortage of public datasets for customs X-ray images of contraband and the difficulties in deploying trained models in engineering applications,a method has been proposed that employs the Extract-Transform-Load(ETL)approach to create an X-ray dataset of contraband items.Initially,X-ray scatter image data is collected and cleaned.Using Kafka message queues and the Elasticsearch(ES)distributed search engine,the data is transmitted in real-time to cloud servers.Subsequently,contraband data is annotated using a combination of neural networks and manual methods to improve annotation efficiency and implemented mean hash algorithm for quick image retrieval.The method of integrating targets with backgrounds has enhanced the X-ray contraband image data,increasing the number of positive samples.Finally,an Airport Customs X-ray dataset(ACXray)compatible with customs business scenarios has been constructed,featuring an increased number of positive contraband samples.Experimental tests using three datasets to train the Mask Region-based Convolutional Neural Network(Mask R-CNN)algorithm and tested on 400 real customs images revealed that the recognition accuracy of algorithms trained with Security Inspection X-ray(SIXray)and Occluded Prohibited Items X-ray(OPIXray)decreased by 16.3%and 15.1%,respectively,while the ACXray dataset trained algorithm’s accuracy was almost unaffected.This indicates that the ACXray dataset-trained algorithm possesses strong generalization capabilities and is more suitable for customs detection scenarios.展开更多
The scientific goal of the Anninghe seismic array is to investigate the detailed geometry of the Anninghe fault and the velocity structure of the fault zone.This 2D seismic array is composed of 161 stations forming su...The scientific goal of the Anninghe seismic array is to investigate the detailed geometry of the Anninghe fault and the velocity structure of the fault zone.This 2D seismic array is composed of 161 stations forming sub-rectangular geometry along the Anninghe fault,which covers 50 km and 150 km in the fault normal and strike directions,respectively,with~5 km intervals.The data were collected between June 2020 and June 2021,with some level of temporal gaps.Two types of instruments,i.e.QS-05A and SmartSolo,are used in this array.Data quality and examples of seismograms are provided in this paper.After the data protection period ends(expected in June 2024),researchers can request a dataset from the National Earthquake Science Data Center.展开更多
As the realm of enterprise-level conversational AI continues to evolve, it becomes evident that while generalized Large Language Models (LLMs) like GPT-3.5 bring remarkable capabilities, they also bring forth formidab...As the realm of enterprise-level conversational AI continues to evolve, it becomes evident that while generalized Large Language Models (LLMs) like GPT-3.5 bring remarkable capabilities, they also bring forth formidable challenges. These models, honed on vast and diverse datasets, have undoubtedly pushed the boundaries of natural language understanding and generation. However, they often stumble when faced with the intricate demands of nuanced enterprise applications. This research advocates for a strategic paradigm shift, urging enterprises to embrace a fine-tuning approach as a means to optimize conversational AI. While generalized LLMs are linguistic marvels, their inability to cater to the specific needs of businesses across various industries poses a critical challenge. This strategic shift involves empowering enterprises to seamlessly integrate their own datasets into LLMs, a process that extends beyond linguistic enhancement. The core concept of this approach centers on customization, enabling businesses to fine-tune the AI’s functionality to fit precisely within their unique business landscapes. By immersing the LLM in industry-specific documents, customer interaction records, internal reports, and regulatory guidelines, the AI transcends its generic capabilities to become a sophisticated conversational partner aligned with the intricacies of the enterprise’s domain. The transformative potential of this fine-tuning approach cannot be overstated. It enables a transition from a universal AI solution to a highly customizable tool. The AI evolves from being a linguistic powerhouse to a contextually aware, industry-savvy assistant. As a result, it not only responds with linguistic accuracy but also with depth, relevance, and resonance, significantly elevating user experiences and operational efficiency. In the subsequent sections, this paper delves into the intricacies of fine-tuning, exploring the multifaceted challenges and abundant opportunities it presents. It addresses the technical intricacies of data integration, ethical considerations surrounding data usage, and the broader implications for the future of enterprise AI. The journey embarked upon in this research holds the potential to redefine the role of conversational AI in enterprises, ushering in an era where AI becomes a dynamic, deeply relevant, and highly effective tool, empowering businesses to excel in an ever-evolving digital landscape.展开更多
Fine-grained recognition of ships based on remote sensing images is crucial to safeguarding maritime rights and interests and maintaining national security.Currently,with the emergence of massive high-resolution multi...Fine-grained recognition of ships based on remote sensing images is crucial to safeguarding maritime rights and interests and maintaining national security.Currently,with the emergence of massive high-resolution multi-modality images,the use of multi-modality images for fine-grained recognition has become a promising technology.Fine-grained recognition of multi-modality images imposes higher requirements on the dataset samples.The key to the problem is how to extract and fuse the complementary features of multi-modality images to obtain more discriminative fusion features.The attention mechanism helps the model to pinpoint the key information in the image,resulting in a significant improvement in the model’s performance.In this paper,a dataset for fine-grained recognition of ships based on visible and near-infrared multi-modality remote sensing images has been proposed first,named Dataset for Multimodal Fine-grained Recognition of Ships(DMFGRS).It includes 1,635 pairs of visible and near-infrared remote sensing images divided into 20 categories,collated from digital orthophotos model provided by commercial remote sensing satellites.DMFGRS provides two types of annotation format files,as well as segmentation mask images corresponding to the ship targets.Then,a Multimodal Information Cross-Enhancement Network(MICE-Net)fusing features of visible and near-infrared remote sensing images,has been proposed.In the network,a dual-branch feature extraction and fusion module has been designed to obtain more expressive features.The Feature Cross Enhancement Module(FCEM)achieves the fusion enhancement of the two modal features by making the channel attention and spatial attention work cross-functionally on the feature map.A benchmark is established by evaluating state-of-the-art object recognition algorithms on DMFGRS.MICE-Net conducted experiments on DMFGRS,and the precision,recall,mAP0.5 and mAP0.5:0.95 reached 87%,77.1%,83.8%and 63.9%,respectively.Extensive experiments demonstrate that the proposed MICE-Net has more excellent performance on DMFGRS.Built on lightweight network YOLO,the model has excellent generalizability,and thus has good potential for application in real-life scenarios.展开更多
Tofacitinib is an oral small-molecule Janus kinase(JAK)inhibitor that preferentially inhibits JAK1 and JAK3.Its efficacy in inducing and maintaining remission in ulcerative colitis(UC)as well as its safety profile has...Tofacitinib is an oral small-molecule Janus kinase(JAK)inhibitor that preferentially inhibits JAK1 and JAK3.Its efficacy in inducing and maintaining remission in ulcerative colitis(UC)as well as its safety profile has been demonstrated in multicenter,randomized,double-blind,placebo-controlled trials.Additionally,real-world studies evaluating the effectiveness and adverse effects of tofacitinib have been conducted,affirming its clinical efficacy in moderate-to-severe UC.展开更多
Named entity recognition(NER)is a fundamental task of information extraction(IE),and it has attracted considerable research attention in recent years.The abundant annotated English NER datasets have significantly prom...Named entity recognition(NER)is a fundamental task of information extraction(IE),and it has attracted considerable research attention in recent years.The abundant annotated English NER datasets have significantly promoted the NER research in the English field.By contrast,much fewer efforts are made to the Chinese NER research,especially in the scientific domain,due to the scarcity of Chinese NER datasets.To alleviate this problem,we present aChinese scientificNER dataset–SciCN,which contains entity annotations of titles and abstracts derived from 3,500 scientific papers.We manually annotate a total of 62,059 entities,and these entities are classified into six types.Compared to English scientific NER datasets,SciCN has a larger scale and is more diverse,for it not only contains more paper abstracts but these abstracts are derived from more research fields.To investigate the properties of SciCN and provide baselines for future research,we adapt a number of previous state-of-theart Chinese NER models to evaluate SciCN.Experimental results show that SciCN is more challenging than other Chinese NER datasets.In addition,previous studies have proven the effectiveness of using lexicons to enhance Chinese NER models.Motivated by this fact,we provide a scientific domain-specific lexicon.Validation results demonstrate that our lexicon delivers better performance gains than lexicons of other domains.We hope that the SciCN dataset and the lexicon will enable us to benchmark the NER task regarding the Chinese scientific domain and make progress for future research.The dataset and lexicon are available at:https://github.com/yangjingla/SciCN.git.展开更多
Battery pack capacity estimation under real-world operating conditions is important for battery performance optimization and health management,contributing to the reliability and longevity of batterypowered systems.Ho...Battery pack capacity estimation under real-world operating conditions is important for battery performance optimization and health management,contributing to the reliability and longevity of batterypowered systems.However,complex operating conditions,coupling cell-to-cell inconsistency,and limited labeled data pose great challenges to accurate and robust battery pack capacity estimation.To address these issues,this paper proposes a hierarchical data-driven framework aimed at enhancing the training of machine learning models with fewer labeled data.Unlike traditional data-driven methods that lack interpretability,the hierarchical data-driven framework unveils the“mechanism”of the black box inside the data-driven framework by splitting the final estimation target into cell-level and pack-level intermediate targets.A generalized feature matrix is devised without requiring all cell voltages,significantly reducing the computational cost and memory resources.The generated intermediate target labels and the corresponding features are hierarchically employed to enhance the training of two machine learning models,effectively alleviating the difficulty of learning the relationship from all features due to fewer labeled data and addressing the dilemma of requiring extensive labeled data for accurate estimation.Using only 10%of degradation data,the proposed framework outperforms the state-of-the-art battery pack capacity estimation methods,achieving mean absolute percentage errors of 0.608%,0.601%,and 1.128%for three battery packs whose degradation load profiles represent real-world operating conditions.Its high accuracy,adaptability,and robustness indicate the potential in different application scenarios,which is promising for reducing laborious and expensive aging experiments at the pack level and facilitating the development of battery technology.展开更多
Intrusion detection systems(IDS)are essential in the field of cybersecurity because they protect networks from a wide range of online threats.The goal of this research is to meet the urgent need for small-footprint,hi...Intrusion detection systems(IDS)are essential in the field of cybersecurity because they protect networks from a wide range of online threats.The goal of this research is to meet the urgent need for small-footprint,highly-adaptable Network Intrusion Detection Systems(NIDS)that can identify anomalies.The NSL-KDD dataset is used in the study;it is a sizable collection comprising 43 variables with the label’s“attack”and“level.”It proposes a novel approach to intrusion detection based on the combination of channel attention and convolutional neural networks(CNN).Furthermore,this dataset makes it easier to conduct a thorough assessment of the suggested intrusion detection strategy.Furthermore,maintaining operating efficiency while improving detection accuracy is the primary goal of this work.Moreover,typical NIDS examines both risky and typical behavior using a variety of techniques.On the NSL-KDD dataset,our CNN-based approach achieves an astounding 99.728%accuracy rate when paired with channel attention.Compared to previous approaches such as ensemble learning,CNN,RBM(Boltzmann machine),ANN,hybrid auto-encoders with CNN,MCNN,and ANN,and adaptive algorithms,our solution significantly improves intrusion detection performance.Moreover,the results highlight the effectiveness of our suggested method in improving intrusion detection precision,signifying a noteworthy advancement in this field.Subsequent efforts will focus on strengthening and expanding our approach in order to counteract growing cyberthreats and adjust to changing network circumstances.展开更多
Real-world engineering design problems with complex objective functions under some constraints are relatively difficult problems to solve.Such design problems are widely experienced in many engineering fields,such as ...Real-world engineering design problems with complex objective functions under some constraints are relatively difficult problems to solve.Such design problems are widely experienced in many engineering fields,such as industry,automotive,construction,machinery,and interdisciplinary research.However,there are established optimization techniques that have shown effectiveness in addressing these types of issues.This research paper gives a comparative study of the implementation of seventeen new metaheuristic methods in order to optimize twelve distinct engineering design issues.The algorithms used in the study are listed as:transient search optimization(TSO),equilibrium optimizer(EO),grey wolf optimizer(GWO),moth-flame optimization(MFO),whale optimization algorithm(WOA),slimemould algorithm(SMA),harris hawks optimization(HHO),chimp optimization algorithm(COA),coot optimization algorithm(COOT),multi-verse optimization(MVO),arithmetic optimization algorithm(AOA),aquila optimizer(AO),sine cosine algorithm(SCA),smell agent optimization(SAO),and seagull optimization algorithm(SOA),pelican optimization algorithm(POA),and coati optimization algorithm(CA).As far as we know,there is no comparative analysis of recent and popular methods against the concrete conditions of real-world engineering problems.Hence,a remarkable research guideline is presented in the study for researchersworking in the fields of engineering and artificial intelligence,especiallywhen applying the optimization methods that have emerged recently.Future research can rely on this work for a literature search on comparisons of metaheuristic optimization methods in real-world problems under similar conditions.展开更多
For the first time, this article introduces a LiDAR Point Clouds Dataset of Ships composed of both collected and simulated data to address the scarcity of LiDAR data in maritime applications. The collected data are ac...For the first time, this article introduces a LiDAR Point Clouds Dataset of Ships composed of both collected and simulated data to address the scarcity of LiDAR data in maritime applications. The collected data are acquired using specialized maritime LiDAR sensors in both inland waterways and wide-open ocean environments. The simulated data is generated by placing a ship in the LiDAR coordinate system and scanning it with a redeveloped Blensor that emulates the operation of a LiDAR sensor equipped with various laser beams. Furthermore,we also render point clouds for foggy and rainy weather conditions. To describe a realistic shipping environment, a dynamic tail wave is modeled by iterating the wave elevation of each point in a time series. Finally, networks serving small objects are migrated to ship applications by feeding our dataset. The positive effect of simulated data is described in object detection experiments, and the negative impact of tail waves as noise is verified in single-object tracking experiments. The Dataset is available at https://github.com/zqy411470859/ship_dataset.展开更多
This study investigates the composition,abundance,and basic biological parameters of krill in Prydz Bay,Antarctic Peninsula and Amundsen Sea by analyzing samples and environmental data from the Chinese National Antarc...This study investigates the composition,abundance,and basic biological parameters of krill in Prydz Bay,Antarctic Peninsula and Amundsen Sea by analyzing samples and environmental data from the Chinese National Antarctic Research Expeditions conducted between 2009/2010 and 2019/2020.The predominant krill species observed were Euphausia superba,Euphausia crystallorophias,and Thysanoessa macrura.T.macrura,although the most widespread,exhibited the lowest mean abundance 9.96 ind:(1000 m^(-3))and biomass 0.31 g(1000 m^(-3)),predominantly found in low-latitude regions of the Amundsen Sea while E.crystallorophias was most concentrated in polynyas of Prydz Bay.E.superba,with an average abundance of 34.05 ind(1000 m^(-3))and biomass of 11.80 g:(1000 m^(-3)),was mainly distributed in the Antarctic Peninsula and Prydz Bay.This study also identified regional variations in mean body length and frequency distributions of kril.The relationship between krill body length and wet weight followed a power-law pattern.Regional differences were observed in the relationship between krill abundance,biomass,and environmental factors with varying correlations.In the Amundsen Sea,no significant correlation was found between krill abundance and environmental factors.Notably,E.crystallorophias in Prydz Bay demonstrated a significant positive correlation with chlorophyll a concentration,while T.macrura abundance and biomass in the Antarctic Peninsula exhibited a significant negative correlation with ice-free days.The findings contribute valuable regional data on krll distribution,abundance,and biomass in the Southern Ocean,serving as foundational information for the conservation of the Southern Ocean ecosystem and Antarctic krill fishery management on a circumpolar scale.展开更多
基金the Natural Science Foundation of China(Grant Numbers 72074014 and 72004012).
文摘Purpose:Many science,technology and innovation(STI)resources are attached with several different labels.To assign automatically the resulting labels to an interested instance,many approaches with good performance on the benchmark datasets have been proposed for multi-label classification task in the literature.Furthermore,several open-source tools implementing these approaches have also been developed.However,the characteristics of real-world multi-label patent and publication datasets are not completely in line with those of benchmark ones.Therefore,the main purpose of this paper is to evaluate comprehensively seven multi-label classification methods on real-world datasets.Research limitations:Three real-world datasets differ in the following aspects:statement,data quality,and purposes.Additionally,open-source tools designed for multi-label classification also have intrinsic differences in their approaches for data processing and feature selection,which in turn impacts the performance of a multi-label classification approach.In the near future,we will enhance experimental precision and reinforce the validity of conclusions by employing more rigorous control over variables through introducing expanded parameter settings.Practical implications:The observed Macro F1 and Micro F1 scores on real-world datasets typically fall short of those achieved on benchmark datasets,underscoring the complexity of real-world multi-label classification tasks.Approaches leveraging deep learning techniques offer promising solutions by accommodating the hierarchical relationships and interdependencies among labels.With ongoing enhancements in deep learning algorithms and large-scale models,it is expected that the efficacy of multi-label classification tasks will be significantly improved,reaching a level of practical utility in the foreseeable future.Originality/value:(1)Seven multi-label classification methods are comprehensively compared on three real-world datasets.(2)The TextCNN and TextRCNN models perform better on small-scale datasets with more complex hierarchical structure of labels and more balanced document-label distribution.(3)The MLkNN method works better on the larger-scale dataset with more unbalanced document-label distribution.
基金National Natural Science Foundation of China(No.61971036)Fundamental Research Funds for the Central Universities(No.2023CX01011)Beijing Nova Program(No.20230484361)。
文摘This paper proposed a method to generate semi-experimental biomedical datasets based on full-wave simulation software.The system noise such as antenna port couplings is fully considered in the proposed datasets,which is more realistic than synthetical datasets.In this paper,datasets containing different shapes are constructed based on the relative permittivities of human tissues.Then,a back-propagation scheme is used to obtain the rough reconstructions,which will be fed into a U-net convolutional neural network(CNN)to recover the high-resolution images.Numerical results show that the network trained on the datasets generated by the proposed method can obtain satisfying reconstruction results and is promising to be applied in real-time biomedical imaging.
文摘This article delves into the analysis of performance and utilization of Support Vector Machines (SVMs) for the critical task of forest fire detection using image datasets. With the increasing threat of forest fires to ecosystems and human settlements, the need for rapid and accurate detection systems is of utmost importance. SVMs, renowned for their strong classification capabilities, exhibit proficiency in recognizing patterns associated with fire within images. By training on labeled data, SVMs acquire the ability to identify distinctive attributes associated with fire, such as flames, smoke, or alterations in the visual characteristics of the forest area. The document thoroughly examines the use of SVMs, covering crucial elements like data preprocessing, feature extraction, and model training. It rigorously evaluates parameters such as accuracy, efficiency, and practical applicability. The knowledge gained from this study aids in the development of efficient forest fire detection systems, enabling prompt responses and improving disaster management. Moreover, the correlation between SVM accuracy and the difficulties presented by high-dimensional datasets is carefully investigated, demonstrated through a revealing case study. The relationship between accuracy scores and the different resolutions used for resizing the training datasets has also been discussed in this article. These comprehensive studies result in a definitive overview of the difficulties faced and the potential sectors requiring further improvement and focus.
文摘BACKGROUND Real-world data on tofacitinib(TOF)covering a period of more than 1 year for a sufficient number of Asian patients with ulcerative colitis(UC)are scarce.AIM To investigate the long-term efficacy and safety of TOF treatment for UC,including clinical issues.METHODS We performed a retrospective single-center observational analysis of 111 UC patients administered TOF at Hyogo Medical University as a tertiary inflammatory bowel disease center.All consecutive UC patients who received TOF between May 2018 and February 2020 were enrolled.Patients were followed up until August 2020.The primary outcome was the clinical response rate at week 8.Secondary outcomes included clinical remission at week 8,cumulative persistence rate of TOF administration,colectomy-free survival,relapse after tapering of TOF and predictors of clinical response at week 8 and week 48.RESULTS The clinical response and remission rates were 66.3%and 50.5%at week 8,and 47.1%and 43.5%at week 48,respectively.The overall cumulative clinical remission rate was 61.7%at week 48 and history of anti-tumor necrosis factor-alpha(TNF-α)agents use had no influence(P=0.25).The cumulative TOF persistence rate at week 48 was significantly lower in patients without clinical remission than in those with remission at week 8(30.9%vs 88.1%;P<0.001).Baseline partial Mayo Score was significantly lower in responders vs non-responders at week 8(odds ratio:0.61,95%confidence interval:0.45-0.82,P=0.001).Relapse occurred in 45.7%of patients after TOF tapering,and 85.7%of patients responded within 4 wk after re-increase.All 6 patients with herpes zoster(HZ)developed the infection after achieving remission by TOF.CONCLUSION TOF was more effective in UC patients with mild activity at baseline and its efficacy was not affected by previous treatment with anti-TNF-αagents.Most relapsed patients responded again after re-increase of TOF and nearly half relapsed after tapering off TOF.Special attention is needed for tapering and HZ.
基金the National Key Clinical Specialty Construction Project,No.ZK108000CAMS Innovation Fund for Medical Sciences,No.2021-I2M-C&T-A-001 and No.2022-I2M-C&T-B-012.
文摘BACKGROUND Although chronic erosive gastritis(CEG)is common,its clinical characteristics have not been fully elucidated.The lack of consensus regarding its treatment has resulted in varied treatment regimens.AIM To explore the clinical characteristics,treatment patterns,and short-term outcomes in CEG patients in China.METHODS We recruited patients with chronic non-atrophic or mild-to-moderate atrophic gastritis with erosion based on endoscopy and pathology.Patients and treating physicians completed a questionnaire regarding history,endoscopic findings,and treatment plans as well as a follow-up questionnaire to investigate changes in symptoms after 4 wk of treatment.RESULTS Three thousand five hundred sixty-three patients from 42 centers across 24 cities in China were included.Epigastric pain(68.0%),abdominal distension(62.6%),and postprandial fullness(47.5%)were the most common presenting symptoms.Gastritis was classified as chronic non-atrophic in 69.9%of patients.Among those with erosive lesions,72.1%of patients had lesions in the antrum,51.0%had multiple lesions,and 67.3%had superficial flat lesions.In patients with epigastric pain,the combination of a mucosal protective agent(MPA)and proton pump inhibitor was more effective.For those with postprandial fullness,acid regurgitation,early satiety,or nausea,a MPA appeared more promising.CONCLUSION CEG is a multifactorial disease which is common in Asian patients and has non-specific symptoms.Gastroscopy may play a major role in its detection and diagnosis.Treatment should be individualized based on symptom profile.
基金supported by the National Natural Science Foundation of China(Nos.52279107 and 52379106)the Qingdao Guoxin Jiaozhou Bay Second Submarine Tunnel Co.,Ltd.,the Academician and Expert Workstation of Yunnan Province(No.202205AF150015)the Science and Technology Innovation Project of YCIC Group Co.,Ltd.(No.YCIC-YF-2022-15)。
文摘Rock mass quality serves as a vital index for predicting the stability and safety status of rock tunnel faces.In tunneling practice,the rock mass quality is often assessed via a combination of qualitative and quantitative parameters.However,due to the harsh on-site construction conditions,it is rather difficult to obtain some of the evaluation parameters which are essential for the rock mass quality prediction.In this study,a novel improved Swin Transformer is proposed to detect,segment,and quantify rock mass characteristic parameters such as water leakage,fractures,weak interlayers.The site experiment results demonstrate that the improved Swin Transformer achieves optimal segmentation results and achieving accuracies of 92%,81%,and 86%for water leakage,fractures,and weak interlayers,respectively.A multisource rock tunnel face characteristic(RTFC)dataset includes 11 parameters for predicting rock mass quality is established.Considering the limitations in predictive performance of incomplete evaluation parameters exist in this dataset,a novel tree-augmented naive Bayesian network(BN)is proposed to address the challenge of the incomplete dataset and achieved a prediction accuracy of 88%.In comparison with other commonly used Machine Learning models the proposed BN-based approach proved an improved performance on predicting the rock mass quality with the incomplete dataset.By utilizing the established BN,a further sensitivity analysis is conducted to quantitatively evaluate the importance of the various parameters,results indicate that the rock strength and fractures parameter exert the most significant influence on rock mass quality.
基金supported in part by the 2021 Autonomous Driving Development Innovation Project of the Ministry of Science and ICT,‘Development of Technology for Security and Ultra-High-Speed Integrity of the Next-Generation Internal Net-Work of Autonomous Vehicles’(No.2021-0-01348)and in part by the National Research Foundation of Korea(NRF)grant funded by the Korean Government Ministry of Science and ICT(MSIT)under Grant NRF-2021R1A2C2014428.
文摘Recently,automotive intrusion detection systems(IDSs)have emerged as promising defense approaches to counter attacks on in-vehicle networks(IVNs).However,the effectiveness of IDSs relies heavily on the quality of the datasets used for training and evaluation.Despite the availability of several datasets for automotive IDSs,there has been a lack of comprehensive analysis focusing on assessing these datasets.This paper aims to address the need for dataset assessment in the context of automotive IDSs.It proposes qualitative and quantitative metrics that are independent of specific automotive IDSs,to evaluate the quality of datasets.These metrics take into consideration various aspects such as dataset description,collection environment,and attack complexity.This paper evaluates eight commonly used datasets for automotive IDSs using the proposed metrics.The evaluation reveals biases in the datasets,particularly in terms of limited contexts and lack of diversity.Additionally,it highlights that the attacks in the datasets were mostly injected without considering normal behaviors,which poses challenges for training and evaluating machine learning-based IDSs.This paper emphasizes the importance of addressing the identified limitations in existing datasets to improve the performance and adaptability of automotive IDSs.The proposed metrics can serve as valuable guidelines for researchers and practitioners in selecting and constructing high-quality datasets for automotive security applications.Finally,this paper presents the requirements for high-quality datasets,including the need for representativeness,diversity,and balance.
文摘Phishing attacks pose a significant security threat by masquerading as trustworthy entities to steal sensitive information,a problem that persists despite user awareness.This study addresses the pressing issue of phishing attacks on websites and assesses the performance of three prominent Machine Learning(ML)models—Artificial Neural Networks(ANN),Convolutional Neural Networks(CNN),and Long Short-Term Memory(LSTM)—utilizing authentic datasets sourced from Kaggle and Mendeley repositories.Extensive experimentation and analysis reveal that the CNN model achieves a better accuracy of 98%.On the other hand,LSTM shows the lowest accuracy of 96%.These findings underscore the potential of ML techniques in enhancing phishing detection systems and bolstering cybersecurity measures against evolving phishing tactics,offering a promising avenue for safeguarding sensitive information and online security.
基金This work was supported by the National Key Research and Development Program of China(Grant No.2021YFA1201100)the National Natural Science Foundation of China(Grant No.82072657).
文摘Objective:Little progress has been made in recent years using first-line chemotherapy,including gemcitabine combined with nab-paclitaxel,FOLFIRINOX,and NALIRIFOX,for advanced pancreatic adenocarcinoma(APC).In addition,the optimal second-line chemotherapy regimen has not been determined.This study aimed to compare the effectiveness of different types of second-line chemotherapy for APC.Methods:Patients with APC who received first-line treatment from January 2008 to January 2021 were considered eligible for this retrospective analysis.The primary and secondary endpoints were overall survival(OS)and progression-free survival(PFS),respectively.Results:Four hundred and thirty-seven and 617 patients were treated with 5-fluorouracil-and gemcitabine-based chemotherapy as first-line treatment,respectively.Demographic and clinical features,except age and liver metastasis,were comparable between the two groups(P<0.05).The median OS was 8.8 and 7.8 months in patients who received a 5-fluorouracil-and gemcitabine-based combined regimen for first-line therapy,respectively(HR=1.244,95%CI=1.090–1.419;P<0.001).The median OS was 5.6 and 1.9 months in patients who received second-line chemotherapy and supportive care,respectively(HR=0.766,95%CI=0.677–0.867;P<0.001).The median PFS was not significantly differently between gemcitabine or 5-fluorouracil monotherapy and combination therapy.Conclusions:A 5-fluorouracil-or gemcitabine-based combined regimen was shown to be as effective as a single 5-fluorouracil or gemcitabine regimen as second-line therapy for patients with APC.
基金supported by the National Natural Science Foundation of China(Grant No.51605069).
文摘To address the shortage of public datasets for customs X-ray images of contraband and the difficulties in deploying trained models in engineering applications,a method has been proposed that employs the Extract-Transform-Load(ETL)approach to create an X-ray dataset of contraband items.Initially,X-ray scatter image data is collected and cleaned.Using Kafka message queues and the Elasticsearch(ES)distributed search engine,the data is transmitted in real-time to cloud servers.Subsequently,contraband data is annotated using a combination of neural networks and manual methods to improve annotation efficiency and implemented mean hash algorithm for quick image retrieval.The method of integrating targets with backgrounds has enhanced the X-ray contraband image data,increasing the number of positive samples.Finally,an Airport Customs X-ray dataset(ACXray)compatible with customs business scenarios has been constructed,featuring an increased number of positive contraband samples.Experimental tests using three datasets to train the Mask Region-based Convolutional Neural Network(Mask R-CNN)algorithm and tested on 400 real customs images revealed that the recognition accuracy of algorithms trained with Security Inspection X-ray(SIXray)and Occluded Prohibited Items X-ray(OPIXray)decreased by 16.3%and 15.1%,respectively,while the ACXray dataset trained algorithm’s accuracy was almost unaffected.This indicates that the ACXray dataset-trained algorithm possesses strong generalization capabilities and is more suitable for customs detection scenarios.
基金supported by the National Key Research and Development Program of China(No.2018YFC1503401).
文摘The scientific goal of the Anninghe seismic array is to investigate the detailed geometry of the Anninghe fault and the velocity structure of the fault zone.This 2D seismic array is composed of 161 stations forming sub-rectangular geometry along the Anninghe fault,which covers 50 km and 150 km in the fault normal and strike directions,respectively,with~5 km intervals.The data were collected between June 2020 and June 2021,with some level of temporal gaps.Two types of instruments,i.e.QS-05A and SmartSolo,are used in this array.Data quality and examples of seismograms are provided in this paper.After the data protection period ends(expected in June 2024),researchers can request a dataset from the National Earthquake Science Data Center.
文摘As the realm of enterprise-level conversational AI continues to evolve, it becomes evident that while generalized Large Language Models (LLMs) like GPT-3.5 bring remarkable capabilities, they also bring forth formidable challenges. These models, honed on vast and diverse datasets, have undoubtedly pushed the boundaries of natural language understanding and generation. However, they often stumble when faced with the intricate demands of nuanced enterprise applications. This research advocates for a strategic paradigm shift, urging enterprises to embrace a fine-tuning approach as a means to optimize conversational AI. While generalized LLMs are linguistic marvels, their inability to cater to the specific needs of businesses across various industries poses a critical challenge. This strategic shift involves empowering enterprises to seamlessly integrate their own datasets into LLMs, a process that extends beyond linguistic enhancement. The core concept of this approach centers on customization, enabling businesses to fine-tune the AI’s functionality to fit precisely within their unique business landscapes. By immersing the LLM in industry-specific documents, customer interaction records, internal reports, and regulatory guidelines, the AI transcends its generic capabilities to become a sophisticated conversational partner aligned with the intricacies of the enterprise’s domain. The transformative potential of this fine-tuning approach cannot be overstated. It enables a transition from a universal AI solution to a highly customizable tool. The AI evolves from being a linguistic powerhouse to a contextually aware, industry-savvy assistant. As a result, it not only responds with linguistic accuracy but also with depth, relevance, and resonance, significantly elevating user experiences and operational efficiency. In the subsequent sections, this paper delves into the intricacies of fine-tuning, exploring the multifaceted challenges and abundant opportunities it presents. It addresses the technical intricacies of data integration, ethical considerations surrounding data usage, and the broader implications for the future of enterprise AI. The journey embarked upon in this research holds the potential to redefine the role of conversational AI in enterprises, ushering in an era where AI becomes a dynamic, deeply relevant, and highly effective tool, empowering businesses to excel in an ever-evolving digital landscape.
文摘Fine-grained recognition of ships based on remote sensing images is crucial to safeguarding maritime rights and interests and maintaining national security.Currently,with the emergence of massive high-resolution multi-modality images,the use of multi-modality images for fine-grained recognition has become a promising technology.Fine-grained recognition of multi-modality images imposes higher requirements on the dataset samples.The key to the problem is how to extract and fuse the complementary features of multi-modality images to obtain more discriminative fusion features.The attention mechanism helps the model to pinpoint the key information in the image,resulting in a significant improvement in the model’s performance.In this paper,a dataset for fine-grained recognition of ships based on visible and near-infrared multi-modality remote sensing images has been proposed first,named Dataset for Multimodal Fine-grained Recognition of Ships(DMFGRS).It includes 1,635 pairs of visible and near-infrared remote sensing images divided into 20 categories,collated from digital orthophotos model provided by commercial remote sensing satellites.DMFGRS provides two types of annotation format files,as well as segmentation mask images corresponding to the ship targets.Then,a Multimodal Information Cross-Enhancement Network(MICE-Net)fusing features of visible and near-infrared remote sensing images,has been proposed.In the network,a dual-branch feature extraction and fusion module has been designed to obtain more expressive features.The Feature Cross Enhancement Module(FCEM)achieves the fusion enhancement of the two modal features by making the channel attention and spatial attention work cross-functionally on the feature map.A benchmark is established by evaluating state-of-the-art object recognition algorithms on DMFGRS.MICE-Net conducted experiments on DMFGRS,and the precision,recall,mAP0.5 and mAP0.5:0.95 reached 87%,77.1%,83.8%and 63.9%,respectively.Extensive experiments demonstrate that the proposed MICE-Net has more excellent performance on DMFGRS.Built on lightweight network YOLO,the model has excellent generalizability,and thus has good potential for application in real-life scenarios.
文摘Tofacitinib is an oral small-molecule Janus kinase(JAK)inhibitor that preferentially inhibits JAK1 and JAK3.Its efficacy in inducing and maintaining remission in ulcerative colitis(UC)as well as its safety profile has been demonstrated in multicenter,randomized,double-blind,placebo-controlled trials.Additionally,real-world studies evaluating the effectiveness and adverse effects of tofacitinib have been conducted,affirming its clinical efficacy in moderate-to-severe UC.
基金This research was supported by the National Key Research and Development Program[2020YFB1006302].
文摘Named entity recognition(NER)is a fundamental task of information extraction(IE),and it has attracted considerable research attention in recent years.The abundant annotated English NER datasets have significantly promoted the NER research in the English field.By contrast,much fewer efforts are made to the Chinese NER research,especially in the scientific domain,due to the scarcity of Chinese NER datasets.To alleviate this problem,we present aChinese scientificNER dataset–SciCN,which contains entity annotations of titles and abstracts derived from 3,500 scientific papers.We manually annotate a total of 62,059 entities,and these entities are classified into six types.Compared to English scientific NER datasets,SciCN has a larger scale and is more diverse,for it not only contains more paper abstracts but these abstracts are derived from more research fields.To investigate the properties of SciCN and provide baselines for future research,we adapt a number of previous state-of-theart Chinese NER models to evaluate SciCN.Experimental results show that SciCN is more challenging than other Chinese NER datasets.In addition,previous studies have proven the effectiveness of using lexicons to enhance Chinese NER models.Motivated by this fact,we provide a scientific domain-specific lexicon.Validation results demonstrate that our lexicon delivers better performance gains than lexicons of other domains.We hope that the SciCN dataset and the lexicon will enable us to benchmark the NER task regarding the Chinese scientific domain and make progress for future research.The dataset and lexicon are available at:https://github.com/yangjingla/SciCN.git.
基金supported by the National Outstanding Youth Science Fund Project of National Natural Science Foundation of China[Grant No.52222708]the Natural Science Foundation of Beijing Municipality[Grant No.3212033]。
文摘Battery pack capacity estimation under real-world operating conditions is important for battery performance optimization and health management,contributing to the reliability and longevity of batterypowered systems.However,complex operating conditions,coupling cell-to-cell inconsistency,and limited labeled data pose great challenges to accurate and robust battery pack capacity estimation.To address these issues,this paper proposes a hierarchical data-driven framework aimed at enhancing the training of machine learning models with fewer labeled data.Unlike traditional data-driven methods that lack interpretability,the hierarchical data-driven framework unveils the“mechanism”of the black box inside the data-driven framework by splitting the final estimation target into cell-level and pack-level intermediate targets.A generalized feature matrix is devised without requiring all cell voltages,significantly reducing the computational cost and memory resources.The generated intermediate target labels and the corresponding features are hierarchically employed to enhance the training of two machine learning models,effectively alleviating the difficulty of learning the relationship from all features due to fewer labeled data and addressing the dilemma of requiring extensive labeled data for accurate estimation.Using only 10%of degradation data,the proposed framework outperforms the state-of-the-art battery pack capacity estimation methods,achieving mean absolute percentage errors of 0.608%,0.601%,and 1.128%for three battery packs whose degradation load profiles represent real-world operating conditions.Its high accuracy,adaptability,and robustness indicate the potential in different application scenarios,which is promising for reducing laborious and expensive aging experiments at the pack level and facilitating the development of battery technology.
基金The authors would like to thank Princess Nourah bint Abdulrahman University for funding this project through the Researchers Supporting Project(PNURSP2023R319)this research was funded by the Prince Sultan University,Riyadh,Saudi Arabia.
文摘Intrusion detection systems(IDS)are essential in the field of cybersecurity because they protect networks from a wide range of online threats.The goal of this research is to meet the urgent need for small-footprint,highly-adaptable Network Intrusion Detection Systems(NIDS)that can identify anomalies.The NSL-KDD dataset is used in the study;it is a sizable collection comprising 43 variables with the label’s“attack”and“level.”It proposes a novel approach to intrusion detection based on the combination of channel attention and convolutional neural networks(CNN).Furthermore,this dataset makes it easier to conduct a thorough assessment of the suggested intrusion detection strategy.Furthermore,maintaining operating efficiency while improving detection accuracy is the primary goal of this work.Moreover,typical NIDS examines both risky and typical behavior using a variety of techniques.On the NSL-KDD dataset,our CNN-based approach achieves an astounding 99.728%accuracy rate when paired with channel attention.Compared to previous approaches such as ensemble learning,CNN,RBM(Boltzmann machine),ANN,hybrid auto-encoders with CNN,MCNN,and ANN,and adaptive algorithms,our solution significantly improves intrusion detection performance.Moreover,the results highlight the effectiveness of our suggested method in improving intrusion detection precision,signifying a noteworthy advancement in this field.Subsequent efforts will focus on strengthening and expanding our approach in order to counteract growing cyberthreats and adjust to changing network circumstances.
文摘Real-world engineering design problems with complex objective functions under some constraints are relatively difficult problems to solve.Such design problems are widely experienced in many engineering fields,such as industry,automotive,construction,machinery,and interdisciplinary research.However,there are established optimization techniques that have shown effectiveness in addressing these types of issues.This research paper gives a comparative study of the implementation of seventeen new metaheuristic methods in order to optimize twelve distinct engineering design issues.The algorithms used in the study are listed as:transient search optimization(TSO),equilibrium optimizer(EO),grey wolf optimizer(GWO),moth-flame optimization(MFO),whale optimization algorithm(WOA),slimemould algorithm(SMA),harris hawks optimization(HHO),chimp optimization algorithm(COA),coot optimization algorithm(COOT),multi-verse optimization(MVO),arithmetic optimization algorithm(AOA),aquila optimizer(AO),sine cosine algorithm(SCA),smell agent optimization(SAO),and seagull optimization algorithm(SOA),pelican optimization algorithm(POA),and coati optimization algorithm(CA).As far as we know,there is no comparative analysis of recent and popular methods against the concrete conditions of real-world engineering problems.Hence,a remarkable research guideline is presented in the study for researchersworking in the fields of engineering and artificial intelligence,especiallywhen applying the optimization methods that have emerged recently.Future research can rely on this work for a literature search on comparisons of metaheuristic optimization methods in real-world problems under similar conditions.
基金supported by the National Natural Science Foundation of China (62173103)the Fundamental Research Funds for the Central Universities of China (3072022JC0402,3072022JC0403)。
文摘For the first time, this article introduces a LiDAR Point Clouds Dataset of Ships composed of both collected and simulated data to address the scarcity of LiDAR data in maritime applications. The collected data are acquired using specialized maritime LiDAR sensors in both inland waterways and wide-open ocean environments. The simulated data is generated by placing a ship in the LiDAR coordinate system and scanning it with a redeveloped Blensor that emulates the operation of a LiDAR sensor equipped with various laser beams. Furthermore,we also render point clouds for foggy and rainy weather conditions. To describe a realistic shipping environment, a dynamic tail wave is modeled by iterating the wave elevation of each point in a time series. Finally, networks serving small objects are migrated to ship applications by feeding our dataset. The positive effect of simulated data is described in object detection experiments, and the negative impact of tail waves as noise is verified in single-object tracking experiments. The Dataset is available at https://github.com/zqy411470859/ship_dataset.
基金supported by Marine S&T Fund of Shandong Province for Qingdao Marine Science and Technology Center(Grant no.2022QNLM030002-1)National Natural Science Foundation of China(Grant no.42276238)National Polar Special Program“Impact and Response of Antarctic Seas to Climate Change”(Grant no.IRASCC 01-02-01D)and Taishan Scholars Program.
文摘This study investigates the composition,abundance,and basic biological parameters of krill in Prydz Bay,Antarctic Peninsula and Amundsen Sea by analyzing samples and environmental data from the Chinese National Antarctic Research Expeditions conducted between 2009/2010 and 2019/2020.The predominant krill species observed were Euphausia superba,Euphausia crystallorophias,and Thysanoessa macrura.T.macrura,although the most widespread,exhibited the lowest mean abundance 9.96 ind:(1000 m^(-3))and biomass 0.31 g(1000 m^(-3)),predominantly found in low-latitude regions of the Amundsen Sea while E.crystallorophias was most concentrated in polynyas of Prydz Bay.E.superba,with an average abundance of 34.05 ind(1000 m^(-3))and biomass of 11.80 g:(1000 m^(-3)),was mainly distributed in the Antarctic Peninsula and Prydz Bay.This study also identified regional variations in mean body length and frequency distributions of kril.The relationship between krill body length and wet weight followed a power-law pattern.Regional differences were observed in the relationship between krill abundance,biomass,and environmental factors with varying correlations.In the Amundsen Sea,no significant correlation was found between krill abundance and environmental factors.Notably,E.crystallorophias in Prydz Bay demonstrated a significant positive correlation with chlorophyll a concentration,while T.macrura abundance and biomass in the Antarctic Peninsula exhibited a significant negative correlation with ice-free days.The findings contribute valuable regional data on krll distribution,abundance,and biomass in the Southern Ocean,serving as foundational information for the conservation of the Southern Ocean ecosystem and Antarctic krill fishery management on a circumpolar scale.