A large number of network security breaches in IoT networks have demonstrated the unreliability of current Network Intrusion Detection Systems(NIDSs).Consequently,network interruptions and loss of sensitive data have ...A large number of network security breaches in IoT networks have demonstrated the unreliability of current Network Intrusion Detection Systems(NIDSs).Consequently,network interruptions and loss of sensitive data have occurred,which led to an active research area for improving NIDS technologies.In an analysis of related works,it was observed that most researchers aim to obtain better classification results by using a set of untried combinations of Feature Reduction(FR)and Machine Learning(ML)techniques on NIDS datasets.However,these datasets are different in feature sets,attack types,and network design.Therefore,this paper aims to discover whether these techniques can be generalised across various datasets.Six ML models are utilised:a Deep Feed Forward(DFF),Convolutional Neural Network(CNN),Recurrent Neural Network(RNN),Decision Tree(DT),Logistic Regression(LR),and Naive Bayes(NB).The accuracy of three Feature Extraction(FE)algorithms is detected;Principal Component Analysis(PCA),Auto-encoder(AE),and Linear Discriminant Analysis(LDA),are evaluated using three benchmark datasets:UNSW-NB15,ToN-IoT and CSE-CIC-IDS2018.Although PCA and AE algorithms have been widely used,the determination of their optimal number of extracted dimensions has been overlooked.The results indicate that no clear FE method or ML model can achieve the best scores for all datasets.The optimal number of extracted dimensions has been identified for each dataset,and LDA degrades the performance of the ML models on two datasets.The variance is used to analyse the extracted dimensions of LDA and PCA.Finally,this paper concludes that the choice of datasets significantly alters the performance of the applied techniques.We believe that a universal(benchmark)feature set is needed to facilitate further advancement and progress of research in this field.展开更多
In minimally invasive surgery,endoscopes or laparoscopes equipped with miniature cameras and tools are used to enter the human body for therapeutic purposes through small incisions or natural cavities.However,in clini...In minimally invasive surgery,endoscopes or laparoscopes equipped with miniature cameras and tools are used to enter the human body for therapeutic purposes through small incisions or natural cavities.However,in clinical operating environments,endoscopic images often suffer from challenges such as low texture,uneven illumination,and non-rigid structures,which affect feature observation and extraction.This can severely impact surgical navigation or clinical diagnosis due to missing feature points in endoscopic images,leading to treatment and postoperative recovery issues for patients.To address these challenges,this paper introduces,for the first time,a Cross-Channel Multi-Modal Adaptive Spatial Feature Fusion(ASFF)module based on the lightweight architecture of EfficientViT.Additionally,a novel lightweight feature extraction and matching network based on attention mechanism is proposed.This network dynamically adjusts attention weights for cross-modal information from grayscale images and optical flow images through a dual-branch Siamese network.It extracts static and dynamic information features ranging from low-level to high-level,and from local to global,ensuring robust feature extraction across different widths,noise levels,and blur scenarios.Global and local matching are performed through a multi-level cascaded attention mechanism,with cross-channel attention introduced to simultaneously extract low-level and high-level features.Extensive ablation experiments and comparative studies are conducted on the HyperKvasir,EAD,M2caiSeg,CVC-ClinicDB,and UCL synthetic datasets.Experimental results demonstrate that the proposed network improves upon the baseline EfficientViT-B3 model by 75.4%in accuracy(Acc),while also enhancing runtime performance and storage efficiency.When compared with the complex DenseDescriptor feature extraction network,the difference in Acc is less than 7.22%,and IoU calculation results on specific datasets outperform complex dense models.Furthermore,this method increases the F1 score by 33.2%and accelerates runtime by 70.2%.It is noteworthy that the speed of CMMCAN surpasses that of comparative lightweight models,with feature extraction and matching performance comparable to existing complex models but with faster speed and higher cost-effectiveness.展开更多
Nowadays,ensuring thequality of networkserviceshas become increasingly vital.Experts are turning toknowledge graph technology,with a significant emphasis on entity extraction in the identification of device configurat...Nowadays,ensuring thequality of networkserviceshas become increasingly vital.Experts are turning toknowledge graph technology,with a significant emphasis on entity extraction in the identification of device configurations.This research paper presents a novel entity extraction method that leverages a combination of active learning and attention mechanisms.Initially,an improved active learning approach is employed to select the most valuable unlabeled samples,which are subsequently submitted for expert labeling.This approach successfully addresses the problems of isolated points and sample redundancy within the network configuration sample set.Then the labeled samples are utilized to train the model for network configuration entity extraction.Furthermore,the multi-head self-attention of the transformer model is enhanced by introducing the Adaptive Weighting method based on the Laplace mixture distribution.This enhancement enables the transformer model to dynamically adapt its focus to words in various positions,displaying exceptional adaptability to abnormal data and further elevating the accuracy of the proposed model.Through comparisons with Random Sampling(RANDOM),Maximum Normalized Log-Probability(MNLP),Least Confidence(LC),Token Entrop(TE),and Entropy Query by Bagging(EQB),the proposed method,Entropy Query by Bagging and Maximum Influence Active Learning(EQBMIAL),achieves comparable performance with only 40% of the samples on both datasets,while other algorithms require 50% of the samples.Furthermore,the entity extraction algorithm with the Adaptive Weighted Multi-head Attention mechanism(AW-MHA)is compared with BILSTM-CRF,Mutil_Attention-Bilstm-Crf,Deep_Neural_Model_NER and BERT_Transformer,achieving precision rates of 75.98% and 98.32% on the two datasets,respectively.Statistical tests demonstrate the statistical significance and effectiveness of the proposed algorithms in this paper.展开更多
Significant advancements have been achieved in road surface extraction based on high-resolution remote sensingimage processing. Most current methods rely on fully supervised learning, which necessitates enormous human...Significant advancements have been achieved in road surface extraction based on high-resolution remote sensingimage processing. Most current methods rely on fully supervised learning, which necessitates enormous humaneffort to label the image. Within this field, other research endeavors utilize weakly supervised methods. Theseapproaches aim to reduce the expenses associated with annotation by leveraging sparsely annotated data, such asscribbles. This paper presents a novel technique called a weakly supervised network using scribble-supervised andedge-mask (WSSE-net). This network is a three-branch network architecture, whereby each branch is equippedwith a distinct decoder module dedicated to road extraction tasks. One of the branches is dedicated to generatingedge masks using edge detection algorithms and optimizing road edge details. The other two branches supervise themodel’s training by employing scribble labels and spreading scribble information throughout the image. To addressthe historical flaw that created pseudo-labels that are not updated with network training, we use mixup to blendprediction results dynamically and continually update new pseudo-labels to steer network training. Our solutiondemonstrates efficient operation by simultaneously considering both edge-mask aid and dynamic pseudo-labelsupport. The studies are conducted on three separate road datasets, which consist primarily of high-resolutionremote-sensing satellite photos and drone images. The experimental findings suggest that our methodologyperforms better than advanced scribble-supervised approaches and specific traditional fully supervised methods.展开更多
When existing deep learning models are used for road extraction tasks from high-resolution images,they are easily affected by noise factors such as tree and building occlusion and complex backgrounds,resulting in inco...When existing deep learning models are used for road extraction tasks from high-resolution images,they are easily affected by noise factors such as tree and building occlusion and complex backgrounds,resulting in incomplete road extraction and low accuracy.We propose the introduction of spatial and channel attention modules to the convolutional neural network ConvNeXt.Then,ConvNeXt is used as the backbone network,which cooperates with the perceptual analysis network UPerNet,retains the detection head of the semantic segmentation,and builds a new model ConvNeXt-UPerNet to suppress noise interference.Training on the open-source DeepGlobe and CHN6-CUG datasets and introducing the DiceLoss on the basis of CrossEntropyLoss solves the problem of positive and negative sample imbalance.Experimental results show that the new network model can achieve the following performance on the DeepGlobe dataset:79.40%for precision(Pre),97.93% for accuracy(Acc),69.28% for intersection over union(IoU),and 83.56% for mean intersection over union(MIoU).On the CHN6-CUG dataset,the model achieves the respective values of 78.17%for Pre,97.63%for Acc,65.4% for IoU,and 81.46% for MIoU.Compared with other network models,the fused ConvNeXt-UPerNet model can extract road information better when faced with the influence of noise contained in high-resolution remote sensing images.It also achieves multiscale image feature information with unified perception,ultimately improving the generalization ability of deep learning technology in extracting complex roads from high-resolution remote sensing images.展开更多
This paper proposes a novel open set recognition method,the Spatial Distribution Feature Extraction Network(SDFEN),to address the problem of electromagnetic signal recognition in an open environment.The spatial distri...This paper proposes a novel open set recognition method,the Spatial Distribution Feature Extraction Network(SDFEN),to address the problem of electromagnetic signal recognition in an open environment.The spatial distribution feature extraction layer in SDFEN replaces convolutional output neural networks with the spatial distribution features that focus more on inter-sample information by incorporating class center vectors.The designed hybrid loss function considers both intra-class distance and inter-class distance,thereby enhancing the similarity among samples of the same class and increasing the dissimilarity between samples of different classes during training.Consequently,this method allows unknown classes to occupy a larger space in the feature space.This reduces the possibility of overlap with known class samples and makes the boundaries between known and unknown samples more distinct.Additionally,the feature comparator threshold can be used to reject unknown samples.For signal open set recognition,seven methods,including the proposed method,are applied to two kinds of electromagnetic signal data:modulation signal and real-world emitter.The experimental results demonstrate that the proposed method outperforms the other six methods overall in a simulated open environment.Specifically,compared to the state-of-the-art Openmax method,the novel method achieves up to 8.87%and 5.25%higher micro-F-measures,respectively.展开更多
Deep neural network-based relational extraction research has made significant progress in recent years,andit provides data support for many natural language processing downstream tasks such as building knowledgegraph,...Deep neural network-based relational extraction research has made significant progress in recent years,andit provides data support for many natural language processing downstream tasks such as building knowledgegraph,sentiment analysis and question-answering systems.However,previous studies ignored much unusedstructural information in sentences that could enhance the performance of the relation extraction task.Moreover,most existing dependency-based models utilize self-attention to distinguish the importance of context,whichhardly deals withmultiple-structure information.To efficiently leverage multiple structure information,this paperproposes a dynamic structure attention mechanism model based on textual structure information,which deeplyintegrates word embedding,named entity recognition labels,part of speech,dependency tree and dependency typeinto a graph convolutional network.Specifically,our model extracts text features of different structures from theinput sentence.Textual Structure information Graph Convolutional Networks employs the dynamic structureattention mechanism to learn multi-structure attention,effectively distinguishing important contextual features invarious structural information.In addition,multi-structure weights are carefully designed as amergingmechanismin the different structure attention to dynamically adjust the final attention.This paper combines these featuresand trains a graph convolutional network for relation extraction.We experiment on supervised relation extractiondatasets including SemEval 2010 Task 8,TACRED,TACREV,and Re-TACED,the result significantly outperformsthe previous.展开更多
Road extraction based on deep learning is one of hot spots of semantic segmentation in the past decade.In this work,we proposed a framework based on codec network for automatic road extraction from remote sensing imag...Road extraction based on deep learning is one of hot spots of semantic segmentation in the past decade.In this work,we proposed a framework based on codec network for automatic road extraction from remote sensing images.Firstly,a pre-trained ResNet34 was migrated to U-Net and its encoding structure was replaced to deepen the number of network layers,which reduces the error rate of road segmentation and the loss of details.Secondly,dilated convolution was used to connect the encoder and the decoder of network to expand the receptive field and retain more low-dimensional information of the image.Afterwards,the channel attention mechanism was used to select the information of the feature image obtained by up-sampling of the encoder,the weights of target features were optimized to enhance the features of target region and suppress the features of background and noise regions,and thus the feature extraction effect of the remote sensing image with complex background was optimized.Finally,an adaptive sigmoid loss function was proposed,which optimizes the imbalance between the road and the background,and makes the model reach the optimal solution.Experimental results show that compared with several semantic segmentation networks,the proposed method can greatly reduce the error rate of road segmentation and effectively improve the accuracy of road extraction from remote sensing images.展开更多
This study aims to identify a natural plant chemical with hypolipidemic effects that can be used to treat high cholesterol without adverse reactions.Through network pharmacology screening,it was found that Rosae Rugos...This study aims to identify a natural plant chemical with hypolipidemic effects that can be used to treat high cholesterol without adverse reactions.Through network pharmacology screening,it was found that Rosae Rugosae Flos(RF)flavonoids had potential therapeutic effects on hyperlipidemia and its mechanism of action was discussed.TCMSP and GeneCards databases were used to obtain active ingredients and disease targets.Venn diagrams were drawn to illustrate the findings.The interaction network diagram was created using Cytoscape 3.8.0 software.The PPI protein network was constructed using String.GO and KEGG enrichment analysis was performed using Metascape.The results revealed 2 active flavonoid ingredients and 60 potential targets in RF.The key targets,including CCL2,PPARG,and PPARA,were found to play a role in multiple pathways such as the AGE-RAGE signaling pathway,lipid and atherosclerosis,and cancer pathway in diabetic complications.The solvent extraction method was optimized for efficient flavonoid extraction based on network pharmacology prediction results.This was achieved through a single factor and orthogonal test,resulting in an optimum process with a reflux time of 1.5 h,a solid-liquid ratio of 1:13 g/mL,and an ethanol concentration of 50%.展开更多
A reasonable classification of deposits holds great significance for identifying prospecting targets and deploying exploration. The world ’s keen demand for lithium resources has expedited the discovery of numerous n...A reasonable classification of deposits holds great significance for identifying prospecting targets and deploying exploration. The world ’s keen demand for lithium resources has expedited the discovery of numerous novel lithium resources. Given the presence of varied classification criteria for lithium resources presently, this study further ascertained and classified the lithium resources according to their occurrence modes, obtaining 10 types and 5 subtypes of lithium deposits(resources) based on endogenetic and exogenetic factors. As indicated by surveys of Cenozoic exogenetic lithium deposits in China and abroad,the formation and distribution of the deposits are primarily determined by plate collision zones, their primary material sources are linked to the anatectic magmas in the deep oceanic crust, and they were formed primarily during the Miocene and Late Paleogene. The researchers ascertained that these deposits,especially those of the salt lake, geothermal, and volcanic deposit types, are formed by unique slightly acidic magmas, tend to migrate and accumulate toward low-lying areas, and display supernormal enrichment. However, the material sources of lithium deposits(resources) of the Neopaleozoic clay subtype and the deep brine type are yet to be further identified. Given the various types and complex origins of lithium deposits(resources), which were formed due to the interactions of multiple spheres, it is recommended that the mineralization of exogenetic lithium deposits(resources) be investigated by integrating tectono-geochemistry, paleoatmospheric circulation, and salinology. So far, industrialized lithium extraction is primarily achieved in lithium deposits of the salt lake, clay, and hard rock types. The lithium extraction employs different processes, with lithium extraction from salt lake-type lithium deposits proving the most energy-saving and cost-effective.展开更多
Copper was extracted from methylchlorosilane slurry residue by a direct hydrogen peroxide leaching method.A number of experimental parameters were analyzed to determine the extraction efficiency of copper.The extracti...Copper was extracted from methylchlorosilane slurry residue by a direct hydrogen peroxide leaching method.A number of experimental parameters were analyzed to determine the extraction efficiency of copper.The extraction efficiency of copper reached 98.5%under the optimal leaching conditions,such as the hydrogen peroxide concentration of 1.875 mol·L^(-1),the leaching temperature of 323 K,the liquid–solid ratio of 20 ml·g^(-1),and the stirring speed of 300 r·min^(-1).The leaching kinetics of the copper extraction process was then described by the shrinking core model.There were two stages.The first stage was controlled by chemical reactions,while the second stage was controlled by interface transfer and product layer diffusion.The activation energy and kinetic control equations were determined,as well as an explanation of the leaching mechanism of copper extraction based on kinetic analysis and materials characterization.Copper resources can be recovered from the methylchlorosilane slurry residue efficiently and inexpensively with the methods used in this study.展开更多
The continuous improvement of the cyber threat intelligence sharing mechanism provides new ideas to deal with Advanced Persistent Threats(APT).Extracting attack behaviors,i.e.,Tactics,Techniques,Procedures(TTP)from Cy...The continuous improvement of the cyber threat intelligence sharing mechanism provides new ideas to deal with Advanced Persistent Threats(APT).Extracting attack behaviors,i.e.,Tactics,Techniques,Procedures(TTP)from Cyber Threat Intelligence(CTI)can facilitate APT actors’profiling for an immediate response.However,it is difficult for traditional manual methods to analyze attack behaviors from cyber threat intelligence due to its heterogeneous nature.Based on the Adversarial Tactics,Techniques and Common Knowledge(ATT&CK)of threat behavior description,this paper proposes a threat behavioral knowledge extraction framework that integrates Heterogeneous Text Network(HTN)and Graph Convolutional Network(GCN)to solve this issue.It leverages the hierarchical correlation relationships of attack techniques and tactics in the ATT&CK to construct a text network of heterogeneous cyber threat intelligence.With the help of the Bidirectional EncoderRepresentation fromTransformers(BERT)pretraining model to analyze the contextual semantics of cyber threat intelligence,the task of threat behavior identification is transformed into a text classification task,which automatically extracts attack behavior in CTI,then identifies the malware and advanced threat actors.The experimental results show that F1 achieve 94.86%and 92.15%for the multi-label classification tasks of tactics and techniques.Extend the experiment to verify the method’s effectiveness in identifying the malware and threat actors in APT attacks.The F1 for malware and advanced threat actors identification task reached 98.45%and 99.48%,which are better than the benchmark model in the experiment and achieve state of the art.The model can effectivelymodel threat intelligence text data and acquire knowledge and experience migration by correlating implied features with a priori knowledge to compensate for insufficient sample data and improve the classification performance and recognition ability of threat behavior in text.展开更多
A network intrusion detection system is critical for cyber security against llegitimate attacks.In terms of feature perspectives,network traffic may include a variety of elements such as attack reference,attack type,a...A network intrusion detection system is critical for cyber security against llegitimate attacks.In terms of feature perspectives,network traffic may include a variety of elements such as attack reference,attack type,a subcategory of attack,host information,malicious scripts,etc.In terms of network perspectives,network traffic may contain an imbalanced number of harmful attacks when compared to normal traffic.It is challenging to identify a specific attack due to complex features and data imbalance issues.To address these issues,this paper proposes an Intrusion Detection System using transformer-based transfer learning for Imbalanced Network Traffic(IDS-INT).IDS-INT uses transformer-based transfer learning to learn feature interactions in both network feature representation and imbalanced data.First,detailed information about each type of attack is gathered from network interaction descriptions,which include network nodes,attack type,reference,host information,etc.Second,the transformer-based transfer learning approach is developed to learn detailed feature representation using their semantic anchors.Third,the Synthetic Minority Oversampling Technique(SMOTE)is implemented to balance abnormal traffic and detect minority attacks.Fourth,the Convolution Neural Network(CNN)model is designed to extract deep features from the balanced network traffic.Finally,the hybrid approach of the CNN-Long Short-Term Memory(CNN-LSTM)model is developed to detect different types of attacks from the deep features.Detailed experiments are conducted to test the proposed approach using three standard datasets,i.e.,UNsWNB15,CIC-IDS2017,and NSL-KDD.An explainable AI approach is implemented to interpret the proposed method and develop a trustable model.展开更多
Condensed and hydrolysable tannins are non-toxic natural polyphenols that are a commercial commodity industrialized for tanning hides to obtain leather and for a growing number of other industrial applications mainly ...Condensed and hydrolysable tannins are non-toxic natural polyphenols that are a commercial commodity industrialized for tanning hides to obtain leather and for a growing number of other industrial applications mainly to substitute petroleum-based products.They are a definite class of sustainable materials of the forestry industry.They have been in operation for hundreds of years to manufacture leather and now for a growing number of applications in a variety of other industries,such as wood adhesives,metal coating,pharmaceutical/medical applications and several others.This review presents the main sources,either already or potentially commercial of this forestry by-materials,their industrial and laboratory extraction systems,their systems of analysis with their advantages and drawbacks,be these methods so simple to even appear primitive but nonetheless of proven effectiveness,or very modern and instrumental.It constitutes a basic but essential summary of what is necessary to know of these sustainable materials.In doing so,the review highlights some of the main challenges that remain to be addressed to deliver the quality and economics of tannin supply necessary to fulfill the industrial production requirements for some materials-based uses.展开更多
An artificial neural network(ANN)method is introduced to predict drop size in two kinds of pulsed columns with small-scale data sets.After training,the deviation between calculate and experimental results are 3.8%and ...An artificial neural network(ANN)method is introduced to predict drop size in two kinds of pulsed columns with small-scale data sets.After training,the deviation between calculate and experimental results are 3.8%and 9.3%,respectively.Through ANN model,the influence of interfacial tension and pulsation intensity on the droplet diameter has been developed.Droplet size gradually increases with the increase of interfacial tension,and decreases with the increase of pulse intensity.It can be seen that the accuracy of ANN model in predicting droplet size outside the training set range is reach the same level as the accuracy of correlation obtained based on experiments within this range.For two kinds of columns,the drop size prediction deviations of ANN model are 9.6%and 18.5%and the deviations in correlations are 11%and 15%.展开更多
Chemical solvents instead of pure water being as hydraulic fracturing fluid could effectively increase permeability and improve clean methane extraction efficiency.However,pore-fracture variation features of lean coal...Chemical solvents instead of pure water being as hydraulic fracturing fluid could effectively increase permeability and improve clean methane extraction efficiency.However,pore-fracture variation features of lean coal synergistically affected by solvents have not been fully understood.Ultrasonic testing,nuclear magnetic resonance analysis,liquid phase mass spectrometry was adopted to comprehensively analyze pore-fracture change characteristics of lean coal treated by combined solvent(NMP and CS_(2)).Meanwhile,quantitative characterization of above changing properties was conducted using geometric fractal theory.Relationship model between permeability,fractal dimension and porosity were established.Results indicate that the end face fractures of coal are well developed after CS2and combined solvent treatments,of which,end face box-counting fractal dimensions range from 1.1227 to 1.4767.Maximum decreases in ultrasonic longitudinal wave velocity of coal affected by NMP,CS_(2)and combined solvent are 2.700%,20.521%,22.454%,respectively.Solvent treatments could lead to increasing amount of both mesopores and macropores.Decrease ratio of fractal dimension Dsis 0.259%–2.159%,while permeability increases ratio of NMR ranges from 0.1904 to 6.4486.Meanwhile,combined solvent could dissolve coal polar and non-polar small molecules and expand flow space.Results could provide reference for solvent selection and parameter optimization of permeability-enhancement technology.展开更多
To solve the problem of poor detection and limited application range of current intrusion detection methods,this paper attempts to use deep learning neural network technology to study a new type of intrusion detection...To solve the problem of poor detection and limited application range of current intrusion detection methods,this paper attempts to use deep learning neural network technology to study a new type of intrusion detection method.Hence,we proposed an intrusion detection algorithm based on convolutional neural network(CNN)and AdaBoost algorithm.This algorithm uses CNN to extract the characteristics of network traffic data,which is particularly suitable for the analysis of continuous and classified attack data.The AdaBoost algorithm is used to classify network attack data that improved the detection effect of unbalanced data classification.We adopt the UNSW-NB15 dataset to test of this algorithm in the PyCharm environment.The results show that the detection rate of algorithm is99.27%and the false positive rate is lower than 0.98%.Comparative analysis shows that this algorithm has advantages over existing methods in terms of detection rate and false positive rate for small proportion of attack data.展开更多
Lithium recovery from spent lithium-ion batteries(LIBs)have attracted extensive attention due to the skyrocketing price of lithium.The medium-temperature carbon reduction roasting was proposed to preferential selectiv...Lithium recovery from spent lithium-ion batteries(LIBs)have attracted extensive attention due to the skyrocketing price of lithium.The medium-temperature carbon reduction roasting was proposed to preferential selective extraction of lithium from spent Li-CoO_(2)(LCO)cathodes to overcome the incomplete recovery and loss of lithium during the recycling process.The LCO layered structure was destroyed and lithium was completely converted into water-soluble Li2CO_(3)under a suitable temperature to control the reduced state of the cobalt oxide.The Co metal agglomerates generated during medium-temperature carbon reduction roasting were broken by wet grinding and ultrasonic crushing to release the entrained lithium.The results showed that 99.10%of the whole lithium could be recovered as Li2CO_(3)with a purity of 99.55%.This work provided a new perspective on the preferentially selective extraction of lithium from spent lithium batteries.展开更多
Cultural relics line graphic serves as a crucial form of traditional artifact information documentation,which is a simple and intuitive product with low cost of displaying compared with 3D models.Dimensionality reduct...Cultural relics line graphic serves as a crucial form of traditional artifact information documentation,which is a simple and intuitive product with low cost of displaying compared with 3D models.Dimensionality reduction is undoubtedly necessary for line drawings.However,most existing methods for artifact drawing rely on the principles of orthographic projection that always cannot avoid angle occlusion and data overlapping while the surface of cultural relics is complex.Therefore,conformal mapping was introduced as a dimensionality reduction way to compensate for the limitation of orthographic projection.Based on the given criteria for assessing surface complexity,this paper proposed a three-dimensional feature guideline extraction method for complex cultural relic surfaces.A 2D and 3D combined factor that measured the importance of points on describing surface features,vertex weight,was designed.Then the selection threshold for feature guideline extraction was determined based on the differences between vertex weight and shape index distributions.The feasibility and stability were verified through experiments conducted on real cultural relic surface data.Results demonstrated the ability of the method to address the challenges associated with the automatic generation of line drawings for complex surfaces.The extraction method and the obtained results will be useful for line graphic drawing,displaying and propaganda of cultural relics.展开更多
Photovoltaic(PV)boards are a perfect way to create eco-friendly power from daylight.The defects in the PV panels are caused by various conditions;such defective PV panels need continuous monitoring.The recent developm...Photovoltaic(PV)boards are a perfect way to create eco-friendly power from daylight.The defects in the PV panels are caused by various conditions;such defective PV panels need continuous monitoring.The recent development of PV panel monitoring systems provides a modest and viable approach to monitoring and managing the condition of the PV plants.In general,conventional procedures are used to identify the faulty modules earlier and to avoid declines in power generation.The existing deep learning architectures provide the required output to predict the faulty PV panels with less accuracy and a more time-consuming process.To increase the accuracy and to reduce the processing time,a new Convolutional Neural Network(CNN)architecture is required.Hence,in the present work,a new Real-time Multi Variant Deep learning Model(RMVDM)architecture is proposed,and it extracts the image features and classifies the defects in PV panels quickly with high accuracy.The defects that arise in the PV panels are identified by the CNN based RMVDM using RGB images.The biggest difference between CNN and its predecessors is that CNN automatically extracts the image features without any help from a person.The technique is quantitatively assessed and compared with existing faulty PV board identification approaches on the large real-time dataset.The results show that 98%of the accuracy and recall values in the fault detection and classification process.展开更多
文摘A large number of network security breaches in IoT networks have demonstrated the unreliability of current Network Intrusion Detection Systems(NIDSs).Consequently,network interruptions and loss of sensitive data have occurred,which led to an active research area for improving NIDS technologies.In an analysis of related works,it was observed that most researchers aim to obtain better classification results by using a set of untried combinations of Feature Reduction(FR)and Machine Learning(ML)techniques on NIDS datasets.However,these datasets are different in feature sets,attack types,and network design.Therefore,this paper aims to discover whether these techniques can be generalised across various datasets.Six ML models are utilised:a Deep Feed Forward(DFF),Convolutional Neural Network(CNN),Recurrent Neural Network(RNN),Decision Tree(DT),Logistic Regression(LR),and Naive Bayes(NB).The accuracy of three Feature Extraction(FE)algorithms is detected;Principal Component Analysis(PCA),Auto-encoder(AE),and Linear Discriminant Analysis(LDA),are evaluated using three benchmark datasets:UNSW-NB15,ToN-IoT and CSE-CIC-IDS2018.Although PCA and AE algorithms have been widely used,the determination of their optimal number of extracted dimensions has been overlooked.The results indicate that no clear FE method or ML model can achieve the best scores for all datasets.The optimal number of extracted dimensions has been identified for each dataset,and LDA degrades the performance of the ML models on two datasets.The variance is used to analyse the extracted dimensions of LDA and PCA.Finally,this paper concludes that the choice of datasets significantly alters the performance of the applied techniques.We believe that a universal(benchmark)feature set is needed to facilitate further advancement and progress of research in this field.
基金This work was supported by Science and Technology Cooperation Special Project of Shijiazhuang(SJZZXA23005).
文摘In minimally invasive surgery,endoscopes or laparoscopes equipped with miniature cameras and tools are used to enter the human body for therapeutic purposes through small incisions or natural cavities.However,in clinical operating environments,endoscopic images often suffer from challenges such as low texture,uneven illumination,and non-rigid structures,which affect feature observation and extraction.This can severely impact surgical navigation or clinical diagnosis due to missing feature points in endoscopic images,leading to treatment and postoperative recovery issues for patients.To address these challenges,this paper introduces,for the first time,a Cross-Channel Multi-Modal Adaptive Spatial Feature Fusion(ASFF)module based on the lightweight architecture of EfficientViT.Additionally,a novel lightweight feature extraction and matching network based on attention mechanism is proposed.This network dynamically adjusts attention weights for cross-modal information from grayscale images and optical flow images through a dual-branch Siamese network.It extracts static and dynamic information features ranging from low-level to high-level,and from local to global,ensuring robust feature extraction across different widths,noise levels,and blur scenarios.Global and local matching are performed through a multi-level cascaded attention mechanism,with cross-channel attention introduced to simultaneously extract low-level and high-level features.Extensive ablation experiments and comparative studies are conducted on the HyperKvasir,EAD,M2caiSeg,CVC-ClinicDB,and UCL synthetic datasets.Experimental results demonstrate that the proposed network improves upon the baseline EfficientViT-B3 model by 75.4%in accuracy(Acc),while also enhancing runtime performance and storage efficiency.When compared with the complex DenseDescriptor feature extraction network,the difference in Acc is less than 7.22%,and IoU calculation results on specific datasets outperform complex dense models.Furthermore,this method increases the F1 score by 33.2%and accelerates runtime by 70.2%.It is noteworthy that the speed of CMMCAN surpasses that of comparative lightweight models,with feature extraction and matching performance comparable to existing complex models but with faster speed and higher cost-effectiveness.
基金supported by the National Key R&D Program of China(2019YFB2103202).
文摘Nowadays,ensuring thequality of networkserviceshas become increasingly vital.Experts are turning toknowledge graph technology,with a significant emphasis on entity extraction in the identification of device configurations.This research paper presents a novel entity extraction method that leverages a combination of active learning and attention mechanisms.Initially,an improved active learning approach is employed to select the most valuable unlabeled samples,which are subsequently submitted for expert labeling.This approach successfully addresses the problems of isolated points and sample redundancy within the network configuration sample set.Then the labeled samples are utilized to train the model for network configuration entity extraction.Furthermore,the multi-head self-attention of the transformer model is enhanced by introducing the Adaptive Weighting method based on the Laplace mixture distribution.This enhancement enables the transformer model to dynamically adapt its focus to words in various positions,displaying exceptional adaptability to abnormal data and further elevating the accuracy of the proposed model.Through comparisons with Random Sampling(RANDOM),Maximum Normalized Log-Probability(MNLP),Least Confidence(LC),Token Entrop(TE),and Entropy Query by Bagging(EQB),the proposed method,Entropy Query by Bagging and Maximum Influence Active Learning(EQBMIAL),achieves comparable performance with only 40% of the samples on both datasets,while other algorithms require 50% of the samples.Furthermore,the entity extraction algorithm with the Adaptive Weighted Multi-head Attention mechanism(AW-MHA)is compared with BILSTM-CRF,Mutil_Attention-Bilstm-Crf,Deep_Neural_Model_NER and BERT_Transformer,achieving precision rates of 75.98% and 98.32% on the two datasets,respectively.Statistical tests demonstrate the statistical significance and effectiveness of the proposed algorithms in this paper.
基金the National Natural Science Foundation of China(42001408,61806097).
文摘Significant advancements have been achieved in road surface extraction based on high-resolution remote sensingimage processing. Most current methods rely on fully supervised learning, which necessitates enormous humaneffort to label the image. Within this field, other research endeavors utilize weakly supervised methods. Theseapproaches aim to reduce the expenses associated with annotation by leveraging sparsely annotated data, such asscribbles. This paper presents a novel technique called a weakly supervised network using scribble-supervised andedge-mask (WSSE-net). This network is a three-branch network architecture, whereby each branch is equippedwith a distinct decoder module dedicated to road extraction tasks. One of the branches is dedicated to generatingedge masks using edge detection algorithms and optimizing road edge details. The other two branches supervise themodel’s training by employing scribble labels and spreading scribble information throughout the image. To addressthe historical flaw that created pseudo-labels that are not updated with network training, we use mixup to blendprediction results dynamically and continually update new pseudo-labels to steer network training. Our solutiondemonstrates efficient operation by simultaneously considering both edge-mask aid and dynamic pseudo-labelsupport. The studies are conducted on three separate road datasets, which consist primarily of high-resolutionremote-sensing satellite photos and drone images. The experimental findings suggest that our methodologyperforms better than advanced scribble-supervised approaches and specific traditional fully supervised methods.
基金This work was supported in part by the Key Project of Natural Science Research of Anhui Provincial Department of Education under Grant KJ2017A416in part by the Fund of National Sensor Network Engineering Technology Research Center(No.NSNC202103).
文摘When existing deep learning models are used for road extraction tasks from high-resolution images,they are easily affected by noise factors such as tree and building occlusion and complex backgrounds,resulting in incomplete road extraction and low accuracy.We propose the introduction of spatial and channel attention modules to the convolutional neural network ConvNeXt.Then,ConvNeXt is used as the backbone network,which cooperates with the perceptual analysis network UPerNet,retains the detection head of the semantic segmentation,and builds a new model ConvNeXt-UPerNet to suppress noise interference.Training on the open-source DeepGlobe and CHN6-CUG datasets and introducing the DiceLoss on the basis of CrossEntropyLoss solves the problem of positive and negative sample imbalance.Experimental results show that the new network model can achieve the following performance on the DeepGlobe dataset:79.40%for precision(Pre),97.93% for accuracy(Acc),69.28% for intersection over union(IoU),and 83.56% for mean intersection over union(MIoU).On the CHN6-CUG dataset,the model achieves the respective values of 78.17%for Pre,97.63%for Acc,65.4% for IoU,and 81.46% for MIoU.Compared with other network models,the fused ConvNeXt-UPerNet model can extract road information better when faced with the influence of noise contained in high-resolution remote sensing images.It also achieves multiscale image feature information with unified perception,ultimately improving the generalization ability of deep learning technology in extracting complex roads from high-resolution remote sensing images.
文摘This paper proposes a novel open set recognition method,the Spatial Distribution Feature Extraction Network(SDFEN),to address the problem of electromagnetic signal recognition in an open environment.The spatial distribution feature extraction layer in SDFEN replaces convolutional output neural networks with the spatial distribution features that focus more on inter-sample information by incorporating class center vectors.The designed hybrid loss function considers both intra-class distance and inter-class distance,thereby enhancing the similarity among samples of the same class and increasing the dissimilarity between samples of different classes during training.Consequently,this method allows unknown classes to occupy a larger space in the feature space.This reduces the possibility of overlap with known class samples and makes the boundaries between known and unknown samples more distinct.Additionally,the feature comparator threshold can be used to reject unknown samples.For signal open set recognition,seven methods,including the proposed method,are applied to two kinds of electromagnetic signal data:modulation signal and real-world emitter.The experimental results demonstrate that the proposed method outperforms the other six methods overall in a simulated open environment.Specifically,compared to the state-of-the-art Openmax method,the novel method achieves up to 8.87%and 5.25%higher micro-F-measures,respectively.
文摘Deep neural network-based relational extraction research has made significant progress in recent years,andit provides data support for many natural language processing downstream tasks such as building knowledgegraph,sentiment analysis and question-answering systems.However,previous studies ignored much unusedstructural information in sentences that could enhance the performance of the relation extraction task.Moreover,most existing dependency-based models utilize self-attention to distinguish the importance of context,whichhardly deals withmultiple-structure information.To efficiently leverage multiple structure information,this paperproposes a dynamic structure attention mechanism model based on textual structure information,which deeplyintegrates word embedding,named entity recognition labels,part of speech,dependency tree and dependency typeinto a graph convolutional network.Specifically,our model extracts text features of different structures from theinput sentence.Textual Structure information Graph Convolutional Networks employs the dynamic structureattention mechanism to learn multi-structure attention,effectively distinguishing important contextual features invarious structural information.In addition,multi-structure weights are carefully designed as amergingmechanismin the different structure attention to dynamically adjust the final attention.This paper combines these featuresand trains a graph convolutional network for relation extraction.We experiment on supervised relation extractiondatasets including SemEval 2010 Task 8,TACRED,TACREV,and Re-TACED,the result significantly outperformsthe previous.
基金supported by National Natural Science Foundation of China(No.61864025)2021 Longyuan Youth Innovation and Entrepreneurship Talent(Team),Young Doctoral Fund of Higher Education Institutions of Gansu Province(No.2021QB-49)+4 种基金Employment and Entrepreneurship Improvement Project of University Students of Gansu Province(No.2021-C-123)Intelligent Tunnel Supervision Robot Research Project(China Railway Scientific Research Institute(Scientific Research)(No.2020-KJ016-Z016-A2)Lanzhou Jiaotong University Youth Foundation(No.2015005)Gansu Higher Education Research Project(No.2016A-018)Gansu Dunhuang Cultural Relics Protection Research Center Open Project(No.GDW2021YB15).
文摘Road extraction based on deep learning is one of hot spots of semantic segmentation in the past decade.In this work,we proposed a framework based on codec network for automatic road extraction from remote sensing images.Firstly,a pre-trained ResNet34 was migrated to U-Net and its encoding structure was replaced to deepen the number of network layers,which reduces the error rate of road segmentation and the loss of details.Secondly,dilated convolution was used to connect the encoder and the decoder of network to expand the receptive field and retain more low-dimensional information of the image.Afterwards,the channel attention mechanism was used to select the information of the feature image obtained by up-sampling of the encoder,the weights of target features were optimized to enhance the features of target region and suppress the features of background and noise regions,and thus the feature extraction effect of the remote sensing image with complex background was optimized.Finally,an adaptive sigmoid loss function was proposed,which optimizes the imbalance between the road and the background,and makes the model reach the optimal solution.Experimental results show that compared with several semantic segmentation networks,the proposed method can greatly reduce the error rate of road segmentation and effectively improve the accuracy of road extraction from remote sensing images.
文摘This study aims to identify a natural plant chemical with hypolipidemic effects that can be used to treat high cholesterol without adverse reactions.Through network pharmacology screening,it was found that Rosae Rugosae Flos(RF)flavonoids had potential therapeutic effects on hyperlipidemia and its mechanism of action was discussed.TCMSP and GeneCards databases were used to obtain active ingredients and disease targets.Venn diagrams were drawn to illustrate the findings.The interaction network diagram was created using Cytoscape 3.8.0 software.The PPI protein network was constructed using String.GO and KEGG enrichment analysis was performed using Metascape.The results revealed 2 active flavonoid ingredients and 60 potential targets in RF.The key targets,including CCL2,PPARG,and PPARA,were found to play a role in multiple pathways such as the AGE-RAGE signaling pathway,lipid and atherosclerosis,and cancer pathway in diabetic complications.The solvent extraction method was optimized for efficient flavonoid extraction based on network pharmacology prediction results.This was achieved through a single factor and orthogonal test,resulting in an optimum process with a reflux time of 1.5 h,a solid-liquid ratio of 1:13 g/mL,and an ethanol concentration of 50%.
基金funded by the major research program of the of National Natural Science Foundation of China entitled Metallogenic Mechanisms and Regularity of the Lithium Ore Concentration Area in the Zabuye Salt Lake, Tibet (91962219)Science and Technology Major Project of the Tibet Autonomous Region ’s Science and Techonlogy Plan (XZ202201ZD0004G01)a geological survey project of China Geological Survey (DD20230037)。
文摘A reasonable classification of deposits holds great significance for identifying prospecting targets and deploying exploration. The world ’s keen demand for lithium resources has expedited the discovery of numerous novel lithium resources. Given the presence of varied classification criteria for lithium resources presently, this study further ascertained and classified the lithium resources according to their occurrence modes, obtaining 10 types and 5 subtypes of lithium deposits(resources) based on endogenetic and exogenetic factors. As indicated by surveys of Cenozoic exogenetic lithium deposits in China and abroad,the formation and distribution of the deposits are primarily determined by plate collision zones, their primary material sources are linked to the anatectic magmas in the deep oceanic crust, and they were formed primarily during the Miocene and Late Paleogene. The researchers ascertained that these deposits,especially those of the salt lake, geothermal, and volcanic deposit types, are formed by unique slightly acidic magmas, tend to migrate and accumulate toward low-lying areas, and display supernormal enrichment. However, the material sources of lithium deposits(resources) of the Neopaleozoic clay subtype and the deep brine type are yet to be further identified. Given the various types and complex origins of lithium deposits(resources), which were formed due to the interactions of multiple spheres, it is recommended that the mineralization of exogenetic lithium deposits(resources) be investigated by integrating tectono-geochemistry, paleoatmospheric circulation, and salinology. So far, industrialized lithium extraction is primarily achieved in lithium deposits of the salt lake, clay, and hard rock types. The lithium extraction employs different processes, with lithium extraction from salt lake-type lithium deposits proving the most energy-saving and cost-effective.
基金supported by the National Key Research and Development Program of China (2018YFC1901804 and 2018YFC1901805)the National Natural Science Foundation of China (21978045 and U1902219)。
文摘Copper was extracted from methylchlorosilane slurry residue by a direct hydrogen peroxide leaching method.A number of experimental parameters were analyzed to determine the extraction efficiency of copper.The extraction efficiency of copper reached 98.5%under the optimal leaching conditions,such as the hydrogen peroxide concentration of 1.875 mol·L^(-1),the leaching temperature of 323 K,the liquid–solid ratio of 20 ml·g^(-1),and the stirring speed of 300 r·min^(-1).The leaching kinetics of the copper extraction process was then described by the shrinking core model.There were two stages.The first stage was controlled by chemical reactions,while the second stage was controlled by interface transfer and product layer diffusion.The activation energy and kinetic control equations were determined,as well as an explanation of the leaching mechanism of copper extraction based on kinetic analysis and materials characterization.Copper resources can be recovered from the methylchlorosilane slurry residue efficiently and inexpensively with the methods used in this study.
基金supported by China’s National Key R&D Program,No.2019QY1404the National Natural Science Foundation of China,Grant No.U20A20161,U1836103the Basic Strengthening Program Project,No.2019-JCJQ-ZD-113.
文摘The continuous improvement of the cyber threat intelligence sharing mechanism provides new ideas to deal with Advanced Persistent Threats(APT).Extracting attack behaviors,i.e.,Tactics,Techniques,Procedures(TTP)from Cyber Threat Intelligence(CTI)can facilitate APT actors’profiling for an immediate response.However,it is difficult for traditional manual methods to analyze attack behaviors from cyber threat intelligence due to its heterogeneous nature.Based on the Adversarial Tactics,Techniques and Common Knowledge(ATT&CK)of threat behavior description,this paper proposes a threat behavioral knowledge extraction framework that integrates Heterogeneous Text Network(HTN)and Graph Convolutional Network(GCN)to solve this issue.It leverages the hierarchical correlation relationships of attack techniques and tactics in the ATT&CK to construct a text network of heterogeneous cyber threat intelligence.With the help of the Bidirectional EncoderRepresentation fromTransformers(BERT)pretraining model to analyze the contextual semantics of cyber threat intelligence,the task of threat behavior identification is transformed into a text classification task,which automatically extracts attack behavior in CTI,then identifies the malware and advanced threat actors.The experimental results show that F1 achieve 94.86%and 92.15%for the multi-label classification tasks of tactics and techniques.Extend the experiment to verify the method’s effectiveness in identifying the malware and threat actors in APT attacks.The F1 for malware and advanced threat actors identification task reached 98.45%and 99.48%,which are better than the benchmark model in the experiment and achieve state of the art.The model can effectivelymodel threat intelligence text data and acquire knowledge and experience migration by correlating implied features with a priori knowledge to compensate for insufficient sample data and improve the classification performance and recognition ability of threat behavior in text.
文摘A network intrusion detection system is critical for cyber security against llegitimate attacks.In terms of feature perspectives,network traffic may include a variety of elements such as attack reference,attack type,a subcategory of attack,host information,malicious scripts,etc.In terms of network perspectives,network traffic may contain an imbalanced number of harmful attacks when compared to normal traffic.It is challenging to identify a specific attack due to complex features and data imbalance issues.To address these issues,this paper proposes an Intrusion Detection System using transformer-based transfer learning for Imbalanced Network Traffic(IDS-INT).IDS-INT uses transformer-based transfer learning to learn feature interactions in both network feature representation and imbalanced data.First,detailed information about each type of attack is gathered from network interaction descriptions,which include network nodes,attack type,reference,host information,etc.Second,the transformer-based transfer learning approach is developed to learn detailed feature representation using their semantic anchors.Third,the Synthetic Minority Oversampling Technique(SMOTE)is implemented to balance abnormal traffic and detect minority attacks.Fourth,the Convolution Neural Network(CNN)model is designed to extract deep features from the balanced network traffic.Finally,the hybrid approach of the CNN-Long Short-Term Memory(CNN-LSTM)model is developed to detect different types of attacks from the deep features.Detailed experiments are conducted to test the proposed approach using three standard datasets,i.e.,UNsWNB15,CIC-IDS2017,and NSL-KDD.An explainable AI approach is implemented to interpret the proposed method and develop a trustable model.
文摘Condensed and hydrolysable tannins are non-toxic natural polyphenols that are a commercial commodity industrialized for tanning hides to obtain leather and for a growing number of other industrial applications mainly to substitute petroleum-based products.They are a definite class of sustainable materials of the forestry industry.They have been in operation for hundreds of years to manufacture leather and now for a growing number of applications in a variety of other industries,such as wood adhesives,metal coating,pharmaceutical/medical applications and several others.This review presents the main sources,either already or potentially commercial of this forestry by-materials,their industrial and laboratory extraction systems,their systems of analysis with their advantages and drawbacks,be these methods so simple to even appear primitive but nonetheless of proven effectiveness,or very modern and instrumental.It constitutes a basic but essential summary of what is necessary to know of these sustainable materials.In doing so,the review highlights some of the main challenges that remain to be addressed to deliver the quality and economics of tannin supply necessary to fulfill the industrial production requirements for some materials-based uses.
基金the support of the National Natural Science Foundation of China(22278234,21776151)。
文摘An artificial neural network(ANN)method is introduced to predict drop size in two kinds of pulsed columns with small-scale data sets.After training,the deviation between calculate and experimental results are 3.8%and 9.3%,respectively.Through ANN model,the influence of interfacial tension and pulsation intensity on the droplet diameter has been developed.Droplet size gradually increases with the increase of interfacial tension,and decreases with the increase of pulse intensity.It can be seen that the accuracy of ANN model in predicting droplet size outside the training set range is reach the same level as the accuracy of correlation obtained based on experiments within this range.For two kinds of columns,the drop size prediction deviations of ANN model are 9.6%and 18.5%and the deviations in correlations are 11%and 15%.
基金financially supported by National Natural Science Foundation of China(No.52274171)Joint National-Local Engineering Research Centre for Safe and Precise Coal Mining Fund(No.EC2023015)+1 种基金Excellent Youth Project of Universities in Anhui Province(No.2023AH030042)Unveiled List of Bidding Projects of Shanxi Province(No.20201101001)。
文摘Chemical solvents instead of pure water being as hydraulic fracturing fluid could effectively increase permeability and improve clean methane extraction efficiency.However,pore-fracture variation features of lean coal synergistically affected by solvents have not been fully understood.Ultrasonic testing,nuclear magnetic resonance analysis,liquid phase mass spectrometry was adopted to comprehensively analyze pore-fracture change characteristics of lean coal treated by combined solvent(NMP and CS_(2)).Meanwhile,quantitative characterization of above changing properties was conducted using geometric fractal theory.Relationship model between permeability,fractal dimension and porosity were established.Results indicate that the end face fractures of coal are well developed after CS2and combined solvent treatments,of which,end face box-counting fractal dimensions range from 1.1227 to 1.4767.Maximum decreases in ultrasonic longitudinal wave velocity of coal affected by NMP,CS_(2)and combined solvent are 2.700%,20.521%,22.454%,respectively.Solvent treatments could lead to increasing amount of both mesopores and macropores.Decrease ratio of fractal dimension Dsis 0.259%–2.159%,while permeability increases ratio of NMR ranges from 0.1904 to 6.4486.Meanwhile,combined solvent could dissolve coal polar and non-polar small molecules and expand flow space.Results could provide reference for solvent selection and parameter optimization of permeability-enhancement technology.
基金supported in part by the National Key R&D Program of China(No.2022YFB3904503)National Natural Science Foundation of China(No.62172418)。
文摘To solve the problem of poor detection and limited application range of current intrusion detection methods,this paper attempts to use deep learning neural network technology to study a new type of intrusion detection method.Hence,we proposed an intrusion detection algorithm based on convolutional neural network(CNN)and AdaBoost algorithm.This algorithm uses CNN to extract the characteristics of network traffic data,which is particularly suitable for the analysis of continuous and classified attack data.The AdaBoost algorithm is used to classify network attack data that improved the detection effect of unbalanced data classification.We adopt the UNSW-NB15 dataset to test of this algorithm in the PyCharm environment.The results show that the detection rate of algorithm is99.27%and the false positive rate is lower than 0.98%.Comparative analysis shows that this algorithm has advantages over existing methods in terms of detection rate and false positive rate for small proportion of attack data.
基金the Science and Technology Key Project of Anhui Province,China(No.2022e03020004).
文摘Lithium recovery from spent lithium-ion batteries(LIBs)have attracted extensive attention due to the skyrocketing price of lithium.The medium-temperature carbon reduction roasting was proposed to preferential selective extraction of lithium from spent Li-CoO_(2)(LCO)cathodes to overcome the incomplete recovery and loss of lithium during the recycling process.The LCO layered structure was destroyed and lithium was completely converted into water-soluble Li2CO_(3)under a suitable temperature to control the reduced state of the cobalt oxide.The Co metal agglomerates generated during medium-temperature carbon reduction roasting were broken by wet grinding and ultrasonic crushing to release the entrained lithium.The results showed that 99.10%of the whole lithium could be recovered as Li2CO_(3)with a purity of 99.55%.This work provided a new perspective on the preferentially selective extraction of lithium from spent lithium batteries.
基金National Natural Science Foundation of China(Nos.42071444,42101444)。
文摘Cultural relics line graphic serves as a crucial form of traditional artifact information documentation,which is a simple and intuitive product with low cost of displaying compared with 3D models.Dimensionality reduction is undoubtedly necessary for line drawings.However,most existing methods for artifact drawing rely on the principles of orthographic projection that always cannot avoid angle occlusion and data overlapping while the surface of cultural relics is complex.Therefore,conformal mapping was introduced as a dimensionality reduction way to compensate for the limitation of orthographic projection.Based on the given criteria for assessing surface complexity,this paper proposed a three-dimensional feature guideline extraction method for complex cultural relic surfaces.A 2D and 3D combined factor that measured the importance of points on describing surface features,vertex weight,was designed.Then the selection threshold for feature guideline extraction was determined based on the differences between vertex weight and shape index distributions.The feasibility and stability were verified through experiments conducted on real cultural relic surface data.Results demonstrated the ability of the method to address the challenges associated with the automatic generation of line drawings for complex surfaces.The extraction method and the obtained results will be useful for line graphic drawing,displaying and propaganda of cultural relics.
文摘Photovoltaic(PV)boards are a perfect way to create eco-friendly power from daylight.The defects in the PV panels are caused by various conditions;such defective PV panels need continuous monitoring.The recent development of PV panel monitoring systems provides a modest and viable approach to monitoring and managing the condition of the PV plants.In general,conventional procedures are used to identify the faulty modules earlier and to avoid declines in power generation.The existing deep learning architectures provide the required output to predict the faulty PV panels with less accuracy and a more time-consuming process.To increase the accuracy and to reduce the processing time,a new Convolutional Neural Network(CNN)architecture is required.Hence,in the present work,a new Real-time Multi Variant Deep learning Model(RMVDM)architecture is proposed,and it extracts the image features and classifies the defects in PV panels quickly with high accuracy.The defects that arise in the PV panels are identified by the CNN based RMVDM using RGB images.The biggest difference between CNN and its predecessors is that CNN automatically extracts the image features without any help from a person.The technique is quantitatively assessed and compared with existing faulty PV board identification approaches on the large real-time dataset.The results show that 98%of the accuracy and recall values in the fault detection and classification process.