The emergence of digital networks and the wide adoption of information on internet platforms have given rise to threats against users’private information.Many intruders actively seek such private data either for sale...The emergence of digital networks and the wide adoption of information on internet platforms have given rise to threats against users’private information.Many intruders actively seek such private data either for sale or other inappropriate purposes.Similarly,national and international organizations have country-level and company-level private information that could be accessed by different network attacks.Therefore,the need for a Network Intruder Detection System(NIDS)becomes essential for protecting these networks and organizations.In the evolution of NIDS,Artificial Intelligence(AI)assisted tools and methods have been widely adopted to provide effective solutions.However,the development of NIDS still faces challenges at the dataset and machine learning levels,such as large deviations in numeric features,the presence of numerous irrelevant categorical features resulting in reduced cardinality,and class imbalance in multiclass-level data.To address these challenges and offer a unified solution to NIDS development,this study proposes a novel framework that preprocesses datasets and applies a box-cox transformation to linearly transform the numeric features and bring them into closer alignment.Cardinality reduction was applied to categorical features through the binning method.Subsequently,the class imbalance dataset was addressed using the adaptive synthetic sampling data generation method.Finally,the preprocessed,refined,and oversampled feature set was divided into training and test sets with an 80–20 ratio,and two experiments were conducted.In Experiment 1,the binary classification was executed using four machine learning classifiers,with the extra trees classifier achieving the highest accuracy of 97.23%and an AUC of 0.9961.In Experiment 2,multiclass classification was performed,and the extra trees classifier emerged as the most effective,achieving an accuracy of 81.27%and an AUC of 0.97.The results were evaluated based on training,testing,and total time,and a comparative analysis with state-of-the-art studies proved the robustness and significance of the applied methods in developing a timely and precision-efficient solution to NIDS.展开更多
In this paper,we propose a new relational schema (R-schema) to XML schema translation algorithm, VQT, which analyzes the value cardinality and user query patterns and extracts the implicit referential integrities by u...In this paper,we propose a new relational schema (R-schema) to XML schema translation algorithm, VQT, which analyzes the value cardinality and user query patterns and extracts the implicit referential integrities by using the cardinality property of foreign key constraints between columns and the equi-join characteristic in user queries. The VQT algorithm can apply the extracted implied referential integrity relation information to the R-schema and create an XML schema as the final result. Therefore, the VQT algorithm prevents the R-schema from being incorrectly converted into the XML schema, and it richly and powerfully represents all the information in the R-schema by creating an XML schema as the translation result on behalf of the XML DTD.展开更多
An excellent cardinality estimation can make the query optimiser produce a good execution plan.Although there are some studies on cardinality estimation,the prediction results of existing cardinality estimators are in...An excellent cardinality estimation can make the query optimiser produce a good execution plan.Although there are some studies on cardinality estimation,the prediction results of existing cardinality estimators are inaccurate and the query efficiency cannot be guaranteed as well.In particular,they are difficult to accurately obtain the complex relationships between multiple tables in complex database systems.When dealing with complex queries,the existing cardinality estimators cannot achieve good results.In this study,a novel cardinality estimator is proposed.It uses the core techniques with the BiLSTM network structure and adds the attention mechanism.First,the columns involved in the query statements in the training set are sampled and compressed into bitmaps.Then,the Word2vec model is used to embed the word vectors about the query statements.Finally,the BiLSTM network and attention mechanism are employed to deal with word vectors.The proposed model takes into consideration not only the correlation between tables but also the processing of complex predicates.Extensive experiments and the evaluation of BiLSTM-Attention Cardinality Estimator(BACE)on the IMDB datasets are conducted.The results show that the deep learning model can significantly improve the quality of cardinality estimation,which is a vital role in query optimisation for complex databases.展开更多
The Cardinality Constraint-Based Optimization problem is investigated in this note. In portfolio optimization problem, the cardinality constraint allows one to invest in assets out of a universe of N assets for a pres...The Cardinality Constraint-Based Optimization problem is investigated in this note. In portfolio optimization problem, the cardinality constraint allows one to invest in assets out of a universe of N assets for a prespecified value of K. It is generally agreed that choosing a “small” value of K forces the implementation of diversification in small portfolios. However, the question of how small must be K has remained unanswered. In the present work, using a comparative approach we show computationally that optimal portfolio selection with a relatively small or large number of assets, K, may produce similar results with differentiated reliabilities.展开更多
This paper studied cardinality constrained portfolio with integer weight.We suggested two optimization models and used two genetic algorithms to solve them.In this paper,after finding well matching stocks,according to...This paper studied cardinality constrained portfolio with integer weight.We suggested two optimization models and used two genetic algorithms to solve them.In this paper,after finding well matching stocks,according to investor’s target by using first genetic algorithm,we gave optimal integer weight of portfolio with well matching stocks by using second genetic algorithm.Through numerical comparisons with other feasible portfolios,we verified advantages of designed portfolio with two genetic algorithms.For a numerical comparison,we used a prepared data consisted of 18 stocks listed in S&P 500 and numerical example strongly supported the designed portfolio in this paper.Also,we made all comparisons visible through all feasible efficient frontiers.展开更多
The concepts of quasi-cardinality, possibility and expectation of a fuzzy set on a measurable set, as a generalization of the quasi-cardinality of a fuzzy set on a finite universe of discourse is proposed. Also some p...The concepts of quasi-cardinality, possibility and expectation of a fuzzy set on a measurable set, as a generalization of the quasi-cardinality of a fuzzy set on a finite universe of discourse is proposed. Also some properties of them are given out and thus some relevant results are elucidated.展开更多
目的研究阴道分娩对子宫骶韧带(uterosacral ligaments,USLs)和主韧带(cardinal ligaments,CLs)生物力学性能的影响,进而探讨阴道分娩对盆腔器官脱垂(pelvic organ prolapse,POP)的影响。方法选用成年母猪(已育母猪5头,未育母猪5头)作...目的研究阴道分娩对子宫骶韧带(uterosacral ligaments,USLs)和主韧带(cardinal ligaments,CLs)生物力学性能的影响,进而探讨阴道分娩对盆腔器官脱垂(pelvic organ prolapse,POP)的影响。方法选用成年母猪(已育母猪5头,未育母猪5头)作为动物模型,通过单轴拉伸实验测量离体母猪的USLs和CLs被动力学行为,分析分娩对USLs和CLs生物力学性能的影响。结果猪子宫韧带组织的被动力学行为呈非线性。无论分娩与否,右骶韧带的最大应力大于左骶韧带(P<0.05);分娩后,二者力学性能存在显著差异。未育母猪左主韧带最大应力略大于右主韧带(P<0.05);分娩后,两者差异降低(P>0.05)。USLs最大应力均大于CLs,表明USLs比CLs承受的张力更大,USLs在POP中起到关键性作用。结论研究结果为认识USLs和CLs力学特性提供参考,可以指导更好治疗方法的发展,如POP手术重建,也为预防POP发生提供理论基础。展开更多
针对传统的数据库管理系统无法很好地学习谓词之间的交互以及无法准确地估计复杂查询的基数问题,提出了一种树形结构的长短期记忆神经网络(Tree Long Short Term Memory, TreeLSTM)模型建模查询,并使用该模型对新的查询基数进行估计.所...针对传统的数据库管理系统无法很好地学习谓词之间的交互以及无法准确地估计复杂查询的基数问题,提出了一种树形结构的长短期记忆神经网络(Tree Long Short Term Memory, TreeLSTM)模型建模查询,并使用该模型对新的查询基数进行估计.所提出的模型考虑了查询语句中包含的合取和析取运算,根据谓词之间的操作符类型将子表达式构建为树形结构,根据组合子表达式向量来表示连续向量空间中的任意逻辑表达式.TreeLSTM模型通过捕捉查询谓词之间的顺序依赖关系从而提升基数估计的性能和准确度,将TreeLSTM与基于直方图方法、基于学习的MSCN和TreeRNN方法进行了比较.实验结果表明:TreeLSTM的估算误差比直方图、MSCN、TreeRNN方法的误差分别降低了60.41%,33.33%和11.57%,该方法显著提高了基数估计器的性能.展开更多
An edge coloring of hypergraph H is a function such that holds for any pair of intersecting edges . The minimum number of colors in edge colorings of H is called the chromatic index of H and is ...An edge coloring of hypergraph H is a function such that holds for any pair of intersecting edges . The minimum number of colors in edge colorings of H is called the chromatic index of H and is denoted by . Erdös, Faber and Lovász proposed a famous conjecture that holds for any loopless linear hypergraph H with n vertices. In this paper, we show that is true for gap-restricted hypergraphs. Our result extends a result of Alesandroni in 2021.展开更多
文摘The emergence of digital networks and the wide adoption of information on internet platforms have given rise to threats against users’private information.Many intruders actively seek such private data either for sale or other inappropriate purposes.Similarly,national and international organizations have country-level and company-level private information that could be accessed by different network attacks.Therefore,the need for a Network Intruder Detection System(NIDS)becomes essential for protecting these networks and organizations.In the evolution of NIDS,Artificial Intelligence(AI)assisted tools and methods have been widely adopted to provide effective solutions.However,the development of NIDS still faces challenges at the dataset and machine learning levels,such as large deviations in numeric features,the presence of numerous irrelevant categorical features resulting in reduced cardinality,and class imbalance in multiclass-level data.To address these challenges and offer a unified solution to NIDS development,this study proposes a novel framework that preprocesses datasets and applies a box-cox transformation to linearly transform the numeric features and bring them into closer alignment.Cardinality reduction was applied to categorical features through the binning method.Subsequently,the class imbalance dataset was addressed using the adaptive synthetic sampling data generation method.Finally,the preprocessed,refined,and oversampled feature set was divided into training and test sets with an 80–20 ratio,and two experiments were conducted.In Experiment 1,the binary classification was executed using four machine learning classifiers,with the extra trees classifier achieving the highest accuracy of 97.23%and an AUC of 0.9961.In Experiment 2,multiclass classification was performed,and the extra trees classifier emerged as the most effective,achieving an accuracy of 81.27%and an AUC of 0.97.The results were evaluated based on training,testing,and total time,and a comparative analysis with state-of-the-art studies proved the robustness and significance of the applied methods in developing a timely and precision-efficient solution to NIDS.
基金Project supported by the 2nd Brain Korea Project
文摘In this paper,we propose a new relational schema (R-schema) to XML schema translation algorithm, VQT, which analyzes the value cardinality and user query patterns and extracts the implicit referential integrities by using the cardinality property of foreign key constraints between columns and the equi-join characteristic in user queries. The VQT algorithm can apply the extracted implied referential integrity relation information to the R-schema and create an XML schema as the final result. Therefore, the VQT algorithm prevents the R-schema from being incorrectly converted into the XML schema, and it richly and powerfully represents all the information in the R-schema by creating an XML schema as the translation result on behalf of the XML DTD.
基金supported by the National Natural Science Foundation of China under grant nos.61772091,61802035,61962006,61962038,U1802271,U2001212,and 62072311the Sichuan Science and Technology Program under grant nos.2021JDJQ0021 and 22ZDYF2680+7 种基金the CCF‐Huawei Database System Innovation Research Plan under grant no.CCF‐HuaweiDBIR2020004ADigital Media Art,Key Laboratory of Sichuan Province,Sichuan Conservatory of Music,Chengdu,China under grant no.21DMAKL02the Chengdu Major Science and Technology Innovation Project under grant no.2021‐YF08‐00156‐GXthe Chengdu Technology Innovation and Research and Development Project under grant no.2021‐YF05‐00491‐SNthe Natural Science Foundation of Guangxi under grant no.2018GXNSFDA138005the Guangdong Basic and Applied Basic Research Foundation under grant no.2020B1515120028the Science and Technology Innovation Seedling Project of Sichuan Province under grant no 2021006the College Student Innovation and Entrepreneurship Training Program of Chengdu University of Information Technology under grant nos.202110621179 and 202110621186.
文摘An excellent cardinality estimation can make the query optimiser produce a good execution plan.Although there are some studies on cardinality estimation,the prediction results of existing cardinality estimators are inaccurate and the query efficiency cannot be guaranteed as well.In particular,they are difficult to accurately obtain the complex relationships between multiple tables in complex database systems.When dealing with complex queries,the existing cardinality estimators cannot achieve good results.In this study,a novel cardinality estimator is proposed.It uses the core techniques with the BiLSTM network structure and adds the attention mechanism.First,the columns involved in the query statements in the training set are sampled and compressed into bitmaps.Then,the Word2vec model is used to embed the word vectors about the query statements.Finally,the BiLSTM network and attention mechanism are employed to deal with word vectors.The proposed model takes into consideration not only the correlation between tables but also the processing of complex predicates.Extensive experiments and the evaluation of BiLSTM-Attention Cardinality Estimator(BACE)on the IMDB datasets are conducted.The results show that the deep learning model can significantly improve the quality of cardinality estimation,which is a vital role in query optimisation for complex databases.
文摘The Cardinality Constraint-Based Optimization problem is investigated in this note. In portfolio optimization problem, the cardinality constraint allows one to invest in assets out of a universe of N assets for a prespecified value of K. It is generally agreed that choosing a “small” value of K forces the implementation of diversification in small portfolios. However, the question of how small must be K has remained unanswered. In the present work, using a comparative approach we show computationally that optimal portfolio selection with a relatively small or large number of assets, K, may produce similar results with differentiated reliabilities.
文摘This paper studied cardinality constrained portfolio with integer weight.We suggested two optimization models and used two genetic algorithms to solve them.In this paper,after finding well matching stocks,according to investor’s target by using first genetic algorithm,we gave optimal integer weight of portfolio with well matching stocks by using second genetic algorithm.Through numerical comparisons with other feasible portfolios,we verified advantages of designed portfolio with two genetic algorithms.For a numerical comparison,we used a prepared data consisted of 18 stocks listed in S&P 500 and numerical example strongly supported the designed portfolio in this paper.Also,we made all comparisons visible through all feasible efficient frontiers.
文摘The concepts of quasi-cardinality, possibility and expectation of a fuzzy set on a measurable set, as a generalization of the quasi-cardinality of a fuzzy set on a finite universe of discourse is proposed. Also some properties of them are given out and thus some relevant results are elucidated.
文摘目的研究阴道分娩对子宫骶韧带(uterosacral ligaments,USLs)和主韧带(cardinal ligaments,CLs)生物力学性能的影响,进而探讨阴道分娩对盆腔器官脱垂(pelvic organ prolapse,POP)的影响。方法选用成年母猪(已育母猪5头,未育母猪5头)作为动物模型,通过单轴拉伸实验测量离体母猪的USLs和CLs被动力学行为,分析分娩对USLs和CLs生物力学性能的影响。结果猪子宫韧带组织的被动力学行为呈非线性。无论分娩与否,右骶韧带的最大应力大于左骶韧带(P<0.05);分娩后,二者力学性能存在显著差异。未育母猪左主韧带最大应力略大于右主韧带(P<0.05);分娩后,两者差异降低(P>0.05)。USLs最大应力均大于CLs,表明USLs比CLs承受的张力更大,USLs在POP中起到关键性作用。结论研究结果为认识USLs和CLs力学特性提供参考,可以指导更好治疗方法的发展,如POP手术重建,也为预防POP发生提供理论基础。
文摘针对传统的数据库管理系统无法很好地学习谓词之间的交互以及无法准确地估计复杂查询的基数问题,提出了一种树形结构的长短期记忆神经网络(Tree Long Short Term Memory, TreeLSTM)模型建模查询,并使用该模型对新的查询基数进行估计.所提出的模型考虑了查询语句中包含的合取和析取运算,根据谓词之间的操作符类型将子表达式构建为树形结构,根据组合子表达式向量来表示连续向量空间中的任意逻辑表达式.TreeLSTM模型通过捕捉查询谓词之间的顺序依赖关系从而提升基数估计的性能和准确度,将TreeLSTM与基于直方图方法、基于学习的MSCN和TreeRNN方法进行了比较.实验结果表明:TreeLSTM的估算误差比直方图、MSCN、TreeRNN方法的误差分别降低了60.41%,33.33%和11.57%,该方法显著提高了基数估计器的性能.
文摘An edge coloring of hypergraph H is a function such that holds for any pair of intersecting edges . The minimum number of colors in edge colorings of H is called the chromatic index of H and is denoted by . Erdös, Faber and Lovász proposed a famous conjecture that holds for any loopless linear hypergraph H with n vertices. In this paper, we show that is true for gap-restricted hypergraphs. Our result extends a result of Alesandroni in 2021.