The inter-agency government information sharing(IAGIS)plays an important role in improving service and efficiency of government agencies.Currently,there is still no effective and secure way for data-driven IAGIS to fu...The inter-agency government information sharing(IAGIS)plays an important role in improving service and efficiency of government agencies.Currently,there is still no effective and secure way for data-driven IAGIS to fulfill dynamic demands of information sharing between government agencies.Motivated by blockchain and data mining,a data-driven framework is proposed for IAGIS in this paper.Firstly,the blockchain is used as the core to design the whole framework for monitoring and preventing leakage and abuse of government information,in order to guarantee information security.Secondly,a four-layer architecture is designed for implementing the proposed framework.Thirdly,the classical data mining algorithms PageRank and Apriori are applied to dynamically design smart contracts for information sharing,for the purposed of flexibly adjusting the information sharing strategies according to the practical demands of government agencies for public management and public service.Finally,a case study is presented to illustrate the operation of the proposed framework.展开更多
A modified multiple-component scattering power decomposition for analyzing polarimetric synthetic aperture radar(PolSAR)data is proposed.The modified decomposition involves two distinct steps.Firstly,ei⁃genvectors of ...A modified multiple-component scattering power decomposition for analyzing polarimetric synthetic aperture radar(PolSAR)data is proposed.The modified decomposition involves two distinct steps.Firstly,ei⁃genvectors of the coherency matrix are used to modify the scattering models.Secondly,the entropy and anisotro⁃py of targets are used to improve the volume scattering power.With the guarantee of high double-bounce scatter⁃ing power in the urban areas,the proposed algorithm effectively improves the volume scattering power of vegeta⁃tion areas.The efficacy of the modified multiple-component scattering power decomposition is validated using ac⁃tual AIRSAR PolSAR data.The scattering power obtained through decomposing the original coherency matrix and the coherency matrix after orientation angle compensation is compared with three algorithms.Results from the experiment demonstrate that the proposed decomposition yields more effective scattering power for different PolSAR data sets.展开更多
Magnesium(Mg)is a promising alternative to lithium(Li)as an anode material in solid-state batteries due to its abundance and high theoretical volumetric capacity.However,the sluggish Mg-ion conduction in the lattice o...Magnesium(Mg)is a promising alternative to lithium(Li)as an anode material in solid-state batteries due to its abundance and high theoretical volumetric capacity.However,the sluggish Mg-ion conduction in the lattice of solidstate electrolytes(SSEs)is one of the key challenges that hamper the development of Mg-ion solid-state batteries.Though various Mg-ion SSEs have been reported in recent years,key insights are hard to be derived from a single literature report.Besides,the structure-performance relationships of Mg-ion SSEs need to be further unraveled to provide a more precise design guideline for SSEs.In this viewpoint article,we analyze the structural characteristics of the Mg-based SSEs with high ionic conductivity reported in the last four decades based upon data mining-we provide big-data-derived insights into the challenges and opportunities in developing next-generation Mg-ion SSEs.展开更多
Under single-satellite observation,the parameter estimation of the boost phase of high-precision space noncooperative targets requires prior information.To improve the accuracy without prior information,we propose a p...Under single-satellite observation,the parameter estimation of the boost phase of high-precision space noncooperative targets requires prior information.To improve the accuracy without prior information,we propose a parameter estimation model of the boost phase based on trajectory plane parametric cutting.The use of the plane passing through the geo-center and the cutting sequence line of sight(LOS)generates the trajectory-cutting plane.With the coefficient of the trajectory cutting plane directly used as the parameter to be estimated,a motion parameter estimation model in space non-cooperative targets is established,and the Gauss-Newton iteration method is used to solve the flight parameters.The experimental results show that the estimation algorithm proposed in this paper weakly relies on prior information and has higher estimation accuracy,providing a practical new idea and method for the parameter estimation of space non-cooperative targets under single-satellite warning.展开更多
Objective Clinical medical record data associated with hepatitis B-related acute-on-chronic liver failure(HBV-ACLF)generally have small sample sizes and a class imbalance.However,most machine learning models are desig...Objective Clinical medical record data associated with hepatitis B-related acute-on-chronic liver failure(HBV-ACLF)generally have small sample sizes and a class imbalance.However,most machine learning models are designed based on balanced data and lack interpretability.This study aimed to propose a traditional Chinese medicine(TCM)diagnostic model for HBV-ACLF based on the TCM syndrome differentiation and treatment theory,which is clinically interpretable and highly accurate.Methods We collected medical records from 261 patients diagnosed with HBV-ACLF,including three syndromes:Yang jaundice(214 cases),Yang-Yin jaundice(41 cases),and Yin jaundice(6 cases).To avoid overfitting of the machine learning model,we excluded the cases of Yin jaundice.After data standardization and cleaning,we obtained 255 relevant medical records of Yang jaundice and Yang-Yin jaundice.To address the class imbalance issue,we employed the oversampling method and five machine learning methods,including logistic regression(LR),support vector machine(SVM),decision tree(DT),random forest(RF),and extreme gradient boosting(XGBoost)to construct the syndrome diagnosis models.This study used precision,F1 score,the area under the receiver operating characteristic(ROC)curve(AUC),and accuracy as model evaluation metrics.The model with the best classification performance was selected to extract the diagnostic rule,and its clinical significance was thoroughly analyzed.Furthermore,we proposed a novel multiple-round stable rule extraction(MRSRE)method to obtain a stable rule set of features that can exhibit the model’s clinical interpretability.Results The precision of the five machine learning models built using oversampled balanced data exceeded 0.90.Among these models,the accuracy of RF classification of syndrome types was 0.92,and the mean F1 scores of the two categories of Yang jaundice and Yang-Yin jaundice were 0.93 and 0.94,respectively.Additionally,the AUC was 0.98.The extraction rules of the RF syndrome differentiation model based on the MRSRE method revealed that the common features of Yang jaundice and Yang-Yin jaundice were wiry pulse,yellowing of the urine,skin,and eyes,normal tongue body,healthy sublingual vessel,nausea,oil loathing,and poor appetite.The main features of Yang jaundice were a red tongue body and thickened sublingual vessels,whereas those of Yang-Yin jaundice were a dark tongue body,pale white tongue body,white tongue coating,lack of strength,slippery pulse,light red tongue body,slimy tongue coating,and abdominal distension.This is aligned with the classifications made by TCM experts based on TCM syndrome differentiation and treatment theory.Conclusion Our model can be utilized for differentiating HBV-ACLF syndromes,which has the potential to be applied to generate other clinically interpretable models with high accuracy on clinical data characterized by small sample sizes and a class imbalance.展开更多
Data organization requires high efficiency for large amount of data applied in the digital mine system. A new method of storing massive data of block model is proposed to meet the characteristics of the database, incl...Data organization requires high efficiency for large amount of data applied in the digital mine system. A new method of storing massive data of block model is proposed to meet the characteristics of the database, including ACID-compliant, concurrency support, data sharing, and efficient access. Each block model is organized by linear octree, stored in LMDB(lightning memory-mapped database). Geological attribute can be queried at any point of 3D space by comparison algorithm of location code and conversion algorithm from address code of geometry space to location code of storage. The performance and robustness of querying geological attribute at 3D spatial region are enhanced greatly by the transformation from 3D to 2D and the method of 2D grid scanning to screen the inner and outer points. Experimental results showed that this method can access the massive data of block model, meeting the database characteristics. The method with LMDB is at least 3 times faster than that with etree, especially when it is used to read. In addition, the larger the amount of data is processed, the more efficient the method would be.展开更多
To fully extract and mine the multi-scale features of reservoirs and geologic structures in time/depth and space dimensions, a new 3D multi-scale volumetric curvature (MSVC) methodology is presented in this paper. W...To fully extract and mine the multi-scale features of reservoirs and geologic structures in time/depth and space dimensions, a new 3D multi-scale volumetric curvature (MSVC) methodology is presented in this paper. We also propose a fast algorithm for computing 3D volumetric curvature. In comparison to conventional volumetric curvature attributes, its main improvements and key algorithms introduce multi-frequency components expansion in time-frequency domain and the corresponding multi-scale adaptive differential operator in the wavenumber domain, into the volumetric curvature calculation. This methodology can simultaneously depict seismic multi-scale features in both time and space. Additionally, we use data fusion of volumetric curvatures at various scales to take full advantage of the geologic features and anomalies extracted by curvature measurements at different scales. The 3D MSVC can highlight geologic anomalies and reduce noise at the same time. Thus, it improves the interpretation efficiency of curvature attributes analysis. The 3D MSVC is applied to both land and marine 3D seismic data. The results demonstrate that it can indicate the spatial distribution of reservoirs, detect faults and fracture zones, and identify their multi-scale properties.展开更多
Traditionally, airborne time-domain electromagnetic (ATEM) data are inverted to derive the earth model by iteration. However, the data are often highly correlated among channels and consequently cause ill-posed and ...Traditionally, airborne time-domain electromagnetic (ATEM) data are inverted to derive the earth model by iteration. However, the data are often highly correlated among channels and consequently cause ill-posed and over-determined problems in the inversion. The correlation complicates the mapping relation between the ATEM data and the earth parameters and thus increases the inversion complexity. To obviate this, we adopt principal component analysis to transform ATEM data into orthogonal principal components (PCs) to reduce the correlations and the data dimensionality and simultaneously suppress the unrelated noise. In this paper, we use an artificial neural network (ANN) to approach the PCs mapping relation with the earth model parameters, avoiding the calculation of Jacobian derivatives. The PC-based ANN algorithm is applied to synthetic data for layered models compared with data-based ANN for airborne time-domain electromagnetic inversion. The results demonstrate the PC-based ANN advantages of simpler network structure, less training steps, and better inversion results over data-based ANN, especially for contaminated data. Furthermore, the PC-based ANN algorithm effectiveness is examined by the inversion of the pseudo 2D model and comparison with data-based ANN and Zhody's methods. The results indicate that PC-based ANN inversion can achieve a better agreement with the true model and also proved that PC-based ANN is feasible to invert large ATEM datasets.展开更多
The location of U-turn bays is an important consideration in indirect driveway left-turn treatments.In order to improve the performance of right-turns followed by U-turns(RTUTs),this study evaluates the impacts of t...The location of U-turn bays is an important consideration in indirect driveway left-turn treatments.In order to improve the performance of right-turns followed by U-turns(RTUTs),this study evaluates the impacts of the separation distances between driveway exits and downstream U-turn locations on the safety and operational performance of vehicles making RTUTs.Crash data are investigated at 179 selected roadway segments,and travel time data are measured using video cameras at 29 locations in the state of Florida,USA.Crash rate models and travel time models are developed based on data collected in the field.It is found that the separation distance between driveway exits and downstream U-turn locations significantly impacts the safety and operational performance of vehicles making right turns followed by U-turns.Based on the research results,the minimum and optimal separation distances between driveways and U-turn locations under different roadway conditions are determined to facilitate driver use of RTUTs.The results of this study can be used for future intersection improvement projects in China.展开更多
Missing data are a problem in geophysical surveys, and interpolation and reconstruction of missing data is part of the data processing and interpretation. Based on the sparseness of the geophysical data or the transfo...Missing data are a problem in geophysical surveys, and interpolation and reconstruction of missing data is part of the data processing and interpretation. Based on the sparseness of the geophysical data or the transform domain, we can improve the accuracy and stability of the reconstruction by transforming it to a sparse optimization problem. In this paper, we propose a mathematical model for the sparse reconstruction of data based on the LO-norm minimization. Furthermore, we discuss two types of the approximation algorithm for the LO- norm minimization according to the size and characteristics of the geophysical data: namely, the iteratively reweighted least-squares algorithm and the fast iterative hard thresholding algorithm. Theoretical and numerical analysis showed that applying the iteratively reweighted least-squares algorithm to the reconstruction of potential field data exploits its fast convergence rate, short calculation time, and high precision, whereas the fast iterative hard thresholding algorithm is more suitable for processing seismic data, moreover, its computational efficiency is better than that of the traditional iterative hard thresholding algorithm.展开更多
Regularization inversion uses constraints and a regularization factor to solve ill- posed inversion problems in geophysics. The choice of the regularization factor and of the initial model is critical in regularizatio...Regularization inversion uses constraints and a regularization factor to solve ill- posed inversion problems in geophysics. The choice of the regularization factor and of the initial model is critical in regularization inversion. To deal with these problems, we propose a multiobjective particle swarm inversion (MOPSOI) algorithm to simultaneously minimize the data misfit and model constraints, and obtain a multiobjective inversion solution set without the gradient information of the objective function and the regularization factor. We then choose the optimum solution from the solution set based on the trade-off between data misfit and constraints that substitute for the regularization factor. The inversion of synthetic two-dimensional magnetic data suggests that the MOPSOI algorithm can obtain as many feasible solutions as possible; thus, deeper insights of the inversion process can be gained and more reasonable solutions can be obtained by balancing the data misfit and constraints. The proposed MOPSOI algorithm can deal with the problems of choosing the right regularization factor and the initial model.展开更多
The conventional nonstationary convolutional model assumes that the seismic signal is recorded at normal incidence. Raw shot gathers are far from this assumption because of the effects of offsets. Because of such prob...The conventional nonstationary convolutional model assumes that the seismic signal is recorded at normal incidence. Raw shot gathers are far from this assumption because of the effects of offsets. Because of such problems, we propose a novel prestack nonstationary deconvolution approach. We introduce the radial trace (RT) transform to the nonstationary deconvolution, we estimate the nonstationary deconvolution factor with hyperbolic smoothing based on variable-step sampling (VSS) in the RT domain, and we obtain the high-resolution prestack nonstationary deconvolution data. The RT transform maps the shot record from the offset and traveltime coordinates to those of apparent velocity and traveltime. The ray paths of the traces in the RT better satisfy the assumptions of the convolutional model. The proposed method combines the advantages of stationary deconvolution and inverse Q filtering, without prior information for Q. The nonstationary deconvolution in the RT domain is more suitable than that in the space-time (XT) domain for prestack data because it is the generalized extension of normal incidence. Tests with synthetic and real data demonstrate that the proposed method is more effective in compensating for large-offset and deep data.展开更多
The HY-2 satellite carrying a satellite-borne GPS receiver is the first Chinese radar altimeter satellite, whose radial orbit determination precision must reach the centimeter level. Now HY-2 is in the test phase so t...The HY-2 satellite carrying a satellite-borne GPS receiver is the first Chinese radar altimeter satellite, whose radial orbit determination precision must reach the centimeter level. Now HY-2 is in the test phase so that the observations are not openly released. In order to study the precise orbit determination precision and procedure for HY-2 based on the satellite- borne GPS technique, the satellite-borne GPS data are simulated in this paper. The HY-2 satellite-borne GPS antenna can receive at least seven GPS satellites each epoch, which can validate the GPS receiver and antenna design. What's more, the precise orbit determination processing flow is given and precise orbit determination experiments are conducted using the HY-2-borne GPS data with both the reduced-dynamic method and the kinematic geometry method. With the 1 and 3 mm phase data random errors, the radial orbit determination precision can achieve the centimeter level using these two methods and the kinematic orbit accuracy is slightly lower than that of the reduced-dynamic orbit. The earth gravity field model is an important factor which seriously affects the precise orbit determination of altimeter satellites. The reduced-dynamic orbit determination experiments are made with different earth gravity field models, such as EIGEN2, EGM96, TEG4, and GEMT3. Using a large number of high precision satellite-bome GPS data, the HY-2 precise orbit determination can reach the centimeter level with commonly used earth gravity field models up to above 50 degrees and orders.展开更多
To solve the query processing correctness problem for semantic-based relational data integration,the semantics of SAPRQL(simple protocol and RDF query language) queries is defined.In the course of query rewriting,al...To solve the query processing correctness problem for semantic-based relational data integration,the semantics of SAPRQL(simple protocol and RDF query language) queries is defined.In the course of query rewriting,all relative tables are found and decomposed into minimal connectable units.Minimal connectable units are joined according to semantic queries to produce the semantically correct query plans.Algorithms for query rewriting and transforming are presented.Computational complexity of the algorithms is discussed.Under the worst case,the query decomposing algorithm can be finished in O(n2) time and the query rewriting algorithm requires O(nm) time.And the performance of the algorithms is verified by experiments,and experimental results show that when the length of query is less than 8,the query processing algorithms can provide satisfactory performance.展开更多
The molecular phylogeny of the Lardizabalaceae is reconstructed based on chloroplast trn L_F sequences alone and combined trn L_F and rbc L sequences. The phylogenetic topologies agree well with Qin's and...The molecular phylogeny of the Lardizabalaceae is reconstructed based on chloroplast trn L_F sequences alone and combined trn L_F and rbc L sequences. The phylogenetic topologies agree well with Qin's and Takhtajan's tribal classification in both analyses. Decaisneae and Sinofranchetieae are basal clades in the phylogenetic trees and external to all other taxa in the family. Lardizabaleae consisting of Boquila and Lardizabala are well supported in both trn L_F (100%) analysis and trn L_F and rbc L combined analysis (99%). Tribe Akebieae are strongly supported by a bootstrap value of 100% in both trn L_F analysis and trn L_F and rbc L combined analysis. However, the new genus Archakebia is nested within the genus Akebia in the trn L_F trees. In the combined trees, Archakebia is sister to Akebia with high bootstrap support. The inter_relationships among three closely related genera Parvatia , Holboellia and Stauntonia are still problematic. P. brunoniana ssp. elliptica is sister to H. latifolia in both analyses with low bootstrap support. H. parviflora is nested within the Stauntonia and sister to S. cavalerieana . Therefore, these three genera of tribe Akebieae may not be monophylytic and their generic boundary and delimitation need to be further studied, by exploring more molecular data, together with more morphological characters.展开更多
This paper suggests that a single class rather than methods should be used as the slice scope to compute class cohesion. First, for a given attribute, the statements in all methods that last define the attribute are c...This paper suggests that a single class rather than methods should be used as the slice scope to compute class cohesion. First, for a given attribute, the statements in all methods that last define the attribute are computed. Then, the forward and backward data slices for this attribute are generated by using the class as the slice scope and are combined to compute the corresponding class data slice. Finally, the class cohesion is computed based on all class data slices for the attributes. Compared to traditional cohesion metrics that use methods as the slice scope, the proposed metrics that use a single class as slice scope take into account the possible interactions between the methods. The experimental results show that class cohesion can be more accurately measured when using the class as the slice scope.展开更多
One-class support vector machine (OCSVM) and support vector data description (SVDD) are two main domain-based one-class (kernel) classifiers. To reveal their relationship with density estimation in the case of t...One-class support vector machine (OCSVM) and support vector data description (SVDD) are two main domain-based one-class (kernel) classifiers. To reveal their relationship with density estimation in the case of the Gaussian kernel, OCSVM and SVDD are firstly unified into the framework of kernel density estimation, and the essential relationship between them is explicitly revealed. Then the result proves that the density estimation induced by OCSVM or SVDD is in agreement with the true density. Meanwhile, it can also reduce the integrated squared error (ISE). Finally, experiments on several simulated datasets verify the revealed relationships.展开更多
An approach to contour extraction and feature point detection in the 3-D fragment reassembly is proposed. A simple and effective technique is used for building the intrinsic topology of the fragment data suitable for ...An approach to contour extraction and feature point detection in the 3-D fragment reassembly is proposed. A simple and effective technique is used for building the intrinsic topology of the fragment data suitable for contour extraction. For the scanned data in which the topology is difficult to be achieved, the corresponding solutions are given to manage this problem. A robust approach is used for the curvature and torsion calculation of the discrete contour in a 3-D space. Finally, a method is developed for detecting feature points of the fragment contour based on total curvature. Therefore, the contour description combines the simple global information with local feature points. Experiments with real contour curves extracted from 3-D fragments demonstrate that the proposed method is robust and efficient.展开更多
Flight delay prediction remains an important research topic due to dynamic nature in flight operation and numerous delay factors.Dynamic data-driven application system in the control area can provide a solution to thi...Flight delay prediction remains an important research topic due to dynamic nature in flight operation and numerous delay factors.Dynamic data-driven application system in the control area can provide a solution to this problem.However,in order to apply the approach,a state-space flight delay model needs to be established to represent the relationship among system states,as well as the relationship between system states and input/output variables.Based on the analysis of delay event sequence in a single flight,a state-space mixture model is established and input variables in the model are studied.Case study is also carried out on historical flight delay data.In addition,the genetic expectation-maximization(EM)algorithm is used to obtain the global optimal estimates of parameters in the mixture model,and results fit the historical data.At last,the model is validated in Kolmogorov-Smirnov tests.Results show that the model has reasonable goodness of fitting the data,and the search performance of traditional EM algorithm can be improved by using the genetic algorithm.展开更多
基金Supported by the Project of Guangdong Science and Technology Department(2020B010166005)the Post-Doctoral Research Project(Z000158)+2 种基金the Ministry of Education Social Science Fund(22YJ630167)the Fund project of Department of Science and Technology of Guangdong Province(GDK TP2021032500)the Guangdong Philosophy and Social Science(GD22YYJ15).
文摘The inter-agency government information sharing(IAGIS)plays an important role in improving service and efficiency of government agencies.Currently,there is still no effective and secure way for data-driven IAGIS to fulfill dynamic demands of information sharing between government agencies.Motivated by blockchain and data mining,a data-driven framework is proposed for IAGIS in this paper.Firstly,the blockchain is used as the core to design the whole framework for monitoring and preventing leakage and abuse of government information,in order to guarantee information security.Secondly,a four-layer architecture is designed for implementing the proposed framework.Thirdly,the classical data mining algorithms PageRank and Apriori are applied to dynamically design smart contracts for information sharing,for the purposed of flexibly adjusting the information sharing strategies according to the practical demands of government agencies for public management and public service.Finally,a case study is presented to illustrate the operation of the proposed framework.
基金Supported by the National Natural Science Foundation of China(62376214)the Natural Science Basic Research Program of Shaanxi(2023-JC-YB-533)Foundation of Ministry of Education Key Lab.of Cognitive Radio and Information Processing(Guilin University of Electronic Technology)(CRKL200203)。
文摘A modified multiple-component scattering power decomposition for analyzing polarimetric synthetic aperture radar(PolSAR)data is proposed.The modified decomposition involves two distinct steps.Firstly,ei⁃genvectors of the coherency matrix are used to modify the scattering models.Secondly,the entropy and anisotro⁃py of targets are used to improve the volume scattering power.With the guarantee of high double-bounce scatter⁃ing power in the urban areas,the proposed algorithm effectively improves the volume scattering power of vegeta⁃tion areas.The efficacy of the modified multiple-component scattering power decomposition is validated using ac⁃tual AIRSAR PolSAR data.The scattering power obtained through decomposing the original coherency matrix and the coherency matrix after orientation angle compensation is compared with three algorithms.Results from the experiment demonstrate that the proposed decomposition yields more effective scattering power for different PolSAR data sets.
基金supported by the Ensemble Grant for Early Career Researchers 2022-2023 and the 2023 Ensemble Continuation Grant of Tohoku University,the Hirose Foundation,and the AIMR Fusion Research Grantsupported by JSPS KAKENHI Nos.JP23K13599,JP23K13703,JP22H01803,JP18H05513,and JP23K13542.F.Y.and Q.W.acknowledge the China Scholarship Council(CSC)to support their studies in Japan.
文摘Magnesium(Mg)is a promising alternative to lithium(Li)as an anode material in solid-state batteries due to its abundance and high theoretical volumetric capacity.However,the sluggish Mg-ion conduction in the lattice of solidstate electrolytes(SSEs)is one of the key challenges that hamper the development of Mg-ion solid-state batteries.Though various Mg-ion SSEs have been reported in recent years,key insights are hard to be derived from a single literature report.Besides,the structure-performance relationships of Mg-ion SSEs need to be further unraveled to provide a more precise design guideline for SSEs.In this viewpoint article,we analyze the structural characteristics of the Mg-based SSEs with high ionic conductivity reported in the last four decades based upon data mining-we provide big-data-derived insights into the challenges and opportunities in developing next-generation Mg-ion SSEs.
基金supported in part by the National Natural Science Foundation of China(Nos.42271448,41701531)the Key Laboratory of Land Satellite Remote Sensing Application,Ministry of Natural Resources of the People’s Republic of China(No.KLSMNRG202317)。
文摘Under single-satellite observation,the parameter estimation of the boost phase of high-precision space noncooperative targets requires prior information.To improve the accuracy without prior information,we propose a parameter estimation model of the boost phase based on trajectory plane parametric cutting.The use of the plane passing through the geo-center and the cutting sequence line of sight(LOS)generates the trajectory-cutting plane.With the coefficient of the trajectory cutting plane directly used as the parameter to be estimated,a motion parameter estimation model in space non-cooperative targets is established,and the Gauss-Newton iteration method is used to solve the flight parameters.The experimental results show that the estimation algorithm proposed in this paper weakly relies on prior information and has higher estimation accuracy,providing a practical new idea and method for the parameter estimation of space non-cooperative targets under single-satellite warning.
基金Key research project of Hunan Provincial Administration of Traditional Chinese Medicine(A2023048)Key Research Foundation of Education Bureau of Hunan Province,China(23A0273).
文摘Objective Clinical medical record data associated with hepatitis B-related acute-on-chronic liver failure(HBV-ACLF)generally have small sample sizes and a class imbalance.However,most machine learning models are designed based on balanced data and lack interpretability.This study aimed to propose a traditional Chinese medicine(TCM)diagnostic model for HBV-ACLF based on the TCM syndrome differentiation and treatment theory,which is clinically interpretable and highly accurate.Methods We collected medical records from 261 patients diagnosed with HBV-ACLF,including three syndromes:Yang jaundice(214 cases),Yang-Yin jaundice(41 cases),and Yin jaundice(6 cases).To avoid overfitting of the machine learning model,we excluded the cases of Yin jaundice.After data standardization and cleaning,we obtained 255 relevant medical records of Yang jaundice and Yang-Yin jaundice.To address the class imbalance issue,we employed the oversampling method and five machine learning methods,including logistic regression(LR),support vector machine(SVM),decision tree(DT),random forest(RF),and extreme gradient boosting(XGBoost)to construct the syndrome diagnosis models.This study used precision,F1 score,the area under the receiver operating characteristic(ROC)curve(AUC),and accuracy as model evaluation metrics.The model with the best classification performance was selected to extract the diagnostic rule,and its clinical significance was thoroughly analyzed.Furthermore,we proposed a novel multiple-round stable rule extraction(MRSRE)method to obtain a stable rule set of features that can exhibit the model’s clinical interpretability.Results The precision of the five machine learning models built using oversampled balanced data exceeded 0.90.Among these models,the accuracy of RF classification of syndrome types was 0.92,and the mean F1 scores of the two categories of Yang jaundice and Yang-Yin jaundice were 0.93 and 0.94,respectively.Additionally,the AUC was 0.98.The extraction rules of the RF syndrome differentiation model based on the MRSRE method revealed that the common features of Yang jaundice and Yang-Yin jaundice were wiry pulse,yellowing of the urine,skin,and eyes,normal tongue body,healthy sublingual vessel,nausea,oil loathing,and poor appetite.The main features of Yang jaundice were a red tongue body and thickened sublingual vessels,whereas those of Yang-Yin jaundice were a dark tongue body,pale white tongue body,white tongue coating,lack of strength,slippery pulse,light red tongue body,slimy tongue coating,and abdominal distension.This is aligned with the classifications made by TCM experts based on TCM syndrome differentiation and treatment theory.Conclusion Our model can be utilized for differentiating HBV-ACLF syndromes,which has the potential to be applied to generate other clinically interpretable models with high accuracy on clinical data characterized by small sample sizes and a class imbalance.
基金Projects(41572317,51374242)supported by the National Natural Science Foundation of ChinaProject(2015CX005)supported by the Innovation Driven Plan of Central South University,China
文摘Data organization requires high efficiency for large amount of data applied in the digital mine system. A new method of storing massive data of block model is proposed to meet the characteristics of the database, including ACID-compliant, concurrency support, data sharing, and efficient access. Each block model is organized by linear octree, stored in LMDB(lightning memory-mapped database). Geological attribute can be queried at any point of 3D space by comparison algorithm of location code and conversion algorithm from address code of geometry space to location code of storage. The performance and robustness of querying geological attribute at 3D spatial region are enhanced greatly by the transformation from 3D to 2D and the method of 2D grid scanning to screen the inner and outer points. Experimental results showed that this method can access the massive data of block model, meeting the database characteristics. The method with LMDB is at least 3 times faster than that with etree, especially when it is used to read. In addition, the larger the amount of data is processed, the more efficient the method would be.
基金supported by the National Natural Science Foundation of China (No. 41004054) Research Fund for the Doctoral Program of Higher Education of China (No. 20105122120002)Natural Science Key Project, Sichuan Provincial Department of Education (No. 092A011)
文摘To fully extract and mine the multi-scale features of reservoirs and geologic structures in time/depth and space dimensions, a new 3D multi-scale volumetric curvature (MSVC) methodology is presented in this paper. We also propose a fast algorithm for computing 3D volumetric curvature. In comparison to conventional volumetric curvature attributes, its main improvements and key algorithms introduce multi-frequency components expansion in time-frequency domain and the corresponding multi-scale adaptive differential operator in the wavenumber domain, into the volumetric curvature calculation. This methodology can simultaneously depict seismic multi-scale features in both time and space. Additionally, we use data fusion of volumetric curvatures at various scales to take full advantage of the geologic features and anomalies extracted by curvature measurements at different scales. The 3D MSVC can highlight geologic anomalies and reduce noise at the same time. Thus, it improves the interpretation efficiency of curvature attributes analysis. The 3D MSVC is applied to both land and marine 3D seismic data. The results demonstrate that it can indicate the spatial distribution of reservoirs, detect faults and fracture zones, and identify their multi-scale properties.
基金supported by the National Natural Science Foundation of China (Grant No. 40974039)High-Tech Research and Development Program of China (Grant No.2006AA06205)Leading Strategic Project of Science and Technology, Chinese Academy of Sciences (XDA08020500)
文摘Traditionally, airborne time-domain electromagnetic (ATEM) data are inverted to derive the earth model by iteration. However, the data are often highly correlated among channels and consequently cause ill-posed and over-determined problems in the inversion. The correlation complicates the mapping relation between the ATEM data and the earth parameters and thus increases the inversion complexity. To obviate this, we adopt principal component analysis to transform ATEM data into orthogonal principal components (PCs) to reduce the correlations and the data dimensionality and simultaneously suppress the unrelated noise. In this paper, we use an artificial neural network (ANN) to approach the PCs mapping relation with the earth model parameters, avoiding the calculation of Jacobian derivatives. The PC-based ANN algorithm is applied to synthetic data for layered models compared with data-based ANN for airborne time-domain electromagnetic inversion. The results demonstrate the PC-based ANN advantages of simpler network structure, less training steps, and better inversion results over data-based ANN, especially for contaminated data. Furthermore, the PC-based ANN algorithm effectiveness is examined by the inversion of the pseudo 2D model and comparison with data-based ANN and Zhody's methods. The results indicate that PC-based ANN inversion can achieve a better agreement with the true model and also proved that PC-based ANN is feasible to invert large ATEM datasets.
文摘The location of U-turn bays is an important consideration in indirect driveway left-turn treatments.In order to improve the performance of right-turns followed by U-turns(RTUTs),this study evaluates the impacts of the separation distances between driveway exits and downstream U-turn locations on the safety and operational performance of vehicles making RTUTs.Crash data are investigated at 179 selected roadway segments,and travel time data are measured using video cameras at 29 locations in the state of Florida,USA.Crash rate models and travel time models are developed based on data collected in the field.It is found that the separation distance between driveway exits and downstream U-turn locations significantly impacts the safety and operational performance of vehicles making right turns followed by U-turns.Based on the research results,the minimum and optimal separation distances between driveways and U-turn locations under different roadway conditions are determined to facilitate driver use of RTUTs.The results of this study can be used for future intersection improvement projects in China.
基金supported by the National Natural Science Foundation of China (Grant No.41074133)
文摘Missing data are a problem in geophysical surveys, and interpolation and reconstruction of missing data is part of the data processing and interpretation. Based on the sparseness of the geophysical data or the transform domain, we can improve the accuracy and stability of the reconstruction by transforming it to a sparse optimization problem. In this paper, we propose a mathematical model for the sparse reconstruction of data based on the LO-norm minimization. Furthermore, we discuss two types of the approximation algorithm for the LO- norm minimization according to the size and characteristics of the geophysical data: namely, the iteratively reweighted least-squares algorithm and the fast iterative hard thresholding algorithm. Theoretical and numerical analysis showed that applying the iteratively reweighted least-squares algorithm to the reconstruction of potential field data exploits its fast convergence rate, short calculation time, and high precision, whereas the fast iterative hard thresholding algorithm is more suitable for processing seismic data, moreover, its computational efficiency is better than that of the traditional iterative hard thresholding algorithm.
基金supported by the Natural Science Foundation of China(No.61273179)Department of Education,Science and Technology Research Project of Hubei Province of China(No.D20131206,No.20141304)
文摘Regularization inversion uses constraints and a regularization factor to solve ill- posed inversion problems in geophysics. The choice of the regularization factor and of the initial model is critical in regularization inversion. To deal with these problems, we propose a multiobjective particle swarm inversion (MOPSOI) algorithm to simultaneously minimize the data misfit and model constraints, and obtain a multiobjective inversion solution set without the gradient information of the objective function and the regularization factor. We then choose the optimum solution from the solution set based on the trade-off between data misfit and constraints that substitute for the regularization factor. The inversion of synthetic two-dimensional magnetic data suggests that the MOPSOI algorithm can obtain as many feasible solutions as possible; thus, deeper insights of the inversion process can be gained and more reasonable solutions can be obtained by balancing the data misfit and constraints. The proposed MOPSOI algorithm can deal with the problems of choosing the right regularization factor and the initial model.
基金financially supported by the National Science and Technology Major Project of China(No.2011ZX05023-005-005)the National Natural Science Foundation of China(No.41274137)
文摘The conventional nonstationary convolutional model assumes that the seismic signal is recorded at normal incidence. Raw shot gathers are far from this assumption because of the effects of offsets. Because of such problems, we propose a novel prestack nonstationary deconvolution approach. We introduce the radial trace (RT) transform to the nonstationary deconvolution, we estimate the nonstationary deconvolution factor with hyperbolic smoothing based on variable-step sampling (VSS) in the RT domain, and we obtain the high-resolution prestack nonstationary deconvolution data. The RT transform maps the shot record from the offset and traveltime coordinates to those of apparent velocity and traveltime. The ray paths of the traces in the RT better satisfy the assumptions of the convolutional model. The proposed method combines the advantages of stationary deconvolution and inverse Q filtering, without prior information for Q. The nonstationary deconvolution in the RT domain is more suitable than that in the space-time (XT) domain for prestack data because it is the generalized extension of normal incidence. Tests with synthetic and real data demonstrate that the proposed method is more effective in compensating for large-offset and deep data.
基金supported partially by the National Natural Science Foundation of China (Nos. 40974004 and 40974016)Key Laboratory of Dynamic Geodesy of CAS, China (No. L09-01) R&I Team Support Program and the Graduate Science and Technology Foundation of SDUST, China (No. YCA110403)
文摘The HY-2 satellite carrying a satellite-borne GPS receiver is the first Chinese radar altimeter satellite, whose radial orbit determination precision must reach the centimeter level. Now HY-2 is in the test phase so that the observations are not openly released. In order to study the precise orbit determination precision and procedure for HY-2 based on the satellite- borne GPS technique, the satellite-borne GPS data are simulated in this paper. The HY-2 satellite-borne GPS antenna can receive at least seven GPS satellites each epoch, which can validate the GPS receiver and antenna design. What's more, the precise orbit determination processing flow is given and precise orbit determination experiments are conducted using the HY-2-borne GPS data with both the reduced-dynamic method and the kinematic geometry method. With the 1 and 3 mm phase data random errors, the radial orbit determination precision can achieve the centimeter level using these two methods and the kinematic orbit accuracy is slightly lower than that of the reduced-dynamic orbit. The earth gravity field model is an important factor which seriously affects the precise orbit determination of altimeter satellites. The reduced-dynamic orbit determination experiments are made with different earth gravity field models, such as EIGEN2, EGM96, TEG4, and GEMT3. Using a large number of high precision satellite-bome GPS data, the HY-2 precise orbit determination can reach the centimeter level with commonly used earth gravity field models up to above 50 degrees and orders.
基金Weaponry Equipment Pre-Research Foundation of PLA Equipment Ministry (No. 9140A06050409JB8102)Pre-Research Foundation of PLA University of Science and Technology (No. 2009JSJ11)
文摘To solve the query processing correctness problem for semantic-based relational data integration,the semantics of SAPRQL(simple protocol and RDF query language) queries is defined.In the course of query rewriting,all relative tables are found and decomposed into minimal connectable units.Minimal connectable units are joined according to semantic queries to produce the semantically correct query plans.Algorithms for query rewriting and transforming are presented.Computational complexity of the algorithms is discussed.Under the worst case,the query decomposing algorithm can be finished in O(n2) time and the query rewriting algorithm requires O(nm) time.And the performance of the algorithms is verified by experiments,and experimental results show that when the length of query is less than 8,the query processing algorithms can provide satisfactory performance.
文摘The molecular phylogeny of the Lardizabalaceae is reconstructed based on chloroplast trn L_F sequences alone and combined trn L_F and rbc L sequences. The phylogenetic topologies agree well with Qin's and Takhtajan's tribal classification in both analyses. Decaisneae and Sinofranchetieae are basal clades in the phylogenetic trees and external to all other taxa in the family. Lardizabaleae consisting of Boquila and Lardizabala are well supported in both trn L_F (100%) analysis and trn L_F and rbc L combined analysis (99%). Tribe Akebieae are strongly supported by a bootstrap value of 100% in both trn L_F analysis and trn L_F and rbc L combined analysis. However, the new genus Archakebia is nested within the genus Akebia in the trn L_F trees. In the combined trees, Archakebia is sister to Akebia with high bootstrap support. The inter_relationships among three closely related genera Parvatia , Holboellia and Stauntonia are still problematic. P. brunoniana ssp. elliptica is sister to H. latifolia in both analyses with low bootstrap support. H. parviflora is nested within the Stauntonia and sister to S. cavalerieana . Therefore, these three genera of tribe Akebieae may not be monophylytic and their generic boundary and delimitation need to be further studied, by exploring more molecular data, together with more morphological characters.
基金The National Natural Science Foundation of China(No.60425206,60633010)the High Technology Research and Development Program of Jiangsu Province(No.BG2005032)
文摘This paper suggests that a single class rather than methods should be used as the slice scope to compute class cohesion. First, for a given attribute, the statements in all methods that last define the attribute are computed. Then, the forward and backward data slices for this attribute are generated by using the class as the slice scope and are combined to compute the corresponding class data slice. Finally, the class cohesion is computed based on all class data slices for the attributes. Compared to traditional cohesion metrics that use methods as the slice scope, the proposed metrics that use a single class as slice scope take into account the possible interactions between the methods. The experimental results show that class cohesion can be more accurately measured when using the class as the slice scope.
基金Supported by the National Natural Science Foundation of China(60603029)the Natural Science Foundation of Jiangsu Province(BK2007074)the Natural Science Foundation for Colleges and Universities in Jiangsu Province(06KJB520132)~~
文摘One-class support vector machine (OCSVM) and support vector data description (SVDD) are two main domain-based one-class (kernel) classifiers. To reveal their relationship with density estimation in the case of the Gaussian kernel, OCSVM and SVDD are firstly unified into the framework of kernel density estimation, and the essential relationship between them is explicitly revealed. Then the result proves that the density estimation induced by OCSVM or SVDD is in agreement with the true density. Meanwhile, it can also reduce the integrated squared error (ISE). Finally, experiments on several simulated datasets verify the revealed relationships.
文摘An approach to contour extraction and feature point detection in the 3-D fragment reassembly is proposed. A simple and effective technique is used for building the intrinsic topology of the fragment data suitable for contour extraction. For the scanned data in which the topology is difficult to be achieved, the corresponding solutions are given to manage this problem. A robust approach is used for the curvature and torsion calculation of the discrete contour in a 3-D space. Finally, a method is developed for detecting feature points of the fragment contour based on total curvature. Therefore, the contour description combines the simple global information with local feature points. Experiments with real contour curves extracted from 3-D fragments demonstrate that the proposed method is robust and efficient.
基金Supported by the High Technology Research and Development Programme of China(2006AA12A106)~~
文摘Flight delay prediction remains an important research topic due to dynamic nature in flight operation and numerous delay factors.Dynamic data-driven application system in the control area can provide a solution to this problem.However,in order to apply the approach,a state-space flight delay model needs to be established to represent the relationship among system states,as well as the relationship between system states and input/output variables.Based on the analysis of delay event sequence in a single flight,a state-space mixture model is established and input variables in the model are studied.Case study is also carried out on historical flight delay data.In addition,the genetic expectation-maximization(EM)algorithm is used to obtain the global optimal estimates of parameters in the mixture model,and results fit the historical data.At last,the model is validated in Kolmogorov-Smirnov tests.Results show that the model has reasonable goodness of fitting the data,and the search performance of traditional EM algorithm can be improved by using the genetic algorithm.