The inter-city linkage heat data provided by Baidu Migration is employed as a characterization of inter-city linkages in order to facilitate the study of the network linkage characteristics and hierarchical structure ...The inter-city linkage heat data provided by Baidu Migration is employed as a characterization of inter-city linkages in order to facilitate the study of the network linkage characteristics and hierarchical structure of urban agglomeration in the Greater Bay Area through the use of social network analysis method.This is the inaugural application of big data based on location services in the study of urban agglomeration network structure,which represents a novel research perspective on this topic.The study reveals that the density of network linkages in the Greater Bay Area urban agglomeration has reached 100%,indicating a mature network-like spatial structure.This structure has given rise to three distinct communities:Shenzhen-Dongguan-Huizhou,Guangzhou-Foshan-Zhaoqing,and Zhuhai-Zhongshan-Jiangmen.Additionally,cities within the Greater Bay Area urban agglomeration play different roles,suggesting that varying development strategies may be necessary to achieve staggered development.The study demonstrates that large datasets represented by LBS can offer novel insights and methodologies for the examination of urban agglomeration network structures,contingent on the appropriate mining and processing of the data.展开更多
The South Yellow Sea basin is filled with Mesozoic-Cenozoic continental sediments overlying pre-Palaeozoic and Mesozoic-Palaeozoic marine sediments.Conventional multi-channel seismic data cannot describe the velocity ...The South Yellow Sea basin is filled with Mesozoic-Cenozoic continental sediments overlying pre-Palaeozoic and Mesozoic-Palaeozoic marine sediments.Conventional multi-channel seismic data cannot describe the velocity structure of the marine residual basin in detail,leading to the lack of a deeper understanding of the distribution and lithology owing to strong energy shielding on the top interface of marine sediments.In this study,we present seismic tomography data from ocean bottom seismographs that describe the NEE-trending velocity distributions of the basin.The results indicate that strong velocity variations occur at shallow crustal levels.Horizontal velocity bodies show good correlation with surface geological features,and multi-layer features exist in the vertical velocity framework(depth:0–10 km).The analyses of the velocity model,gravity data,magnetic data,multichannel seismic profiles,and drilling data showed that high-velocity anomalies(>6.5 km/s)of small(thickness:1–2 km)and large(thickness:>5 km)scales were caused by igneous complexes in the multi-layer structure,which were active during the Palaeogene.Possible locations of good Mesozoic and Palaeozoic marine strata are limited to the Central Uplift and the western part of the Northern Depression along the wide-angle ocean bottom seismograph array.Following the Indosinian movement,a strong compression existed in the Northern Depression during the extensional phase that caused the formation of folds in the middle of the survey line.This study is useful for reconstructing the regional tectonic evolution and delineating the distribution of the marine residual basin in the South Yellow Sea basin.展开更多
Data protection in databases is critical for any organization,as unauthorized access or manipulation can have severe negative consequences.Intrusion detection systems are essential for keeping databases secure.Advance...Data protection in databases is critical for any organization,as unauthorized access or manipulation can have severe negative consequences.Intrusion detection systems are essential for keeping databases secure.Advancements in technology will lead to significant changes in the medical field,improving healthcare services through real-time information sharing.However,reliability and consistency still need to be solved.Safeguards against cyber-attacks are necessary due to the risk of unauthorized access to sensitive information and potential data corruption.Dis-ruptions to data items can propagate throughout the database,making it crucial to reverse fraudulent transactions without delay,especially in the healthcare industry,where real-time data access is vital.This research presents a role-based access control architecture for an anomaly detection technique.Additionally,the Structured Query Language(SQL)queries are stored in a new data structure called Pentaplet.These pentaplets allow us to maintain the correlation between SQL statements within the same transaction by employing the transaction-log entry information,thereby increasing detection accuracy,particularly for individuals within the company exhibiting unusual behavior.To identify anomalous queries,this system employs a supervised machine learning technique called Support Vector Machine(SVM).According to experimental findings,the proposed model performed well in terms of detection accuracy,achieving 99.92%through SVM with One Hot Encoding and Principal Component Analysis(PCA).展开更多
To capitalize on the primary role of major course teaching and to facilitate students’understanding of abstract concepts in the data structure course,it is essential to increase their interest in learning and develop...To capitalize on the primary role of major course teaching and to facilitate students’understanding of abstract concepts in the data structure course,it is essential to increase their interest in learning and develop case studies that highlight fine traditional culture.By incorporating these culture-rich case studies into classroom instruction,we employ a project-driven teaching approach.This not only allows students to master professional knowledge,but also enhances their abilities to solve specific engineering problems,ultimately fostering cultural confidence.Over the past few years,during which educational reforms have been conducted for trial runs,the feasibility and effectiveness of these reform schemes have been demonstrated.展开更多
Bayesian networks are a powerful class of graphical decision models used to represent causal relationships among variables.However,the reliability and integrity of learned Bayesian network models are highly dependent ...Bayesian networks are a powerful class of graphical decision models used to represent causal relationships among variables.However,the reliability and integrity of learned Bayesian network models are highly dependent on the quality of incoming data streams.One of the primary challenges with Bayesian networks is their vulnerability to adversarial data poisoning attacks,wherein malicious data is injected into the training dataset to negatively influence the Bayesian network models and impair their performance.In this research paper,we propose an efficient framework for detecting data poisoning attacks against Bayesian network structure learning algorithms.Our framework utilizes latent variables to quantify the amount of belief between every two nodes in each causal model over time.We use our innovative methodology to tackle an important issue with data poisoning assaults in the context of Bayesian networks.With regard to four different forms of data poisoning attacks,we specifically aim to strengthen the security and dependability of Bayesian network structure learning techniques,such as the PC algorithm.By doing this,we explore the complexity of this area and offer workablemethods for identifying and reducing these sneaky dangers.Additionally,our research investigates one particular use case,the“Visit to Asia Network.”The practical consequences of using uncertainty as a way to spot cases of data poisoning are explored in this inquiry,which is of utmost relevance.Our results demonstrate the promising efficacy of latent variables in detecting and mitigating the threat of data poisoning attacks.Additionally,our proposed latent-based framework proves to be sensitive in detecting malicious data poisoning attacks in the context of stream data.展开更多
The Dialafara area is part of the highly endowed Kédougou-Kéniéba Inlier (KKI), West-Malian gold belt, which corresponds to a Paleoproterozoic window through the West African Craton (WAC). This study pr...The Dialafara area is part of the highly endowed Kédougou-Kéniéba Inlier (KKI), West-Malian gold belt, which corresponds to a Paleoproterozoic window through the West African Craton (WAC). This study presents, first of all, an integration of geophysical data interpretation with litho-structural field reconnaissance and then proposes a new litho-structural map of the Dialafara area. The Dialafara area shows a variety of lithology characterized by volcanic and volcano-sedimentary units, metasediments and plutonic intrusion. These lithologies were affected by a complex superposition of structures of unequal importance defining three deformation phases (D<sub>D1</sub> to D<sub>D3</sub>) under ductile to brittle regimes. These features permit to portray a new litho-structural map, which shows that the Dialafara area presents a more complex lithological and structural context than the one presented in regional map of the KKI. This leads to the evidence that this area could be a potential site for exploration as it is situated between two world-class gold districts.展开更多
The research purpose of this dissertation is threefold: to innovate artificial intelligence methods, to create the intersection of artificial intelligence and biological research, and to innovate human methodology. Th...The research purpose of this dissertation is threefold: to innovate artificial intelligence methods, to create the intersection of artificial intelligence and biological research, and to innovate human methodology. The work I have done in my research includes: improving logical structure and logical engineering, using my theory to study the innovation of the development path of artificial intelligence, using my theory to create biomimetic logic, a new intersection of artificial intelligence and biological research, and exploring the innovation of human methodology through the previous two works. The results of the research are as follows: 1) Introduction to bionic logic, incorporating simulations of people, society, and life as core principles. 2) Definition of the logical structure as the primary focus of research, with logic mechanics serving as foundational research principles. 3) Examination of the logical structure’s environment through logical fields and networks. 4) Study of logical structure communication via logical networks and main lines. 5) Proposal of data logic. 6) Investigation into the logic of logical structures, employing structural diagrams of logical equations. 7) Development of a theory of life activity within logical structures, encompassing information reasoning, its corresponding control structure, and structural reasoning. 8) Introduction of the lifecycle theory for logical structures and examination of the clock equation. 9) Exploration of logical structure intelligence. 10) Study of logical structures in mathematical forms. 11) Introduction of logic engineering. 12) Examination of artificial intelligence’s significance. 13) Investigation into the significance of human methodology.展开更多
With the deepening of the Guangdong-Hong Kong-Macao Greater Bay Area strategy and the accelerated integration and development of the east and west sides of the Pearl River Estuary,Zhuhai’s hub position is becoming mo...With the deepening of the Guangdong-Hong Kong-Macao Greater Bay Area strategy and the accelerated integration and development of the east and west sides of the Pearl River Estuary,Zhuhai’s hub position is becoming more and more prominent.The city of Zhuhai has a dense water network and is divided into two urban areas,the east and the west,under the influence of the Mordor Gate waterway.Based on the theory of spatial syntax,this paper carries out an analytical study on the urban spatial structure of Zhuhai,identifies the distribution characteristics of urban POIs,and provides theoretical support for the urban development of Zhuhai.展开更多
A robust and efficient algorithm is presented to build multiresolution models (MRMs) of arbitrary meshes without requirement of subdivision connectivity. To overcome the sampling difficulty of arbitrary meshes, edge c...A robust and efficient algorithm is presented to build multiresolution models (MRMs) of arbitrary meshes without requirement of subdivision connectivity. To overcome the sampling difficulty of arbitrary meshes, edge contraction and vertex expansion are used as downsampling and upsampling methods. Our MRMs of a mesh are composed of a base mesh and a series of edge split operations, which are organized as a directed graph. Each split operation encodes two parts of information. One is the modification to the mesh, and the other is the dependency relation among splits. Such organization ensures the efficiency and robustness of our MRM algorithm. Examples demonstrate the functionality of our method.展开更多
More web pages are widely applying AJAX (Asynchronous JavaScript XML) due to the rich interactivity and incremental communication. By observing, it is found that the AJAX contents, which could not be seen by traditi...More web pages are widely applying AJAX (Asynchronous JavaScript XML) due to the rich interactivity and incremental communication. By observing, it is found that the AJAX contents, which could not be seen by traditional crawler, are well-structured and belong to one specific domain generally. Extracting the structured data from AJAX contents and annotating its semantic are very significant for further applications. In this paper, a structured AJAX data extraction method for agricultural domain based on agricultural ontology was proposed. Firstly, Crawljax, an open AJAX crawling tool, was overridden to explore and retrieve the AJAX contents; secondly, the retrieved contents were partitioned into items and then classified by combining with agricultural ontology. HTML tags and punctuations were used to segment the retrieved contents into entity items. Finally, the entity items were clustered and the semantic annotation was assigned to clustering results according to agricultural ontology. By experimental evaluation, the proposed approach was proved effectively in resource exploring, entity extraction, and semantic annotation.展开更多
In this paper, a new concept called numerical structure of seismic data is introduced and the difference between numerical structure and numerical value of seismic data is explained. Our study shows that the numerical...In this paper, a new concept called numerical structure of seismic data is introduced and the difference between numerical structure and numerical value of seismic data is explained. Our study shows that the numerical seismic structure is closely related to oil and gas-bearing reservoir, so it is very useful for a geologist or a geophysicist to precisely interpret the oil-bearing layers from the seismic data. This technology can be applied to any exploration or production stage. The new method has been tested on a series of exploratory or development wells and proved to be reliable in China. Hydrocarbon-detection with this new method for 39 exploration wells on 25 structures indi- cates a success ratio of over 80 percent. The new method of hydrocarbon prediction can be applied for: (1) depositional environment of reservoirs with marine fades, delta, or non-marine fades (including fluvial facies, lacustrine fades); (2) sedimentary rocks of reservoirs that are non-marine clastic rocks and carbonate rock; and (3) burial depths range from 300 m to 7000 m, and the minimum thickness of these reservoirs is over 8 m (main frequency is about 50 Hz).展开更多
In order to improve the quality of web search,a new query expansion method by choosing meaningful structure data from a domain database is proposed.It categories attributes into three different classes,named as concep...In order to improve the quality of web search,a new query expansion method by choosing meaningful structure data from a domain database is proposed.It categories attributes into three different classes,named as concept attribute,context attribute and meaningless attribute,according to their semantic features which are document frequency features and distinguishing capability features.It also defines the semantic relevance between two attributes when they have correlations in the database.Then it proposes trie-bitmap structure and pair pointer tables to implement efficient algorithms for discovering attribute semantic feature and detecting their semantic relevances.By using semantic attributes and their semantic relevances,expansion words can be generated and embedded into a vector space model with interpolation parameters.The experiments use an IMDB movie database and real texts collections to evaluate the proposed method by comparing its performance with a classical vector space model.The results show that the proposed method can improve text search efficiently and also improve both semantic features and semantic relevances with good separation capabilities.展开更多
Seismic data structure characteristics means the waveform character arranged in the time sequence at discrete data points in each 2-D or 3-D seismic trace. Hydrocarbon prediction using seismic data structure character...Seismic data structure characteristics means the waveform character arranged in the time sequence at discrete data points in each 2-D or 3-D seismic trace. Hydrocarbon prediction using seismic data structure characteristics is a new reservoir prediction technique. When the main pay interval is in carbonate fracture and fissure-cavern type reservoirs with very strong inhomogeneity, there are some difficulties with hydrocarbon prediction. Because of the special geological conditions of the eighth zone in the Tahe oil field, we apply seismic data structure characteristics to hydrocarbon prediction for the Ordovician reservoir in this zone. We divide the area oil zone into favorable and unfavorable blocks. Eighteen well locations were proposed in the favorable oil block, drilled, and recovered higher output of oil and gas.展开更多
Aiming to increase the efficiency of gem design and manufacturing, a new method in computer-aided-design (CAD) of convex faceted gem cuts (CFGC) based on Half-edge data structure (HDS), including the algorithms for th...Aiming to increase the efficiency of gem design and manufacturing, a new method in computer-aided-design (CAD) of convex faceted gem cuts (CFGC) based on Half-edge data structure (HDS), including the algorithms for the implementation is presented in this work. By using object-oriented methods, geometrical elements of CFGC are classified and responding geometrical feature classes are established. Each class is implemented and embedded based on the gem process. Matrix arithmetic and analytical geometry are used to derive the affine transformation and the cutting algorithm. Based on the demand for a diversity of gem cuts, CAD functions both for free-style faceted cuts and parametric designs of typical cuts and visualization and human-computer interactions of the CAD system including two-dimensional and three-dimensional interactions have been realized which enhances the flexibility and universality of the CAD system. Furthermore, data in this CAD system can also be used directly by the gem CAM module, which will promote the gem CAD/CAM integration.展开更多
In conjunction with association rules for data mining, the connections between testing indices and strong and weak association rules were determined, and new derivative rules were obtained by further reasoning. Associ...In conjunction with association rules for data mining, the connections between testing indices and strong and weak association rules were determined, and new derivative rules were obtained by further reasoning. Association rules were used to analyze correlation and check consistency between indices. This study shows that the judgment obtained by weak association rules or non-association rules is more accurate and more credible than that obtained by strong association rules. When the testing grades of two indices in the weak association rules are inconsistent, the testing grades of indices are more likely to be erroneous, and the mistakes are often caused by human factors. Clustering data mining technology was used to analyze the reliability of a diagnosis, or to perform health diagnosis directly. Analysis showed that the clustering results are related to the indices selected, and that if the indices selected are more significant, the characteristics of clustering results are also more significant, and the analysis or diagnosis is more credible. The indices and diagnosis analysis function produced by this study provide a necessary theoretical foundation and new ideas for the development of hydraulic metal structure health diagnosis technology.展开更多
The distribution of oil and gas resources in the South China Sea and adjacent areas is closely related to the structural pattern that helped to define the controlling effect of deep processes on oil-bearing basins.Ign...The distribution of oil and gas resources in the South China Sea and adjacent areas is closely related to the structural pattern that helped to define the controlling effect of deep processes on oil-bearing basins.Igneous rocks can record important information from deep processes.Deep structures such as faults,basin uplift and depression,Cenozoic basement and magnetic basement are all the results of energy exchange within the earth.The study of the relationship between igneous rocks and deep structures is of great significance for the study of the South China Sea.By using the minimum curvature potential field separation technique and the correlation analysis technique of gravitational and magnetic anomalies,the fusion of gravitational and magnetic data reflecting igneous rocks can be obtained,through which the igneous rocks with high susceptibility/high density or high susceptibility/low density can be identified.In this study area,igneous rocks do not develop in the Yinggehai basin,Qiongdongnan basin,Zengmu basin and Brunei-Sabah basin whilst igneous rocks with high susceptibility/high density or high susceptibility/low density are widely-developed in other basins.In undeveloped igneous areas,faults are also undeveloped the Cenozoic thickness is greater,the magnetic basement depth is greater and the Cenozoic thickness is highly positively correlated with the magnetic basement depth.In igneously developed regions,the distribution pattern of the Qiongtai block is mainly controlled by primary faults,while the distribution of the Zhongxisha block,Xunta block and Yongshu-Taiping block is mainly controlled by secondary faults,the Cenozoic thickness having a low correlation with the depth of the magnetic basement.展开更多
In a sensor network with a large number of densely populated sensor nodes, a single target of interest may be detected by multiple sensor nodes simultaneously. Data collected from the sensor nodes are usually highly c...In a sensor network with a large number of densely populated sensor nodes, a single target of interest may be detected by multiple sensor nodes simultaneously. Data collected from the sensor nodes are usually highly correlated, and hence energy saving using in-network data fusion becomes possible. A traditional data fusion scheme starts with dividing the network into clusters, followed by electing a sensor node as cluster head in each cluster. A cluster head is responsible for collecting data from all its cluster members, performing data fusion on these data and transmitting the fused data to the base station. Assuming that a sensor node is only capable of handling a single node-to-node transmission at a time and each transmission takes T time-slots, a cluster head with n cluster members will take at least nT time-slots to collect data from all its cluster members. In this paper, a tree-based network structure and its formation algorithms are proposed. Simulation results show that the proposed network structure can greatly reduce the delay in data collection.展开更多
Taking autonomous driving and driverless as the research object,we discuss and define intelligent high-precision map.Intelligent high-precision map is considered as a key link of future travel,a carrier of real-time p...Taking autonomous driving and driverless as the research object,we discuss and define intelligent high-precision map.Intelligent high-precision map is considered as a key link of future travel,a carrier of real-time perception of traffic resources in the entire space-time range,and the criterion for the operation and control of the whole process of the vehicle.As a new form of map,it has distinctive features in terms of cartography theory and application requirements compared with traditional navigation electronic maps.Thus,it is necessary to analyze and discuss its key features and problems to promote the development of research and application of intelligent high-precision map.Accordingly,we propose an information transmission model based on the cartography theory and combine the wheeled robot’s control flow in practical application.Next,we put forward the data logic structure of intelligent high-precision map,and analyze its application in autonomous driving.Then,we summarize the computing mode of“Crowdsourcing+Edge-Cloud Collaborative Computing”,and carry out key technical analysis on how to improve the quality of crowdsourced data.We also analyze the effective application scenarios of intelligent high-precision map in the future.Finally,we present some thoughts and suggestions for the future development of this field.展开更多
The key to develop 3-D GISs is the study on 3-D data model and data structure. Some of the data models and data structures have been presented by scholars. Because of the complexity of 3-D spatial phenomenon, there ar...The key to develop 3-D GISs is the study on 3-D data model and data structure. Some of the data models and data structures have been presented by scholars. Because of the complexity of 3-D spatial phenomenon, there are no perfect data structures that can describe all spatial entities. Every data structure has its own advantages and disadvantages. It is difficult to design a single data structure to meet different needs. The important subject in the3-D data models is developing a data model that has integrated vector and raster data structures. A special 3-D spatial data model based on distributing features of spatial entities should be designed. We took the geological exploration engineering as the research background and designed an integrated data model whose data structures integrats vector and raster data byadopting object-oriented technique. Research achievements are presented in this paper.展开更多
文摘The inter-city linkage heat data provided by Baidu Migration is employed as a characterization of inter-city linkages in order to facilitate the study of the network linkage characteristics and hierarchical structure of urban agglomeration in the Greater Bay Area through the use of social network analysis method.This is the inaugural application of big data based on location services in the study of urban agglomeration network structure,which represents a novel research perspective on this topic.The study reveals that the density of network linkages in the Greater Bay Area urban agglomeration has reached 100%,indicating a mature network-like spatial structure.This structure has given rise to three distinct communities:Shenzhen-Dongguan-Huizhou,Guangzhou-Foshan-Zhaoqing,and Zhuhai-Zhongshan-Jiangmen.Additionally,cities within the Greater Bay Area urban agglomeration play different roles,suggesting that varying development strategies may be necessary to achieve staggered development.The study demonstrates that large datasets represented by LBS can offer novel insights and methodologies for the examination of urban agglomeration network structures,contingent on the appropriate mining and processing of the data.
基金The National Natural Science Foundation of China under contract No.41806048the Open Fund of the Hubei Key Laboratory of Marine Geological Resources under contract No.MGR202009+2 种基金the Fund from the Key Laboratory of Deep-Earth Dynamics of Ministry of Natural Resource,Institute of Geology,Chinese Academy of Geological Sciences under contract No.J1901-16the Aoshan Science and Technology Innovation Project of Pilot National Laboratory for Marine Science and Technology(Qingdao)under contract No.2015ASKJ03-Seabed Resourcesthe Fund from the Korea Institute of Ocean Science and Technology(KIOST)under contract No.PE99741.
文摘The South Yellow Sea basin is filled with Mesozoic-Cenozoic continental sediments overlying pre-Palaeozoic and Mesozoic-Palaeozoic marine sediments.Conventional multi-channel seismic data cannot describe the velocity structure of the marine residual basin in detail,leading to the lack of a deeper understanding of the distribution and lithology owing to strong energy shielding on the top interface of marine sediments.In this study,we present seismic tomography data from ocean bottom seismographs that describe the NEE-trending velocity distributions of the basin.The results indicate that strong velocity variations occur at shallow crustal levels.Horizontal velocity bodies show good correlation with surface geological features,and multi-layer features exist in the vertical velocity framework(depth:0–10 km).The analyses of the velocity model,gravity data,magnetic data,multichannel seismic profiles,and drilling data showed that high-velocity anomalies(>6.5 km/s)of small(thickness:1–2 km)and large(thickness:>5 km)scales were caused by igneous complexes in the multi-layer structure,which were active during the Palaeogene.Possible locations of good Mesozoic and Palaeozoic marine strata are limited to the Central Uplift and the western part of the Northern Depression along the wide-angle ocean bottom seismograph array.Following the Indosinian movement,a strong compression existed in the Northern Depression during the extensional phase that caused the formation of folds in the middle of the survey line.This study is useful for reconstructing the regional tectonic evolution and delineating the distribution of the marine residual basin in the South Yellow Sea basin.
基金thankful to the Dean of Scientific Research at Najran University for funding this work under the Research Groups Funding Program,Grant Code(NU/RG/SERC/12/6).
文摘Data protection in databases is critical for any organization,as unauthorized access or manipulation can have severe negative consequences.Intrusion detection systems are essential for keeping databases secure.Advancements in technology will lead to significant changes in the medical field,improving healthcare services through real-time information sharing.However,reliability and consistency still need to be solved.Safeguards against cyber-attacks are necessary due to the risk of unauthorized access to sensitive information and potential data corruption.Dis-ruptions to data items can propagate throughout the database,making it crucial to reverse fraudulent transactions without delay,especially in the healthcare industry,where real-time data access is vital.This research presents a role-based access control architecture for an anomaly detection technique.Additionally,the Structured Query Language(SQL)queries are stored in a new data structure called Pentaplet.These pentaplets allow us to maintain the correlation between SQL statements within the same transaction by employing the transaction-log entry information,thereby increasing detection accuracy,particularly for individuals within the company exhibiting unusual behavior.To identify anomalous queries,this system employs a supervised machine learning technique called Support Vector Machine(SVM).According to experimental findings,the proposed model performed well in terms of detection accuracy,achieving 99.92%through SVM with One Hot Encoding and Principal Component Analysis(PCA).
基金the research outcomes of a blended top-tier undergraduate course in Henan ProvinceData Structures and Algorithms(Jiao Gao[2022]324)a research-based teaching demonstration course in Henan Province-Data Structures and Algorithms(Jiao Gao[2023]36)a model course of ideological and political education of Anyang Normal University-Data Structures and Algorithms(No.YBKC20210012)。
文摘To capitalize on the primary role of major course teaching and to facilitate students’understanding of abstract concepts in the data structure course,it is essential to increase their interest in learning and develop case studies that highlight fine traditional culture.By incorporating these culture-rich case studies into classroom instruction,we employ a project-driven teaching approach.This not only allows students to master professional knowledge,but also enhances their abilities to solve specific engineering problems,ultimately fostering cultural confidence.Over the past few years,during which educational reforms have been conducted for trial runs,the feasibility and effectiveness of these reform schemes have been demonstrated.
文摘Bayesian networks are a powerful class of graphical decision models used to represent causal relationships among variables.However,the reliability and integrity of learned Bayesian network models are highly dependent on the quality of incoming data streams.One of the primary challenges with Bayesian networks is their vulnerability to adversarial data poisoning attacks,wherein malicious data is injected into the training dataset to negatively influence the Bayesian network models and impair their performance.In this research paper,we propose an efficient framework for detecting data poisoning attacks against Bayesian network structure learning algorithms.Our framework utilizes latent variables to quantify the amount of belief between every two nodes in each causal model over time.We use our innovative methodology to tackle an important issue with data poisoning assaults in the context of Bayesian networks.With regard to four different forms of data poisoning attacks,we specifically aim to strengthen the security and dependability of Bayesian network structure learning techniques,such as the PC algorithm.By doing this,we explore the complexity of this area and offer workablemethods for identifying and reducing these sneaky dangers.Additionally,our research investigates one particular use case,the“Visit to Asia Network.”The practical consequences of using uncertainty as a way to spot cases of data poisoning are explored in this inquiry,which is of utmost relevance.Our results demonstrate the promising efficacy of latent variables in detecting and mitigating the threat of data poisoning attacks.Additionally,our proposed latent-based framework proves to be sensitive in detecting malicious data poisoning attacks in the context of stream data.
文摘The Dialafara area is part of the highly endowed Kédougou-Kéniéba Inlier (KKI), West-Malian gold belt, which corresponds to a Paleoproterozoic window through the West African Craton (WAC). This study presents, first of all, an integration of geophysical data interpretation with litho-structural field reconnaissance and then proposes a new litho-structural map of the Dialafara area. The Dialafara area shows a variety of lithology characterized by volcanic and volcano-sedimentary units, metasediments and plutonic intrusion. These lithologies were affected by a complex superposition of structures of unequal importance defining three deformation phases (D<sub>D1</sub> to D<sub>D3</sub>) under ductile to brittle regimes. These features permit to portray a new litho-structural map, which shows that the Dialafara area presents a more complex lithological and structural context than the one presented in regional map of the KKI. This leads to the evidence that this area could be a potential site for exploration as it is situated between two world-class gold districts.
文摘The research purpose of this dissertation is threefold: to innovate artificial intelligence methods, to create the intersection of artificial intelligence and biological research, and to innovate human methodology. The work I have done in my research includes: improving logical structure and logical engineering, using my theory to study the innovation of the development path of artificial intelligence, using my theory to create biomimetic logic, a new intersection of artificial intelligence and biological research, and exploring the innovation of human methodology through the previous two works. The results of the research are as follows: 1) Introduction to bionic logic, incorporating simulations of people, society, and life as core principles. 2) Definition of the logical structure as the primary focus of research, with logic mechanics serving as foundational research principles. 3) Examination of the logical structure’s environment through logical fields and networks. 4) Study of logical structure communication via logical networks and main lines. 5) Proposal of data logic. 6) Investigation into the logic of logical structures, employing structural diagrams of logical equations. 7) Development of a theory of life activity within logical structures, encompassing information reasoning, its corresponding control structure, and structural reasoning. 8) Introduction of the lifecycle theory for logical structures and examination of the clock equation. 9) Exploration of logical structure intelligence. 10) Study of logical structures in mathematical forms. 11) Introduction of logic engineering. 12) Examination of artificial intelligence’s significance. 13) Investigation into the significance of human methodology.
基金funded by The Guangdong Province General Universities Young Innovative Talent Project(Grant No.2023WQNCX122)The Zhuhai Philosophy and Social Science Planning Project(Grant No.2023YBB049)。
文摘With the deepening of the Guangdong-Hong Kong-Macao Greater Bay Area strategy and the accelerated integration and development of the east and west sides of the Pearl River Estuary,Zhuhai’s hub position is becoming more and more prominent.The city of Zhuhai has a dense water network and is divided into two urban areas,the east and the west,under the influence of the Mordor Gate waterway.Based on the theory of spatial syntax,this paper carries out an analytical study on the urban spatial structure of Zhuhai,identifies the distribution characteristics of urban POIs,and provides theoretical support for the urban development of Zhuhai.
文摘A robust and efficient algorithm is presented to build multiresolution models (MRMs) of arbitrary meshes without requirement of subdivision connectivity. To overcome the sampling difficulty of arbitrary meshes, edge contraction and vertex expansion are used as downsampling and upsampling methods. Our MRMs of a mesh are composed of a base mesh and a series of edge split operations, which are organized as a directed graph. Each split operation encodes two parts of information. One is the modification to the mesh, and the other is the dependency relation among splits. Such organization ensures the efficiency and robustness of our MRM algorithm. Examples demonstrate the functionality of our method.
基金supported by the Knowledge Innovation Program of the Chinese Academy of Sciencesthe National High-Tech R&D Program of China(2008BAK49B05)
文摘More web pages are widely applying AJAX (Asynchronous JavaScript XML) due to the rich interactivity and incremental communication. By observing, it is found that the AJAX contents, which could not be seen by traditional crawler, are well-structured and belong to one specific domain generally. Extracting the structured data from AJAX contents and annotating its semantic are very significant for further applications. In this paper, a structured AJAX data extraction method for agricultural domain based on agricultural ontology was proposed. Firstly, Crawljax, an open AJAX crawling tool, was overridden to explore and retrieve the AJAX contents; secondly, the retrieved contents were partitioned into items and then classified by combining with agricultural ontology. HTML tags and punctuations were used to segment the retrieved contents into entity items. Finally, the entity items were clustered and the semantic annotation was assigned to clustering results according to agricultural ontology. By experimental evaluation, the proposed approach was proved effectively in resource exploring, entity extraction, and semantic annotation.
基金Mainly presented at the 6-th international meeting of acoustics in Aug. 2003, and The 1999 SPE Asia Pacific Oil and GasConference and Exhibition held in Jakarta, Indonesia, 20-22 April 1999, SPE 54274.
文摘In this paper, a new concept called numerical structure of seismic data is introduced and the difference between numerical structure and numerical value of seismic data is explained. Our study shows that the numerical seismic structure is closely related to oil and gas-bearing reservoir, so it is very useful for a geologist or a geophysicist to precisely interpret the oil-bearing layers from the seismic data. This technology can be applied to any exploration or production stage. The new method has been tested on a series of exploratory or development wells and proved to be reliable in China. Hydrocarbon-detection with this new method for 39 exploration wells on 25 structures indi- cates a success ratio of over 80 percent. The new method of hydrocarbon prediction can be applied for: (1) depositional environment of reservoirs with marine fades, delta, or non-marine fades (including fluvial facies, lacustrine fades); (2) sedimentary rocks of reservoirs that are non-marine clastic rocks and carbonate rock; and (3) burial depths range from 300 m to 7000 m, and the minimum thickness of these reservoirs is over 8 m (main frequency is about 50 Hz).
基金Program for New Century Excellent Talents in University(No.NCET-06-0290)the National Natural Science Foundation of China(No.60503036)the Fok Ying Tong Education Foundation Award(No.104027)
文摘In order to improve the quality of web search,a new query expansion method by choosing meaningful structure data from a domain database is proposed.It categories attributes into three different classes,named as concept attribute,context attribute and meaningless attribute,according to their semantic features which are document frequency features and distinguishing capability features.It also defines the semantic relevance between two attributes when they have correlations in the database.Then it proposes trie-bitmap structure and pair pointer tables to implement efficient algorithms for discovering attribute semantic feature and detecting their semantic relevances.By using semantic attributes and their semantic relevances,expansion words can be generated and embedded into a vector space model with interpolation parameters.The experiments use an IMDB movie database and real texts collections to evaluate the proposed method by comparing its performance with a classical vector space model.The results show that the proposed method can improve text search efficiently and also improve both semantic features and semantic relevances with good separation capabilities.
基金This reservoir research is sponsored by the National 973 Subject Project (No. 2001CB209).
文摘Seismic data structure characteristics means the waveform character arranged in the time sequence at discrete data points in each 2-D or 3-D seismic trace. Hydrocarbon prediction using seismic data structure characteristics is a new reservoir prediction technique. When the main pay interval is in carbonate fracture and fissure-cavern type reservoirs with very strong inhomogeneity, there are some difficulties with hydrocarbon prediction. Because of the special geological conditions of the eighth zone in the Tahe oil field, we apply seismic data structure characteristics to hydrocarbon prediction for the Ordovician reservoir in this zone. We divide the area oil zone into favorable and unfavorable blocks. Eighteen well locations were proposed in the favorable oil block, drilled, and recovered higher output of oil and gas.
基金Supported by the National Natural Science Foundation of China(21576240)Experimental Technology Research Program of China University of Geosciences(Key Program)(SJ-201422)
文摘Aiming to increase the efficiency of gem design and manufacturing, a new method in computer-aided-design (CAD) of convex faceted gem cuts (CFGC) based on Half-edge data structure (HDS), including the algorithms for the implementation is presented in this work. By using object-oriented methods, geometrical elements of CFGC are classified and responding geometrical feature classes are established. Each class is implemented and embedded based on the gem process. Matrix arithmetic and analytical geometry are used to derive the affine transformation and the cutting algorithm. Based on the demand for a diversity of gem cuts, CAD functions both for free-style faceted cuts and parametric designs of typical cuts and visualization and human-computer interactions of the CAD system including two-dimensional and three-dimensional interactions have been realized which enhances the flexibility and universality of the CAD system. Furthermore, data in this CAD system can also be used directly by the gem CAM module, which will promote the gem CAD/CAM integration.
基金supported by the Key Program of the National Natural Science Foundation of China(Grant No.50539010)the Special Fund for Public Welfare Industry of the Ministry of Water Resources of China(Grant No.200801019)
文摘In conjunction with association rules for data mining, the connections between testing indices and strong and weak association rules were determined, and new derivative rules were obtained by further reasoning. Association rules were used to analyze correlation and check consistency between indices. This study shows that the judgment obtained by weak association rules or non-association rules is more accurate and more credible than that obtained by strong association rules. When the testing grades of two indices in the weak association rules are inconsistent, the testing grades of indices are more likely to be erroneous, and the mistakes are often caused by human factors. Clustering data mining technology was used to analyze the reliability of a diagnosis, or to perform health diagnosis directly. Analysis showed that the clustering results are related to the indices selected, and that if the indices selected are more significant, the characteristics of clustering results are also more significant, and the analysis or diagnosis is more credible. The indices and diagnosis analysis function produced by this study provide a necessary theoretical foundation and new ideas for the development of hydraulic metal structure health diagnosis technology.
基金supported by CNOOC Research Institute,the Major National R&D project(Grant No.2008 ZX05025)the National R&D project(Grant No.2017 yfc0602202)。
文摘The distribution of oil and gas resources in the South China Sea and adjacent areas is closely related to the structural pattern that helped to define the controlling effect of deep processes on oil-bearing basins.Igneous rocks can record important information from deep processes.Deep structures such as faults,basin uplift and depression,Cenozoic basement and magnetic basement are all the results of energy exchange within the earth.The study of the relationship between igneous rocks and deep structures is of great significance for the study of the South China Sea.By using the minimum curvature potential field separation technique and the correlation analysis technique of gravitational and magnetic anomalies,the fusion of gravitational and magnetic data reflecting igneous rocks can be obtained,through which the igneous rocks with high susceptibility/high density or high susceptibility/low density can be identified.In this study area,igneous rocks do not develop in the Yinggehai basin,Qiongdongnan basin,Zengmu basin and Brunei-Sabah basin whilst igneous rocks with high susceptibility/high density or high susceptibility/low density are widely-developed in other basins.In undeveloped igneous areas,faults are also undeveloped the Cenozoic thickness is greater,the magnetic basement depth is greater and the Cenozoic thickness is highly positively correlated with the magnetic basement depth.In igneously developed regions,the distribution pattern of the Qiongtai block is mainly controlled by primary faults,while the distribution of the Zhongxisha block,Xunta block and Yongshu-Taiping block is mainly controlled by secondary faults,the Cenozoic thickness having a low correlation with the depth of the magnetic basement.
基金The Hong Kong Polytechnic University under internal Grant No. G-YF51.
文摘In a sensor network with a large number of densely populated sensor nodes, a single target of interest may be detected by multiple sensor nodes simultaneously. Data collected from the sensor nodes are usually highly correlated, and hence energy saving using in-network data fusion becomes possible. A traditional data fusion scheme starts with dividing the network into clusters, followed by electing a sensor node as cluster head in each cluster. A cluster head is responsible for collecting data from all its cluster members, performing data fusion on these data and transmitting the fused data to the base station. Assuming that a sensor node is only capable of handling a single node-to-node transmission at a time and each transmission takes T time-slots, a cluster head with n cluster members will take at least nT time-slots to collect data from all its cluster members. In this paper, a tree-based network structure and its formation algorithms are proposed. Simulation results show that the proposed network structure can greatly reduce the delay in data collection.
基金National Key Research and Development Program(No.2018YFB1305001)Major Consulting and Research Project of Chinese Academy of Engineering(No.2018-ZD-02-07)。
文摘Taking autonomous driving and driverless as the research object,we discuss and define intelligent high-precision map.Intelligent high-precision map is considered as a key link of future travel,a carrier of real-time perception of traffic resources in the entire space-time range,and the criterion for the operation and control of the whole process of the vehicle.As a new form of map,it has distinctive features in terms of cartography theory and application requirements compared with traditional navigation electronic maps.Thus,it is necessary to analyze and discuss its key features and problems to promote the development of research and application of intelligent high-precision map.Accordingly,we propose an information transmission model based on the cartography theory and combine the wheeled robot’s control flow in practical application.Next,we put forward the data logic structure of intelligent high-precision map,and analyze its application in autonomous driving.Then,we summarize the computing mode of“Crowdsourcing+Edge-Cloud Collaborative Computing”,and carry out key technical analysis on how to improve the quality of crowdsourced data.We also analyze the effective application scenarios of intelligent high-precision map in the future.Finally,we present some thoughts and suggestions for the future development of this field.
基金Project supported by the National Outstanding Youth Researchers Foundation (No.49525101)the Opening Research Foundation from LIESMARS(WKL(96)0302)
文摘The key to develop 3-D GISs is the study on 3-D data model and data structure. Some of the data models and data structures have been presented by scholars. Because of the complexity of 3-D spatial phenomenon, there are no perfect data structures that can describe all spatial entities. Every data structure has its own advantages and disadvantages. It is difficult to design a single data structure to meet different needs. The important subject in the3-D data models is developing a data model that has integrated vector and raster data structures. A special 3-D spatial data model based on distributing features of spatial entities should be designed. We took the geological exploration engineering as the research background and designed an integrated data model whose data structures integrats vector and raster data byadopting object-oriented technique. Research achievements are presented in this paper.