Expenditure on wells constitute a significant part of the operational costs for a petroleum enterprise, where most of the cost results from drilling. This has prompted drilling departments to continuously look for wa...Expenditure on wells constitute a significant part of the operational costs for a petroleum enterprise, where most of the cost results from drilling. This has prompted drilling departments to continuously look for ways to reduce their drilling costs and be as efficient as possible. A system called the Drilling Comprehensive Information Management and Application System (DCIMAS) is developed and presented here, with an aim at collecting, storing and making full use of the valuable well data and information relating to all drilling activities and operations. The DCIMAS comprises three main parts, including a data collection and transmission system, a data warehouse (DW) management system, and an integrated platform of core applications. With the support of the application platform, the DW management system is introduced, whereby the operation data are captured at well sites and transmitted electronically to a data warehouse via transmission equipment and ETL (extract, transformation and load) tools. With the high quality of the data guaranteed, our central task is to make the best use of the operation data and information for drilling analysis and to provide further information to guide later production stages. Applications have been developed and integrated on a uniform platform to interface directly with different layers of the multi-tier DW. Now, engineers in every department spend less time on data handling and more time on applying technology in their real work with the system.展开更多
Based on the experience and achievement of the"China Digital Ocean", the classification plan for Marine data elements is made, which can be classified into five, including marine point elements, marine line elements...Based on the experience and achievement of the"China Digital Ocean", the classification plan for Marine data elements is made, which can be classified into five, including marine point elements, marine line elements, marine polygon elements, marine grid elements and marine dynamic elements. In this paper, the technology of features and object-oriented method, a spatial-temporal data model is proposed, which can be applied in the large information system engineering like the "Digital Ocean", and this paper discusses the application of spatial data model, marine three-dimensional raster data model and relation data model in the building of Data Warehouse in "China Digital Ocean", and concludes the merits of these models.展开更多
The characters of marine data, such as multi-source, polymorphism, diversity and large amount, determine their differences from other data. How to store and manage marine data rationally and effectively to provide pow...The characters of marine data, such as multi-source, polymorphism, diversity and large amount, determine their differences from other data. How to store and manage marine data rationally and effectively to provide powerful data support for marine management information system and "Digital Ocean" prototype system construction is an urgent problem to solve. Different types of system planning data, such as marine resource, marine environment, marine econotny and marine management, and establishing marine data architecture frame with uniform standard are to realize the effective management of all level marine data, such as national marine data, the provincial (municipal) marine data, and meet the need of fundamental information-platform construction.展开更多
Data warehouse provides storage and management for mass data, but data schema evolves with time on. When data schema is changed, added or deleted, the data in data warehouse must comply with the changed data schema, ...Data warehouse provides storage and management for mass data, but data schema evolves with time on. When data schema is changed, added or deleted, the data in data warehouse must comply with the changed data schema, so data warehouse must be re organized or re constructed, but this process is exhausting and wasteful. In order to cope with these problems, this paper develops an approach to model data cube with XML, which emerges as a universal format for data exchange on the Web and which can make data warehouse flexible and scalable. This paper also extends OLAP algebra for XML based data cube, which is called X OLAP.展开更多
A uniform metadata representation is introduced for heterogeneous databases, multi media information and other information sources. Some features about metadata are analyzed. The limitation of existing metadata model...A uniform metadata representation is introduced for heterogeneous databases, multi media information and other information sources. Some features about metadata are analyzed. The limitation of existing metadata model is compared with the new one. The metadata model is described in XML which is fit for metadata denotation and exchange. The well structured data, semi structured data and those exterior file data without structure are described in the metadata model. The model provides feasibility and extensibility for constructing uniform metadata model of data warehouse.展开更多
Some key technologies of actualizing customer relationship management(CRM)systems are researched.The customer analysis mining prototype systems on the basis of on-line analytical mining(OLAM)is designed.After transact...Some key technologies of actualizing customer relationship management(CRM)systems are researched.The customer analysis mining prototype systems on the basis of on-line analytical mining(OLAM)is designed.After transaction analysis,the data warehouse of CRM is constructed.The CLV/CL/CC customer division model based on customer lifetime value,customer loyalty and customer credit is emphatically researched.Three parameters of customer division—customer lifetime value,customer loyalty and customer credit—are calculated by corresponding algorithms,which can realize customer divisions effectively and improve the accuracy of distinguishing among customers.The data of product sales are analyzed by the sequence association rules algorithm,the potential rules of the products relevance are discovered,which can provide evidence for supporting decisions such as promotion strategies.The transaction data such as product sales volumes and order lists are analyzed on-line through multi-dimensional and multi-level up-drills,down-drills,and horizontal/longitudinal sections.The customer property factors are analyzed as well.The theory and practice of OLAM and its visualization are further explored.展开更多
Current technology for frequent itemset mining mostly applies to the data stored in a single transaction database. This paper presents a novel algorithm MultiClose for frequent itemset mining in data warehouses. Multi...Current technology for frequent itemset mining mostly applies to the data stored in a single transaction database. This paper presents a novel algorithm MultiClose for frequent itemset mining in data warehouses. MultiClose respectively computes the results in single dimension tables and merges the results with a very efficient approach. Close itemsets technique is used to improve the performance of the algorithm. The authors propose an efficient implementation for star schemas in which their al- gorithm outperforms state-of-the-art single-table algorithms.展开更多
This paper first briefly looks back on the history of census GIS around the world,and then summarizes some preponderances of developing urban census GIS in China.And then a three-tier construction architecture for the...This paper first briefly looks back on the history of census GIS around the world,and then summarizes some preponderances of developing urban census GIS in China.And then a three-tier construction architecture for the urban census GIS is proposed.Finally,using the illustration of the census data of Guangzhou city,the paper analyzes and elaborates some issues on the urban census GIS designing,such as data management,data warehouse building,and data analysis.展开更多
Integrating heterogeneous data sources is a precondition to share data for enterprises. Highly-efficient data updating can both save system expenses, and offer real-time data. It is one of the hot issues to modify dat...Integrating heterogeneous data sources is a precondition to share data for enterprises. Highly-efficient data updating can both save system expenses, and offer real-time data. It is one of the hot issues to modify data rapidly in the pre-processing area of the data warehouse. An extract transform loading design is proposed based on a new data algorithm called Diff-Match,which is developed by utilizing mode matching and data-filtering technology. It can accelerate data renewal, filter the heterogeneous data, and seek out different sets of data. Its efficiency has been proved by its successful application in an enterprise of electric apparatus groups.展开更多
The analysis of relevant standards and guidelines proved the lack of information on actions and activities concerning data warehouse testing. The absence of the complex data warehouse testing methodology seems to be c...The analysis of relevant standards and guidelines proved the lack of information on actions and activities concerning data warehouse testing. The absence of the complex data warehouse testing methodology seems to be crucial particularly in the phase of the data warehouse implementation. The aim of this article is to suggest basic data warehouse testing activities as a final part of data warehouse testing methodology. The testing activities that must be implemented in the process of the data warehouse testing can be split into four logical units regarding the multidimensional database testing, data pump testing, metadata and OLAP (Online Analytical Processing) testing. Between main testing activities can be included: revision of the multidimensional database scheme, optimizing of fact tables number, problem of data explosion, testing for correctness of aggregation and summation of data etc.展开更多
Discussing the matter of organizational data management implies, almost automatically, the concept of data warehousing as one of the most important parts of decision support system (DSS), as it supports the integrat...Discussing the matter of organizational data management implies, almost automatically, the concept of data warehousing as one of the most important parts of decision support system (DSS), as it supports the integration of information management by aggregating all data formats and provisioning external systems with consistent data content and flows, together with the metadata concept, as one of the easiest ways of integration for software and database systems. Since organizational data management uses the metadata channel for creating a bi-directional flow, when correctly managed, metadata can save both time and resources for organizations. This paperI will focus on providing theoretical aspects of the two concepts, together with a short brief over a proposed model of design for an organizational management tool.展开更多
This research aims to validate a structural equation modeling (SEM) model for measuring warehouse performance using data from an international company in Australia (company G). Moreover, a methodological triangula...This research aims to validate a structural equation modeling (SEM) model for measuring warehouse performance using data from an international company in Australia (company G). Moreover, a methodological triangulation method was also adopted to test whether different methodological approaches produce convergent findings about warehouse performance measurement (WPM). These three different methods are the SEM model, the multiple case research study, and validation of the SEM model using data from 80 companies in Thailand and a company in Australia. With the results from the triangulation method, it is obvious that the SEM model can be used to measure the performance of warehouses in Thailand and Australia. Since the SEM model consists only of significant indicators, it is more appropriate than company G's scorecard. Furthermore, the SEM model can overcome the limitations of traditional models by allowing companies to compare their performance over time.展开更多
Data structure and semantics of the traditional data model cannot effectively represent the data warehouse, it is difficult to effectively support online analytical processing (referred to as OLAP). This paper is pr...Data structure and semantics of the traditional data model cannot effectively represent the data warehouse, it is difficult to effectively support online analytical processing (referred to as OLAP). This paper is propose a new multidimensional data model based on the partial ordering and mapping. The data model can fully express the complex data structure and semantics of data warehouse, and provide an OLAP operation as the core of the operation of algebra, support structure in levels of complex aggregation operation sequence, which can effectively support the application of OLAE The data model supports the concept of aggregation function constraint, and provides constraint mechanism of the hierarchy aggregation function.展开更多
Objective To establish a warehouse on acupuncture-moxibution (acup-mox) methods to explore valuable laws about research and clinical application of acup-mox in a great number of literature by use of data mining tech...Objective To establish a warehouse on acupuncture-moxibution (acup-mox) methods to explore valuable laws about research and clinical application of acup-mox in a great number of literature by use of data mining technique and to promote acup-mox research and effective treatment of diseases. Methods According to the acup-mox literature information of different types, different subjects of the aeup-mox literature are determined and the relevant database is established. In the continuously enriched subject database, the data warehouse catering to multi-subjects and multi-dimensions is set up so as to provide a platform for wider application of aeup-mox literature information. Results Based on characteristics of the acup-mox literature, many subject databases, such as needling with filiform needle, moxibustion, etc., are established and clinical treatment laws of acup-mox are revealed by use of data mining method in the database established. Conclusion Establishment of the acup-mox literature warehouse provides a standard data expression model, rich attributes and relation between different literature information for study of aeup-mox literature by more effective techniques, and a rich and standard data basis for acup-mox researches.展开更多
A regional groundwater management system has been elaborated, integrating Relational Database Management System (RDBMS) and various web services. It consists of web geospatial application so-called HydrlS ( Hydroge...A regional groundwater management system has been elaborated, integrating Relational Database Management System (RDBMS) and various web services. It consists of web geospatial application so-called HydrlS ( Hydrogeological Information System) based on Open Source components and technologies, leading to a feasible and low-cost solution. Therefore, HydrlS permits delivery of data from a number of heterogeneous sources to standards supported by the Open Geospatial Consortium (OGC). The protocols used for exchanging data are also derived from OGC standards, i.e., WMS (Web Mapping Service), WFS (Web Feature Service), and WCS (Web Coverage Service). Finally, a geoportal was developed, which consists of client-applications that communicate with different Web Services (WMS, WCS, and WFS) through HTTP-requests. A prototype for web-based GIS application was designed using the deegree Framework to provide systematic interfaces and functions. This system was developed to demonstrate the value of making hydrogeological data more widely accessible through client/server architecture. This experience and knowledge already gained in this project will be a source for technology transfer and policy decisions. Otherwise, this will enable user groups to improve the management of their groundwater resources and contribute to enhanced decision support capabilities.展开更多
The issues and challenges involved in controlling the collaborative changes in a Building Information Modeling(BIM) data repository, in a multi-model collaboration environment, are discussed. It is suggested that mana...The issues and challenges involved in controlling the collaborative changes in a Building Information Modeling(BIM) data repository, in a multi-model collaboration environment, are discussed. It is suggested that managing iterative changes in BIMs is a database problem, exacerbated by the long transaction times needed to support collaborative design progression. This is yet to be resolved in the construction industry and better solutions are needed to support the underlying workflows and computing operations for seamless collaboration on BIMs. With this in mind, this paper proposes the use of the structural and semantic characteristics of BIM objects as a mechanism for tracking changes across co-developed solutions. The creation of object signatures, using hash codes derived from their characteristics, provides a potential mechanism for object comparison and effective change recognition and management.展开更多
文摘Expenditure on wells constitute a significant part of the operational costs for a petroleum enterprise, where most of the cost results from drilling. This has prompted drilling departments to continuously look for ways to reduce their drilling costs and be as efficient as possible. A system called the Drilling Comprehensive Information Management and Application System (DCIMAS) is developed and presented here, with an aim at collecting, storing and making full use of the valuable well data and information relating to all drilling activities and operations. The DCIMAS comprises three main parts, including a data collection and transmission system, a data warehouse (DW) management system, and an integrated platform of core applications. With the support of the application platform, the DW management system is introduced, whereby the operation data are captured at well sites and transmitted electronically to a data warehouse via transmission equipment and ETL (extract, transformation and load) tools. With the high quality of the data guaranteed, our central task is to make the best use of the operation data and information for drilling analysis and to provide further information to guide later production stages. Applications have been developed and integrated on a uniform platform to interface directly with different layers of the multi-tier DW. Now, engineers in every department spend less time on data handling and more time on applying technology in their real work with the system.
文摘Based on the experience and achievement of the"China Digital Ocean", the classification plan for Marine data elements is made, which can be classified into five, including marine point elements, marine line elements, marine polygon elements, marine grid elements and marine dynamic elements. In this paper, the technology of features and object-oriented method, a spatial-temporal data model is proposed, which can be applied in the large information system engineering like the "Digital Ocean", and this paper discusses the application of spatial data model, marine three-dimensional raster data model and relation data model in the building of Data Warehouse in "China Digital Ocean", and concludes the merits of these models.
文摘The characters of marine data, such as multi-source, polymorphism, diversity and large amount, determine their differences from other data. How to store and manage marine data rationally and effectively to provide powerful data support for marine management information system and "Digital Ocean" prototype system construction is an urgent problem to solve. Different types of system planning data, such as marine resource, marine environment, marine econotny and marine management, and establishing marine data architecture frame with uniform standard are to realize the effective management of all level marine data, such as national marine data, the provincial (municipal) marine data, and meet the need of fundamental information-platform construction.
文摘Data warehouse provides storage and management for mass data, but data schema evolves with time on. When data schema is changed, added or deleted, the data in data warehouse must comply with the changed data schema, so data warehouse must be re organized or re constructed, but this process is exhausting and wasteful. In order to cope with these problems, this paper develops an approach to model data cube with XML, which emerges as a universal format for data exchange on the Web and which can make data warehouse flexible and scalable. This paper also extends OLAP algebra for XML based data cube, which is called X OLAP.
文摘A uniform metadata representation is introduced for heterogeneous databases, multi media information and other information sources. Some features about metadata are analyzed. The limitation of existing metadata model is compared with the new one. The metadata model is described in XML which is fit for metadata denotation and exchange. The well structured data, semi structured data and those exterior file data without structure are described in the metadata model. The model provides feasibility and extensibility for constructing uniform metadata model of data warehouse.
基金The National Key Technology R&D Program of China during the 11th Five-Year Plan Period(No.2006BAH02A06)
文摘Some key technologies of actualizing customer relationship management(CRM)systems are researched.The customer analysis mining prototype systems on the basis of on-line analytical mining(OLAM)is designed.After transaction analysis,the data warehouse of CRM is constructed.The CLV/CL/CC customer division model based on customer lifetime value,customer loyalty and customer credit is emphatically researched.Three parameters of customer division—customer lifetime value,customer loyalty and customer credit—are calculated by corresponding algorithms,which can realize customer divisions effectively and improve the accuracy of distinguishing among customers.The data of product sales are analyzed by the sequence association rules algorithm,the potential rules of the products relevance are discovered,which can provide evidence for supporting decisions such as promotion strategies.The transaction data such as product sales volumes and order lists are analyzed on-line through multi-dimensional and multi-level up-drills,down-drills,and horizontal/longitudinal sections.The customer property factors are analyzed as well.The theory and practice of OLAM and its visualization are further explored.
文摘Current technology for frequent itemset mining mostly applies to the data stored in a single transaction database. This paper presents a novel algorithm MultiClose for frequent itemset mining in data warehouses. MultiClose respectively computes the results in single dimension tables and merges the results with a very efficient approach. Close itemsets technique is used to improve the performance of the algorithm. The authors propose an efficient implementation for star schemas in which their al- gorithm outperforms state-of-the-art single-table algorithms.
文摘This paper first briefly looks back on the history of census GIS around the world,and then summarizes some preponderances of developing urban census GIS in China.And then a three-tier construction architecture for the urban census GIS is proposed.Finally,using the illustration of the census data of Guangzhou city,the paper analyzes and elaborates some issues on the urban census GIS designing,such as data management,data warehouse building,and data analysis.
基金Supported by National Natural Science Foundation of China (No. 50475117)Tianjin Natural Science Foundation (No.06YFJMJC03700).
文摘Integrating heterogeneous data sources is a precondition to share data for enterprises. Highly-efficient data updating can both save system expenses, and offer real-time data. It is one of the hot issues to modify data rapidly in the pre-processing area of the data warehouse. An extract transform loading design is proposed based on a new data algorithm called Diff-Match,which is developed by utilizing mode matching and data-filtering technology. It can accelerate data renewal, filter the heterogeneous data, and seek out different sets of data. Its efficiency has been proved by its successful application in an enterprise of electric apparatus groups.
文摘The analysis of relevant standards and guidelines proved the lack of information on actions and activities concerning data warehouse testing. The absence of the complex data warehouse testing methodology seems to be crucial particularly in the phase of the data warehouse implementation. The aim of this article is to suggest basic data warehouse testing activities as a final part of data warehouse testing methodology. The testing activities that must be implemented in the process of the data warehouse testing can be split into four logical units regarding the multidimensional database testing, data pump testing, metadata and OLAP (Online Analytical Processing) testing. Between main testing activities can be included: revision of the multidimensional database scheme, optimizing of fact tables number, problem of data explosion, testing for correctness of aggregation and summation of data etc.
文摘Discussing the matter of organizational data management implies, almost automatically, the concept of data warehousing as one of the most important parts of decision support system (DSS), as it supports the integration of information management by aggregating all data formats and provisioning external systems with consistent data content and flows, together with the metadata concept, as one of the easiest ways of integration for software and database systems. Since organizational data management uses the metadata channel for creating a bi-directional flow, when correctly managed, metadata can save both time and resources for organizations. This paperI will focus on providing theoretical aspects of the two concepts, together with a short brief over a proposed model of design for an organizational management tool.
文摘This research aims to validate a structural equation modeling (SEM) model for measuring warehouse performance using data from an international company in Australia (company G). Moreover, a methodological triangulation method was also adopted to test whether different methodological approaches produce convergent findings about warehouse performance measurement (WPM). These three different methods are the SEM model, the multiple case research study, and validation of the SEM model using data from 80 companies in Thailand and a company in Australia. With the results from the triangulation method, it is obvious that the SEM model can be used to measure the performance of warehouses in Thailand and Australia. Since the SEM model consists only of significant indicators, it is more appropriate than company G's scorecard. Furthermore, the SEM model can overcome the limitations of traditional models by allowing companies to compare their performance over time.
文摘Data structure and semantics of the traditional data model cannot effectively represent the data warehouse, it is difficult to effectively support online analytical processing (referred to as OLAP). This paper is propose a new multidimensional data model based on the partial ordering and mapping. The data model can fully express the complex data structure and semantics of data warehouse, and provide an OLAP operation as the core of the operation of algebra, support structure in levels of complex aggregation operation sequence, which can effectively support the application of OLAE The data model supports the concept of aggregation function constraint, and provides constraint mechanism of the hierarchy aggregation function.
基金Supported by National Natural Science Foundation of China: No.81072883
文摘Objective To establish a warehouse on acupuncture-moxibution (acup-mox) methods to explore valuable laws about research and clinical application of acup-mox in a great number of literature by use of data mining technique and to promote acup-mox research and effective treatment of diseases. Methods According to the acup-mox literature information of different types, different subjects of the aeup-mox literature are determined and the relevant database is established. In the continuously enriched subject database, the data warehouse catering to multi-subjects and multi-dimensions is set up so as to provide a platform for wider application of aeup-mox literature information. Results Based on characteristics of the acup-mox literature, many subject databases, such as needling with filiform needle, moxibustion, etc., are established and clinical treatment laws of acup-mox are revealed by use of data mining method in the database established. Conclusion Establishment of the acup-mox literature warehouse provides a standard data expression model, rich attributes and relation between different literature information for study of aeup-mox literature by more effective techniques, and a rich and standard data basis for acup-mox researches.
文摘A regional groundwater management system has been elaborated, integrating Relational Database Management System (RDBMS) and various web services. It consists of web geospatial application so-called HydrlS ( Hydrogeological Information System) based on Open Source components and technologies, leading to a feasible and low-cost solution. Therefore, HydrlS permits delivery of data from a number of heterogeneous sources to standards supported by the Open Geospatial Consortium (OGC). The protocols used for exchanging data are also derived from OGC standards, i.e., WMS (Web Mapping Service), WFS (Web Feature Service), and WCS (Web Coverage Service). Finally, a geoportal was developed, which consists of client-applications that communicate with different Web Services (WMS, WCS, and WFS) through HTTP-requests. A prototype for web-based GIS application was designed using the deegree Framework to provide systematic interfaces and functions. This system was developed to demonstrate the value of making hydrogeological data more widely accessible through client/server architecture. This experience and knowledge already gained in this project will be a source for technology transfer and policy decisions. Otherwise, this will enable user groups to improve the management of their groundwater resources and contribute to enhanced decision support capabilities.
文摘The issues and challenges involved in controlling the collaborative changes in a Building Information Modeling(BIM) data repository, in a multi-model collaboration environment, are discussed. It is suggested that managing iterative changes in BIMs is a database problem, exacerbated by the long transaction times needed to support collaborative design progression. This is yet to be resolved in the construction industry and better solutions are needed to support the underlying workflows and computing operations for seamless collaboration on BIMs. With this in mind, this paper proposes the use of the structural and semantic characteristics of BIM objects as a mechanism for tracking changes across co-developed solutions. The creation of object signatures, using hash codes derived from their characteristics, provides a potential mechanism for object comparison and effective change recognition and management.