With the explosive growth of data available, there is an urgent need to develop continuous data mining which reduces manual interaction evidently. A novel model for data mining is proposed in evolving environment. Fir...With the explosive growth of data available, there is an urgent need to develop continuous data mining which reduces manual interaction evidently. A novel model for data mining is proposed in evolving environment. First, some valid mining task schedules are generated, and then au tonomous and local mining are executed periodically, finally, previous results are merged and refined. The framework based on the model creates a communication mechanism to in corporate domain knowledge into continuous process through ontology service. The local and merge mining are transparent to the end user and heterogeneous data ,source by ontology. Experiments suggest that the framework should be useful in guiding the continuous mining process.展开更多
Reliability and remaining useful life(RUL)estimation for a satellite rechargeable lithium battery(RLB)are significant for prognostic and health management(PHM).A novel Bayesian framework is proposed to do reliability ...Reliability and remaining useful life(RUL)estimation for a satellite rechargeable lithium battery(RLB)are significant for prognostic and health management(PHM).A novel Bayesian framework is proposed to do reliability analysis by synthesizing multisource data,including bivariate degradation data and lifetime data.Bivariate degradation means that there are two degraded performance characteristics leading to the failure of the system.First,linear Wiener process and Frank Copula function are used to model the dependent degradation processes of the RLB's temperature and discharge voltage.Next,the Bayesian method,in combination with Markov Chain Monte Carlo(MCMC)simulations,is provided to integrate limited bivariate degradation data with other congeneric RLBs'lifetime data.Then reliability evaluation and RUL prediction are carried out for PHM.A simulation study demonstrates that due to the data fusion,parameter estimations and predicted RUL obtained from our model are more precise than models only using degradation data or ignoring the dependency of different degradation processes.Finally,a practical case study of a satellite RLB verifies the usability of the model.展开更多
This paper investigates the problem of data scarcity in spectrum prediction.A cognitive radio equipment may frequently switch the target frequency as the electromagnetic environment changes.The previously trained mode...This paper investigates the problem of data scarcity in spectrum prediction.A cognitive radio equipment may frequently switch the target frequency as the electromagnetic environment changes.The previously trained model for prediction often cannot maintain a good performance when facing small amount of historical data of the new target frequency.Moreover,the cognitive radio equipment usually implements the dynamic spectrum access in real time which means the time to recollect the data of the new task frequency band and retrain the model is very limited.To address the above issues,we develop a crossband data augmentation framework for spectrum prediction by leveraging the recent advances of generative adversarial network(GAN)and deep transfer learning.Firstly,through the similarity measurement,we pre-train a GAN model using the historical data of the frequency band that is the most similar to the target frequency band.Then,through the data augmentation by feeding the small amount of the target data into the pre-trained GAN,temporal-spectral residual network is further trained using deep transfer learning and the generated data with high similarity from GAN.Finally,experiment results demonstrate the effectiveness of the proposed framework.展开更多
The tremendous growth of the cloud computing environments requires new architecture for security services. Cloud computing is the utilization of many servers/data centers or cloud data storages (CDSs) housed in many d...The tremendous growth of the cloud computing environments requires new architecture for security services. Cloud computing is the utilization of many servers/data centers or cloud data storages (CDSs) housed in many different locations and interconnected by high speed networks. CDS, like any other emerging technology, is experiencing growing pains. It is immature, it is fragmented and it lacks standardization. Although security issues are delaying its fast adoption, cloud computing is an unstoppable force and we need to provide security mechanisms to ensure its secure adoption. In this paper a comprehensive security framework based on Multi-Agent System (MAS) architecture for CDS to facilitate confidentiality, correctness assurance, availability and integrity of users' data in the cloud is proposed. Our security framework consists of two main layers as agent layer and CDS layer. Our propose MAS architecture includes main five types of agents: Cloud Service Provider Agent (CSPA), Cloud Data Confidentiality Agent (CDConA), Cloud Data Correctness Agent (CDCorA), Cloud Data Availability Agent (CDAA) and Cloud Data Integrity Agent (CDIA). In order to verify our proposed security framework based on MAS architecture, pilot study is conducted using a questionnaire survey. Rasch Methodology is used to analyze the pilot data. Item reliability is found to be poor and a few respondents and items are identified as misfits with distorted measurements. As a result, some problematic questions are revised and some predictably easy questions are excluded from the questionnaire. A prototype of the system is implemented using Java. To simulate the agents, oracle database packages and triggers are used to implement agent functions and oracle jobs are utilized to create agents.展开更多
Cloud accounting is based on the traditional financial work process, the context of big data, and the necessary trend of future corporate accounting development. Its emergence and rapid development will have a fundame...Cloud accounting is based on the traditional financial work process, the context of big data, and the necessary trend of future corporate accounting development. Its emergence and rapid development will have a fundamental impact on corporate environmental information disclosure. In the big data era of information sharing, companies will have a new understanding of the emergence, balance, and final consideration of social responsibility, and will have new changes in their overall decision-making and information disclosure methods. "Knowing" and "behavior" will be combined on the basis of rational judgment, so that corporate environmental information disclosure is more in line with the overall social development requirements. Based on the background of big data, this article starts with the disclosure of impact factors, footholds, and path choices. It describes the evolution of corporate environmental information disclosure and provides reference suggestions for enterprises to disclose environmental information truthfully and perform social responsibilities.展开更多
In several fields like financial dealing,industry,business,medicine,et cetera,Big Data(BD)has been utilized extensively,which is nothing but a collection of a huge amount of data.However,it is highly complicated along...In several fields like financial dealing,industry,business,medicine,et cetera,Big Data(BD)has been utilized extensively,which is nothing but a collection of a huge amount of data.However,it is highly complicated along with time-consuming to process a massive amount of data.Thus,to design the Distribution Preserving Framework for BD,a novel methodology has been proposed utilizing Manhattan Distance(MD)-centered Partition Around Medoid(MD–PAM)along with Conjugate Gradient Artificial Neural Network(CG-ANN),which undergoes various steps to reduce the complications of BD.Firstly,the data are processed in the pre-processing phase by mitigating the data repetition utilizing the map-reduce function;subsequently,the missing data are handled by substituting or by ignoring the missed values.After that,the data are transmuted into a normalized form.Next,to enhance the classification performance,the data’s dimensionalities are minimized by employing Gaussian Kernel(GK)-Fisher Discriminant Analysis(GK-FDA).Afterwards,the processed data is submitted to the partitioning phase after transmuting it into a structured format.In the partition phase,by utilizing the MD-PAM,the data are partitioned along with grouped into a cluster.Lastly,by employing CG-ANN,the data are classified in the classification phase so that the needed data can be effortlessly retrieved by the user.To analogize the outcomes of the CG-ANN with the prevailing methodologies,the NSL-KDD openly accessible datasets are utilized.The experiential outcomes displayed that an efficient result along with a reduced computation cost was shown by the proposed CG-ANN.The proposed work outperforms well in terms of accuracy,sensitivity and specificity than the existing systems.展开更多
An idea is presented about the development of a data processing and analysis system for ICF experiments, which is based on an object oriented framework. The design and preliminary implementation of the data processing...An idea is presented about the development of a data processing and analysis system for ICF experiments, which is based on an object oriented framework. The design and preliminary implementation of the data processing and analysis framework based on the ROOT system have been completed. Software for unfolding soft X-ray spectra has been developed to test the functions of this framework.展开更多
Antarctic data management is the research focus, which the international Antarctic organizations, e.g. Antarctic Treaty Consultative Meeting(ATCM) , Scientific Committee on Antarctic Research(SCAR), and Council of Man...Antarctic data management is the research focus, which the international Antarctic organizations, e.g. Antarctic Treaty Consultative Meeting(ATCM) , Scientific Committee on Antarctic Research(SCAR), and Council of Managers of National Antarctic Programmes(COMNAP) have been paying close attention to and promoting actively. Through the co effort of international Antarctic organizations and member countries concerned in recent years, Antarctic Data Directory Syatem(ADDS) is established as the most important basic programme for development of the international Antarctic data management system. At present, Joint Committee on Antarctic Data Management(JCADM) is responsible for organizing and coordinating the international Antarctic data management, and implementing the project ADDS.In this paper the background on Antarctic data management in time sequence and the structure of international framework are introduced, meanwhile, it is necessary to develop ADDS first of all. The ADDS mainly consists of the two principal parts: National Antarctic Data Center(NADCs) of all the party members and Antarctic Main Directory(AMD), the best available technology for creating ADDS is to make full use of International Directory Network(IDN) and adopt its Directory Interchange Formats(DIF). In the light of the above requirements, combined with Chinese specific situation, the contents, technical and administrative methods on Chinese Antarctic data management are discussed to promote our related work.展开更多
With an increase in population and economic development,water withdrawals are close to or even exceed the amount of water available in many regions of the world.Modelling water withdrawals could help water planners im...With an increase in population and economic development,water withdrawals are close to or even exceed the amount of water available in many regions of the world.Modelling water withdrawals could help water planners improve the efficiency of water use,water resources allocation,and management in order to alleviate water crises.However,minimal information has been obtained on how water withdrawals have changed over space and time,especially on a regional or local scale.This research proposes a data-driven framework to help estimate county-level distribution of water withdrawals.Using this framework,spatial statistical methods are used to estimate water withdrawals for agricultural,industrial,and domestic purposes in the Huaihe River watershed in China for the period 1978–2018.Total water withdrawals were found to have more than doubled,from 292.55×10^(8)m^(3) in 1978 to 642.93×10^(8)m^(3) in 2009,and decreased to 602.63×10^(8)m^(3) in 2018.Agricultural water increased from 208.17×10^(8)m^(3) in 1978 to 435.80×10^(8)m^(3) in 2009 and decreased to 360.84×10^(8)m^(3) in 2018.Industrial and domestic water usage constantly increased throughout the 1978–2018 period.In 1978,industrial and domestic demands were 20.35×10^(8)m^(3) and 60.04×10^(8)m^(3),respectively,and up until 2018,the figures were 105.58×10^(8)m^(3) and 136.20×10^(8)m^(3).From a spatial distribution perspective,Moran’s I statistical results show that the total water withdrawal has significant spatial autocorrelation during 1978–2018.The overall trend was a gradual increase in 1978–2010 with withdrawal beginning to decline in 2010–2018.The results of Getis-Ord G_(i)^(*)statistical calculations showed spatially contiguous clusters of total water withdrawal in the Huaihe River watershed during1978–2010,and the spatial agglomeration weakened from 2010 to 2018.This study provides a data-driven framework for assessing water withdrawals to enable a deeper understanding of competing water use among economic sectors as well as water withdrawal modelled with proper data resource and method.展开更多
This paper proposes a two-step general framework for reversible data hiding(RDH)schemes with controllable contrast enhancement.The first step aims at preserving visual perception as much as possible on the basis of ac...This paper proposes a two-step general framework for reversible data hiding(RDH)schemes with controllable contrast enhancement.The first step aims at preserving visual perception as much as possible on the basis of achieving high embedding capacity(EC),while the second step is used for increasing image contrast.In the second step,some peak-pairs are utilized so that the histogram of pixel values is modified to perform histogram equalization(HE),which would lead to the image contrast enhancement.However,for HE,the utilization of some peak-pairs easily leads to over-enhanced image contrast when a large number of bits are embedded.Therefore,in our proposed framework,contrast over-enhancement is avoided by controlling the degree of contrast enhancement.Since the second step can only provide a small amount of data due to controlled contrast enhancement,the first one helps to achieve a large amount of data without degrading visual quality.Any RDH method which can achieve high EC while preserve good visual quality,can be selected for the first step.In fact,Gao et al.’s method is a special case of our proposed framework.In addition,two simple and commonly-used RDH methods are also introduced to further demonstrate the generalization of our framework.展开更多
Point of Care (PoC) devices and systems can be categorized into three broad classes (CAT 1, CAT 2, and CAT 3) based on the context of operation and usage. In this paper, the categories are defined to address certain u...Point of Care (PoC) devices and systems can be categorized into three broad classes (CAT 1, CAT 2, and CAT 3) based on the context of operation and usage. In this paper, the categories are defined to address certain usage models of the PoC device. PoC devices that are used for PoC testing and diagnostic applications are defined CAT 1 devices;PoC devices that are used for patient monitoring are defined as CAT 2 devices (PoCM);PoC devices that are used for as interfacing with other devices are defined as CAT 3 devices (PoCI). The PoCI devices provide an interface gateway for collecting and aggregating data from other medical devices. In all categories, data security is an important aspect. This paper presents a security framework concept, which is applicable for all of the classes of PoC operation. It outlines the concepts and security framework for preventing security challenges in unauthorized access to data, unintended data flow, and data tampering during communication between system entities, the user, and the PoC system. The security framework includes secure layering of basic PoC system architecture, protection of PoC devices in the context of application and network. Developing the security framework is taken into account of a thread model of the PoC system. A proposal for a low-level protocol is discussed. This protocol is independent of communications technologies, and it is elaborated in relation to providing security. An algorithm that can be used to overcome the threat challenges has been shown using the elements in the protocol. The paper further discusses the vulnerability scanning process for the PoC system interconnected network. The paper also presents a four-step process of authentication and authorization framework for providing the security for the PoC system. Finally, the paper concludes with the machine to machine (M2M) security viewpoint and discusses the key stakeholders within an actual deployment of the PoC system and its security challenges.展开更多
The importance of the project selection phase in any six sigma initiative cannot be emphasized enough. The successfulness of the six sigma initiative is affected by successful project selection. Recently, Data Envelop...The importance of the project selection phase in any six sigma initiative cannot be emphasized enough. The successfulness of the six sigma initiative is affected by successful project selection. Recently, Data Envelopment Analysis (DEA) has been proposed as a six sigma project selection tool. However, there exist a number of different DEA formulations which may affect the selection process and the wining project being selected. This work initially applies nine different DEA formulations to several case studies and concludes that different DEA formulations select different wining projects. Also in this work, a Multi-DEA Unified Scoring Framework is proposed to overcome this problem. This framework is applied to several case studies and proved to successfully select the six sigma project with the best performance. The framework is also successful in filtering out some of the projects that have “selective” excellent performance, i.e. projects with excellent performance in some of the DEA formulations and worse performance in others. It is also successful in selecting stable projects;these are projects that perform well in the majority of the DEA formulations, even if it has not been selected as a wining project by any of the DEA formulations.展开更多
Digital educational content is gaining importance as an incubator of pedagogical methodologies in formal and informal online educational settings. Its educational efficiency is directly dependent on its quality, howev...Digital educational content is gaining importance as an incubator of pedagogical methodologies in formal and informal online educational settings. Its educational efficiency is directly dependent on its quality, however educational content is more than information and data. This paper presents a new data quality framework for assessing digital educational content used for teaching in distance learning environments. The model relies on the ISO2500 series quality standard and beside providing the mechanisms for multi-facet quality assessment it also supports organizations that design, create, manage and use educational content with the quality tools (expressed as quality metrics and measurement methods) to provide a more efficient distance education experience. The model describes the quality characteristics of the educational material content using data and software quality characteristics.展开更多
When data privacy is imposed as a necessity,Federated learning(FL)emerges as a relevant artificial intelligence field for developing machine learning(ML)models in a distributed and decentralized environment.FL allows ...When data privacy is imposed as a necessity,Federated learning(FL)emerges as a relevant artificial intelligence field for developing machine learning(ML)models in a distributed and decentralized environment.FL allows ML models to be trained on local devices without any need for centralized data transfer,thereby reducing both the exposure of sensitive data and the possibility of data interception by malicious third parties.This paradigm has gained momentum in the last few years,spurred by the plethora of real-world applications that have leveraged its ability to improve the efficiency of distributed learning and to accommodate numerous participants with their data sources.By virtue of FL,models can be learned from all such distributed data sources while preserving data privacy.The aim of this paper is to provide a practical tutorial on FL,including a short methodology and a systematic analysis of existing software frameworks.Furthermore,our tutorial provides exemplary cases of study from three complementary perspectives:i)Foundations of FL,describing the main components of FL,from key elements to FL categories;ii)Implementation guidelines and exemplary cases of study,by systematically examining the functionalities provided by existing software frameworks for FL deployment,devising a methodology to design a FL scenario,and providing exemplary cases of study with source code for different ML approaches;and iii)Trends,shortly reviewing a non-exhaustive list of research directions that are under active investigation in the current FL landscape.The ultimate purpose of this work is to establish itself as a referential work for researchers,developers,and data scientists willing to explore the capabilities of FL in practical applications.展开更多
Big data analytics has been widely adopted by large companies to achieve measurable benefits including increased profitability,customer demand forecasting,cheaper development of products,and improved stock control.Sma...Big data analytics has been widely adopted by large companies to achieve measurable benefits including increased profitability,customer demand forecasting,cheaper development of products,and improved stock control.Small and medium sized enterprises(SMEs)are the backbone of the global economy,comprising of 90%of businesses worldwide.However,only 10%SMEs have adopted big data analytics despite the competitive advantage they could achieve.Previous research has analysed the barriers to adoption and a strategic framework has been developed to help SMEs adopt big data analytics.The framework was converted into a scoring tool which has been applied to multiple case studies of SMEs in the UK.This paper documents the process of evaluating the framework based on the structured feedback from a focus group composed of experienced practitioners.The results of the evaluation are presented with a discussion on the results,and the paper concludes with recommendations to improve the scoring tool based on the proposed framework.The research demonstrates that this positioning tool is beneficial for SMEs to achieve competitive advantages by increasing the application of business intelligence and big data analytics.展开更多
The Zambian mining industry is crucial to the national economy but struggles with inconsistent equipment maintenance practices. This study developed an Equipment Maintenance Management Framework (EMMF) tailored to the...The Zambian mining industry is crucial to the national economy but struggles with inconsistent equipment maintenance practices. This study developed an Equipment Maintenance Management Framework (EMMF) tailored to the industry’s needs. Using surveys, interviews, and on-site visits at eight major mining companies, we identified significant variations in maintenance strategies, CMMS usage, and reliability engineering. The EMMF prioritizes predictive maintenance, efficient CMMS implementation, ongoing training, and robust reliability engineering to shift from reactive to proactive maintenance. We recommend adopting continuous improvement practices and data-driven decision-making based on performance metrics, with a phased EMMF implementation aligning maintenance with strategic business objectives. This framework is poised to enhance operational efficiency, equipment reliability, and safety, fostering sustainable growth in the Zambian mining sector.展开更多
基金Supported by the National Natural Science Foun-dation of China (60173058 ,70372024)
文摘With the explosive growth of data available, there is an urgent need to develop continuous data mining which reduces manual interaction evidently. A novel model for data mining is proposed in evolving environment. First, some valid mining task schedules are generated, and then au tonomous and local mining are executed periodically, finally, previous results are merged and refined. The framework based on the model creates a communication mechanism to in corporate domain knowledge into continuous process through ontology service. The local and merge mining are transparent to the end user and heterogeneous data ,source by ontology. Experiments suggest that the framework should be useful in guiding the continuous mining process.
基金Project(71371182) supported by the National Natural Science Foundation of China
文摘Reliability and remaining useful life(RUL)estimation for a satellite rechargeable lithium battery(RLB)are significant for prognostic and health management(PHM).A novel Bayesian framework is proposed to do reliability analysis by synthesizing multisource data,including bivariate degradation data and lifetime data.Bivariate degradation means that there are two degraded performance characteristics leading to the failure of the system.First,linear Wiener process and Frank Copula function are used to model the dependent degradation processes of the RLB's temperature and discharge voltage.Next,the Bayesian method,in combination with Markov Chain Monte Carlo(MCMC)simulations,is provided to integrate limited bivariate degradation data with other congeneric RLBs'lifetime data.Then reliability evaluation and RUL prediction are carried out for PHM.A simulation study demonstrates that due to the data fusion,parameter estimations and predicted RUL obtained from our model are more precise than models only using degradation data or ignoring the dependency of different degradation processes.Finally,a practical case study of a satellite RLB verifies the usability of the model.
基金This work was supported by the Science and Technology Innovation 2030-Key Project of“New Generation Artificial Intelligence”of China under Grant 2018AAA0102303the Natural Science Foundation for Distinguished Young Scholars of Jiangsu Province(No.BK20190030)the National Natural Science Foundation of China(No.61631020,No.61871398,No.61931011 and No.U20B2038).
文摘This paper investigates the problem of data scarcity in spectrum prediction.A cognitive radio equipment may frequently switch the target frequency as the electromagnetic environment changes.The previously trained model for prediction often cannot maintain a good performance when facing small amount of historical data of the new target frequency.Moreover,the cognitive radio equipment usually implements the dynamic spectrum access in real time which means the time to recollect the data of the new task frequency band and retrain the model is very limited.To address the above issues,we develop a crossband data augmentation framework for spectrum prediction by leveraging the recent advances of generative adversarial network(GAN)and deep transfer learning.Firstly,through the similarity measurement,we pre-train a GAN model using the historical data of the frequency band that is the most similar to the target frequency band.Then,through the data augmentation by feeding the small amount of the target data into the pre-trained GAN,temporal-spectral residual network is further trained using deep transfer learning and the generated data with high similarity from GAN.Finally,experiment results demonstrate the effectiveness of the proposed framework.
文摘The tremendous growth of the cloud computing environments requires new architecture for security services. Cloud computing is the utilization of many servers/data centers or cloud data storages (CDSs) housed in many different locations and interconnected by high speed networks. CDS, like any other emerging technology, is experiencing growing pains. It is immature, it is fragmented and it lacks standardization. Although security issues are delaying its fast adoption, cloud computing is an unstoppable force and we need to provide security mechanisms to ensure its secure adoption. In this paper a comprehensive security framework based on Multi-Agent System (MAS) architecture for CDS to facilitate confidentiality, correctness assurance, availability and integrity of users' data in the cloud is proposed. Our security framework consists of two main layers as agent layer and CDS layer. Our propose MAS architecture includes main five types of agents: Cloud Service Provider Agent (CSPA), Cloud Data Confidentiality Agent (CDConA), Cloud Data Correctness Agent (CDCorA), Cloud Data Availability Agent (CDAA) and Cloud Data Integrity Agent (CDIA). In order to verify our proposed security framework based on MAS architecture, pilot study is conducted using a questionnaire survey. Rasch Methodology is used to analyze the pilot data. Item reliability is found to be poor and a few respondents and items are identified as misfits with distorted measurements. As a result, some problematic questions are revised and some predictably easy questions are excluded from the questionnaire. A prototype of the system is implemented using Java. To simulate the agents, oracle database packages and triggers are used to implement agent functions and oracle jobs are utilized to create agents.
基金supported by the Philosophy and Social Science Research Project of Daqing City (Grant No. DSGB2017112)the Postgraduate Innovation Research Project of Heilongjiang Bayi Agricultural University (Grant No. YJSCX2017Y79)
文摘Cloud accounting is based on the traditional financial work process, the context of big data, and the necessary trend of future corporate accounting development. Its emergence and rapid development will have a fundamental impact on corporate environmental information disclosure. In the big data era of information sharing, companies will have a new understanding of the emergence, balance, and final consideration of social responsibility, and will have new changes in their overall decision-making and information disclosure methods. "Knowing" and "behavior" will be combined on the basis of rational judgment, so that corporate environmental information disclosure is more in line with the overall social development requirements. Based on the background of big data, this article starts with the disclosure of impact factors, footholds, and path choices. It describes the evolution of corporate environmental information disclosure and provides reference suggestions for enterprises to disclose environmental information truthfully and perform social responsibilities.
文摘In several fields like financial dealing,industry,business,medicine,et cetera,Big Data(BD)has been utilized extensively,which is nothing but a collection of a huge amount of data.However,it is highly complicated along with time-consuming to process a massive amount of data.Thus,to design the Distribution Preserving Framework for BD,a novel methodology has been proposed utilizing Manhattan Distance(MD)-centered Partition Around Medoid(MD–PAM)along with Conjugate Gradient Artificial Neural Network(CG-ANN),which undergoes various steps to reduce the complications of BD.Firstly,the data are processed in the pre-processing phase by mitigating the data repetition utilizing the map-reduce function;subsequently,the missing data are handled by substituting or by ignoring the missed values.After that,the data are transmuted into a normalized form.Next,to enhance the classification performance,the data’s dimensionalities are minimized by employing Gaussian Kernel(GK)-Fisher Discriminant Analysis(GK-FDA).Afterwards,the processed data is submitted to the partitioning phase after transmuting it into a structured format.In the partition phase,by utilizing the MD-PAM,the data are partitioned along with grouped into a cluster.Lastly,by employing CG-ANN,the data are classified in the classification phase so that the needed data can be effortlessly retrieved by the user.To analogize the outcomes of the CG-ANN with the prevailing methodologies,the NSL-KDD openly accessible datasets are utilized.The experiential outcomes displayed that an efficient result along with a reduced computation cost was shown by the proposed CG-ANN.The proposed work outperforms well in terms of accuracy,sensitivity and specificity than the existing systems.
基金This project supported by the National High-Tech Research and Development Plan (863-804-3)
文摘An idea is presented about the development of a data processing and analysis system for ICF experiments, which is based on an object oriented framework. The design and preliminary implementation of the data processing and analysis framework based on the ROOT system have been completed. Software for unfolding soft X-ray spectra has been developed to test the functions of this framework.
文摘Antarctic data management is the research focus, which the international Antarctic organizations, e.g. Antarctic Treaty Consultative Meeting(ATCM) , Scientific Committee on Antarctic Research(SCAR), and Council of Managers of National Antarctic Programmes(COMNAP) have been paying close attention to and promoting actively. Through the co effort of international Antarctic organizations and member countries concerned in recent years, Antarctic Data Directory Syatem(ADDS) is established as the most important basic programme for development of the international Antarctic data management system. At present, Joint Committee on Antarctic Data Management(JCADM) is responsible for organizing and coordinating the international Antarctic data management, and implementing the project ADDS.In this paper the background on Antarctic data management in time sequence and the structure of international framework are introduced, meanwhile, it is necessary to develop ADDS first of all. The ADDS mainly consists of the two principal parts: National Antarctic Data Center(NADCs) of all the party members and Antarctic Main Directory(AMD), the best available technology for creating ADDS is to make full use of International Directory Network(IDN) and adopt its Directory Interchange Formats(DIF). In the light of the above requirements, combined with Chinese specific situation, the contents, technical and administrative methods on Chinese Antarctic data management are discussed to promote our related work.
基金Under the auspices of the National Natural Science Foundation of China(No.71203200)the National Social Science Fund Project(No.20&ZD138)+1 种基金the National Science and Technology Platform Construction Project(No.2005DKA32300)the Major Research Projects of the Ministry of Education(No.16JJD770019)。
文摘With an increase in population and economic development,water withdrawals are close to or even exceed the amount of water available in many regions of the world.Modelling water withdrawals could help water planners improve the efficiency of water use,water resources allocation,and management in order to alleviate water crises.However,minimal information has been obtained on how water withdrawals have changed over space and time,especially on a regional or local scale.This research proposes a data-driven framework to help estimate county-level distribution of water withdrawals.Using this framework,spatial statistical methods are used to estimate water withdrawals for agricultural,industrial,and domestic purposes in the Huaihe River watershed in China for the period 1978–2018.Total water withdrawals were found to have more than doubled,from 292.55×10^(8)m^(3) in 1978 to 642.93×10^(8)m^(3) in 2009,and decreased to 602.63×10^(8)m^(3) in 2018.Agricultural water increased from 208.17×10^(8)m^(3) in 1978 to 435.80×10^(8)m^(3) in 2009 and decreased to 360.84×10^(8)m^(3) in 2018.Industrial and domestic water usage constantly increased throughout the 1978–2018 period.In 1978,industrial and domestic demands were 20.35×10^(8)m^(3) and 60.04×10^(8)m^(3),respectively,and up until 2018,the figures were 105.58×10^(8)m^(3) and 136.20×10^(8)m^(3).From a spatial distribution perspective,Moran’s I statistical results show that the total water withdrawal has significant spatial autocorrelation during 1978–2018.The overall trend was a gradual increase in 1978–2010 with withdrawal beginning to decline in 2010–2018.The results of Getis-Ord G_(i)^(*)statistical calculations showed spatially contiguous clusters of total water withdrawal in the Huaihe River watershed during1978–2010,and the spatial agglomeration weakened from 2010 to 2018.This study provides a data-driven framework for assessing water withdrawals to enable a deeper understanding of competing water use among economic sectors as well as water withdrawal modelled with proper data resource and method.
基金This work was supported in part by National NSF of China(Nos.61872095,61872128,61571139 and 61201393)New Star of Pearl River on Science and Technology of Guangzhou(No.2014J2200085)+2 种基金the Open Project Program of Shenzhen Key Laboratory of Media Security(Grant No.ML-2018-03)the Opening Project of Guang Dong Province Key Laboratory of Information Security Technology(Grant No.2017B030314131-15)Natural Science Foundation of Xizang(No.2016ZR-MZ-01).
文摘This paper proposes a two-step general framework for reversible data hiding(RDH)schemes with controllable contrast enhancement.The first step aims at preserving visual perception as much as possible on the basis of achieving high embedding capacity(EC),while the second step is used for increasing image contrast.In the second step,some peak-pairs are utilized so that the histogram of pixel values is modified to perform histogram equalization(HE),which would lead to the image contrast enhancement.However,for HE,the utilization of some peak-pairs easily leads to over-enhanced image contrast when a large number of bits are embedded.Therefore,in our proposed framework,contrast over-enhancement is avoided by controlling the degree of contrast enhancement.Since the second step can only provide a small amount of data due to controlled contrast enhancement,the first one helps to achieve a large amount of data without degrading visual quality.Any RDH method which can achieve high EC while preserve good visual quality,can be selected for the first step.In fact,Gao et al.’s method is a special case of our proposed framework.In addition,two simple and commonly-used RDH methods are also introduced to further demonstrate the generalization of our framework.
文摘Point of Care (PoC) devices and systems can be categorized into three broad classes (CAT 1, CAT 2, and CAT 3) based on the context of operation and usage. In this paper, the categories are defined to address certain usage models of the PoC device. PoC devices that are used for PoC testing and diagnostic applications are defined CAT 1 devices;PoC devices that are used for patient monitoring are defined as CAT 2 devices (PoCM);PoC devices that are used for as interfacing with other devices are defined as CAT 3 devices (PoCI). The PoCI devices provide an interface gateway for collecting and aggregating data from other medical devices. In all categories, data security is an important aspect. This paper presents a security framework concept, which is applicable for all of the classes of PoC operation. It outlines the concepts and security framework for preventing security challenges in unauthorized access to data, unintended data flow, and data tampering during communication between system entities, the user, and the PoC system. The security framework includes secure layering of basic PoC system architecture, protection of PoC devices in the context of application and network. Developing the security framework is taken into account of a thread model of the PoC system. A proposal for a low-level protocol is discussed. This protocol is independent of communications technologies, and it is elaborated in relation to providing security. An algorithm that can be used to overcome the threat challenges has been shown using the elements in the protocol. The paper further discusses the vulnerability scanning process for the PoC system interconnected network. The paper also presents a four-step process of authentication and authorization framework for providing the security for the PoC system. Finally, the paper concludes with the machine to machine (M2M) security viewpoint and discusses the key stakeholders within an actual deployment of the PoC system and its security challenges.
文摘The importance of the project selection phase in any six sigma initiative cannot be emphasized enough. The successfulness of the six sigma initiative is affected by successful project selection. Recently, Data Envelopment Analysis (DEA) has been proposed as a six sigma project selection tool. However, there exist a number of different DEA formulations which may affect the selection process and the wining project being selected. This work initially applies nine different DEA formulations to several case studies and concludes that different DEA formulations select different wining projects. Also in this work, a Multi-DEA Unified Scoring Framework is proposed to overcome this problem. This framework is applied to several case studies and proved to successfully select the six sigma project with the best performance. The framework is also successful in filtering out some of the projects that have “selective” excellent performance, i.e. projects with excellent performance in some of the DEA formulations and worse performance in others. It is also successful in selecting stable projects;these are projects that perform well in the majority of the DEA formulations, even if it has not been selected as a wining project by any of the DEA formulations.
文摘Digital educational content is gaining importance as an incubator of pedagogical methodologies in formal and informal online educational settings. Its educational efficiency is directly dependent on its quality, however educational content is more than information and data. This paper presents a new data quality framework for assessing digital educational content used for teaching in distance learning environments. The model relies on the ISO2500 series quality standard and beside providing the mechanisms for multi-facet quality assessment it also supports organizations that design, create, manage and use educational content with the quality tools (expressed as quality metrics and measurement methods) to provide a more efficient distance education experience. The model describes the quality characteristics of the educational material content using data and software quality characteristics.
基金the R&D&I,Spain grants PID2020-119478GB-I00 and,PID2020-115832GB-I00 funded by MCIN/AEI/10.13039/501100011033.N.Rodríguez-Barroso was supported by the grant FPU18/04475 funded by MCIN/AEI/10.13039/501100011033 and by“ESF Investing in your future”Spain.J.Moyano was supported by a postdoctoral Juan de la Cierva Formación grant FJC2020-043823-I funded by MCIN/AEI/10.13039/501100011033 and by European Union NextGenerationEU/PRTR.J.Del Ser acknowledges funding support from the Spanish Centro para el Desarrollo Tecnológico Industrial(CDTI)through the AI4ES projectthe Department of Education of the Basque Government(consolidated research group MATHMODE,IT1456-22)。
文摘When data privacy is imposed as a necessity,Federated learning(FL)emerges as a relevant artificial intelligence field for developing machine learning(ML)models in a distributed and decentralized environment.FL allows ML models to be trained on local devices without any need for centralized data transfer,thereby reducing both the exposure of sensitive data and the possibility of data interception by malicious third parties.This paradigm has gained momentum in the last few years,spurred by the plethora of real-world applications that have leveraged its ability to improve the efficiency of distributed learning and to accommodate numerous participants with their data sources.By virtue of FL,models can be learned from all such distributed data sources while preserving data privacy.The aim of this paper is to provide a practical tutorial on FL,including a short methodology and a systematic analysis of existing software frameworks.Furthermore,our tutorial provides exemplary cases of study from three complementary perspectives:i)Foundations of FL,describing the main components of FL,from key elements to FL categories;ii)Implementation guidelines and exemplary cases of study,by systematically examining the functionalities provided by existing software frameworks for FL deployment,devising a methodology to design a FL scenario,and providing exemplary cases of study with source code for different ML approaches;and iii)Trends,shortly reviewing a non-exhaustive list of research directions that are under active investigation in the current FL landscape.The ultimate purpose of this work is to establish itself as a referential work for researchers,developers,and data scientists willing to explore the capabilities of FL in practical applications.
文摘Big data analytics has been widely adopted by large companies to achieve measurable benefits including increased profitability,customer demand forecasting,cheaper development of products,and improved stock control.Small and medium sized enterprises(SMEs)are the backbone of the global economy,comprising of 90%of businesses worldwide.However,only 10%SMEs have adopted big data analytics despite the competitive advantage they could achieve.Previous research has analysed the barriers to adoption and a strategic framework has been developed to help SMEs adopt big data analytics.The framework was converted into a scoring tool which has been applied to multiple case studies of SMEs in the UK.This paper documents the process of evaluating the framework based on the structured feedback from a focus group composed of experienced practitioners.The results of the evaluation are presented with a discussion on the results,and the paper concludes with recommendations to improve the scoring tool based on the proposed framework.The research demonstrates that this positioning tool is beneficial for SMEs to achieve competitive advantages by increasing the application of business intelligence and big data analytics.
文摘The Zambian mining industry is crucial to the national economy but struggles with inconsistent equipment maintenance practices. This study developed an Equipment Maintenance Management Framework (EMMF) tailored to the industry’s needs. Using surveys, interviews, and on-site visits at eight major mining companies, we identified significant variations in maintenance strategies, CMMS usage, and reliability engineering. The EMMF prioritizes predictive maintenance, efficient CMMS implementation, ongoing training, and robust reliability engineering to shift from reactive to proactive maintenance. We recommend adopting continuous improvement practices and data-driven decision-making based on performance metrics, with a phased EMMF implementation aligning maintenance with strategic business objectives. This framework is poised to enhance operational efficiency, equipment reliability, and safety, fostering sustainable growth in the Zambian mining sector.