This work surveys and illustrates multiple open challenges in the field of industrial Internet of Things(IoT)-based big data management and analysis in cloud environments.Challenges arising from the fields of machine ...This work surveys and illustrates multiple open challenges in the field of industrial Internet of Things(IoT)-based big data management and analysis in cloud environments.Challenges arising from the fields of machine learning in cloud infrastructures,artificial intelligence techniques for big data analytics in cloud environments,and federated learning cloud systems are elucidated.Additionally,reinforcement learning,which is a novel technique that allows large cloud-based data centers,to allocate more energy-efficient resources is examined.Moreover,we propose an architecture that attempts to combine the features offered by several cloud providers to achieve an energy-efficient industrial IoT-based big data management framework(EEIBDM)established outside of every user in the cloud.IoT data can be integrated with techniques such as reinforcement and federated learning to achieve a digital twin scenario for the virtual representation of industrial IoT-based big data of machines and room tem-peratures.Furthermore,we propose an algorithm for determining the energy consumption of the infrastructure by evaluating the EEIBDM framework.Finally,future directions for the expansion of this research are discussed.展开更多
Data pre-deployment in the HDFS (Hadoop distributed file systems) is more complicated than that in traditional file systems. There are many key issues need to be addressed, such as determining the target location of...Data pre-deployment in the HDFS (Hadoop distributed file systems) is more complicated than that in traditional file systems. There are many key issues need to be addressed, such as determining the target location of the data prefetching, the amount of data to be prefetched, the balance between data prefetching services and normal data accesses. Aiming to solve these problems, we employ the characteristics of digital ocean information service flows and propose a deployment scheme which combines input data prefetching with output data oriented storage strategies. The method achieves the parallelism of data preparation and data processing, thereby massively reducing I/O time cost of digital ocean cloud computing platforms when processing multi-source information synergistic tasks. The experimental results show that the scheme has a higher degree of parallelism than traditional Hadoop mechanisms, shortens the waiting time of a running service node, and significantly reduces data access conflicts.展开更多
Cloud computing offers numerous web-based services.The adoption of many Cloud applications has been hindered by concerns about data security and privacy.Cloud service providers’access to private information raises mo...Cloud computing offers numerous web-based services.The adoption of many Cloud applications has been hindered by concerns about data security and privacy.Cloud service providers’access to private information raises more security issues.In addition,Cloud computing is incompatible with several industries,including finance and government.Public-key cryptography is frequently cited as a significant advancement in cryptography.In contrast,the Digital Envelope that will be used combines symmetric and asymmetric methods to secure sensitive data.This study aims to design a Digital Envelope for distributed Cloud-based large data security using public-key cryptography.Through strategic design,the hybrid Envelope model adequately supports enterprises delivering routine customer services via independent multi-sourced entities.Both the Cloud service provider and the consumer benefit from the proposed scheme since it results in more resilient and secure services.The suggested approach employs a secret version of the distributed equation to ensure the highest level of security and confidentiality for large amounts of data.Based on the proposed scheme,a Digital Envelope application is developed which prohibits Cloud service providers from directly accessing insufficient or encrypted data.展开更多
The Corona Virus Disease 2019(COVID-19) pandemic has taught us many valuable lessons regarding the importance of our physical and mental health. Even with so many technological advancements, we still lag in developing...The Corona Virus Disease 2019(COVID-19) pandemic has taught us many valuable lessons regarding the importance of our physical and mental health. Even with so many technological advancements, we still lag in developing a system that can fully digitalize the medical data of each individual and make it readily accessible for both the patient and health worker at any point in time. Moreover, there are also no ways for the government to identify the legitimacy of a particular clinic. This study merges modern technology with traditional approaches,thereby highlighting a scenario where artificial intelligence(AI) merges with traditional Chinese medicine(TCM), proposing a way to advance the conventional approaches. The main objective of our research is to provide a one-stop platform for the government, doctors,nurses, and patients to access their data effortlessly. The proposed portal will also check the doctors’ authenticity. Data is one of the most critical assets of an organization, so a breach of data can risk users’ lives. Data security is of primary importance and must be prioritized. The proposed methodology is based on cloud computing technology which assures the security of the data and avoids any kind of breach. The study also accounts for the difficulties encountered in creating such an infrastructure in the cloud and overcomes the hurdles faced during the project, keeping enough room for possible future innovations. To summarize, this study focuses on the digitalization of medical data and suggests some possible ways to achieve it. Moreover, it also focuses on some related aspects like security and potential digitalization difficulties.展开更多
With increasingly complex website structure and continuously advancing web technologies,accurate user clicks recognition from massive HTTP data,which is critical for web usage mining,becomes more difficult.In this pap...With increasingly complex website structure and continuously advancing web technologies,accurate user clicks recognition from massive HTTP data,which is critical for web usage mining,becomes more difficult.In this paper,we propose a dependency graph model to describe the relationships between web requests.Based on this model,we design and implement a heuristic parallel algorithm to distinguish user clicks with the assistance of cloud computing technology.We evaluate the proposed algorithm with real massive data.The size of the dataset collected from a mobile core network is 228.7GB.It covers more than three million users.The experiment results demonstrate that the proposed algorithm can achieve higher accuracy than previous methods.展开更多
Reversible data hiding techniques are capable of reconstructing the original cover image from stego-images. Recently, many researchers have focused on reversible data hiding to protect intellectual property rights. In...Reversible data hiding techniques are capable of reconstructing the original cover image from stego-images. Recently, many researchers have focused on reversible data hiding to protect intellectual property rights. In this paper, we combine reversible data hiding with the chaotic Henon map as an encryption technique to achieve an acceptable level of confidentiality in cloud computing environments. And, Haar digital wavelet transformation (HDWT) is also applied to convert an image from a spatial domain into a frequency domain. And then the decimal of coefficients and integer of high frequency band are modified for hiding secret bits. Finally, the modified coefficients are inversely transformed to stego-images.展开更多
Data outsourcing through cloud storage enables the users to share on-demand resources with cost effective IT services but several security issues arise like confidentiality, integrity and authentication. Each of them ...Data outsourcing through cloud storage enables the users to share on-demand resources with cost effective IT services but several security issues arise like confidentiality, integrity and authentication. Each of them plays an important role in the successful achievement of the other. In cloud computing data integrity assurance is one of the major challenges because the user has no control over the security mechanism to protect the data. Data integrity insures that data received are the same as data stored. It is a result of data security but data integrity refers to validity and accuracy of data rather than protect the data. Data security refers to protection of data against unauthorized access, modification or corruption and it is necessary to ensure data integrity. This paper proposed a new approach using Matrix Dialing Method in block level to enhance the performance of both data integrity and data security without using Third Party Auditor (TPA). In this approach, the data are partitioned into number of blocks and each block converted into a square matrix. Determinant factor of each matrix is generated dynamically to ensure data integrity. This model also implements a combination of AES algorithm and SHA-1 algorithm for digital signature generation. Data coloring on digital signature is applied to ensure data security with better performance. The performance analysis using cloud simulator shows that the proposed scheme is highly efficient and secure as it overcomes the limitations of previous approaches of data security using encryption and decryption algorithms and data integrity assurance using TPA due to server computation time and accuracy.展开更多
Digital data have become a torrent engulfing every area of business, science and engineering disciplines, gushing into every economy, every organization and every user of digital technology. In the age of big data, de...Digital data have become a torrent engulfing every area of business, science and engineering disciplines, gushing into every economy, every organization and every user of digital technology. In the age of big data, deriving values and insights from big data using rich analytics becomes important for achieving competitiveness, success and leadership in every field. The Internet of Things (IoT) is causing the number and types of products to emit data at an unprecedented rate. Heterogeneity, scale, timeliness, complexity, and privacy problems with large data impede progress at all phases of the pipeline that can create value from data issues. With the push of such massive data, we are entering a new era of computing driven by novel and ground breaking research innovation on elastic parallelism, partitioning and scalability. Designing a scalable system for analysing, processing and mining huge real world datasets has become one of the challenging problems facing both systems researchers and data management researchers. In this paper, we will give an overview of computing infrastructure for IoT data processing, focusing on architectural and major challenges of massive data. We will briefly discuss about emerging computing infrastructure and technologies that are promising for improving massive data management.展开更多
In light of the coronavirus disease 2019(COVID-19)outbreak caused by the novel coronavirus,companies and institutions have instructed their employees to work from home as a precautionary measure to reduce the risk of ...In light of the coronavirus disease 2019(COVID-19)outbreak caused by the novel coronavirus,companies and institutions have instructed their employees to work from home as a precautionary measure to reduce the risk of contagion.Employees,however,have been exposed to different security risks because of working from home.Moreover,the rapid global spread of COVID-19 has increased the volume of data generated from various sources.Working from home depends mainly on cloud computing(CC)applications that help employees to efficiently accomplish their tasks.The cloud computing environment(CCE)is an unsung hero in the COVID-19 pandemic crisis.It consists of the fast-paced practices for services that reflect the trend of rapidly deployable applications for maintaining data.Despite the increase in the use of CC applications,there is an ongoing research challenge in the domains of CCE concerning data,guaranteeing security,and the availability of CC applications.This paper,to the best of our knowledge,is the first paper that thoroughly explains the impact of the COVID-19 pandemic on CCE.Additionally,this paper also highlights the security risks of working from home during the COVID-19 pandemic.展开更多
文摘This work surveys and illustrates multiple open challenges in the field of industrial Internet of Things(IoT)-based big data management and analysis in cloud environments.Challenges arising from the fields of machine learning in cloud infrastructures,artificial intelligence techniques for big data analytics in cloud environments,and federated learning cloud systems are elucidated.Additionally,reinforcement learning,which is a novel technique that allows large cloud-based data centers,to allocate more energy-efficient resources is examined.Moreover,we propose an architecture that attempts to combine the features offered by several cloud providers to achieve an energy-efficient industrial IoT-based big data management framework(EEIBDM)established outside of every user in the cloud.IoT data can be integrated with techniques such as reinforcement and federated learning to achieve a digital twin scenario for the virtual representation of industrial IoT-based big data of machines and room tem-peratures.Furthermore,we propose an algorithm for determining the energy consumption of the infrastructure by evaluating the EEIBDM framework.Finally,future directions for the expansion of this research are discussed.
基金The Ocean Public Welfare Scientific Research Project of State Oceanic Administration of China under contract No.20110533
文摘Data pre-deployment in the HDFS (Hadoop distributed file systems) is more complicated than that in traditional file systems. There are many key issues need to be addressed, such as determining the target location of the data prefetching, the amount of data to be prefetched, the balance between data prefetching services and normal data accesses. Aiming to solve these problems, we employ the characteristics of digital ocean information service flows and propose a deployment scheme which combines input data prefetching with output data oriented storage strategies. The method achieves the parallelism of data preparation and data processing, thereby massively reducing I/O time cost of digital ocean cloud computing platforms when processing multi-source information synergistic tasks. The experimental results show that the scheme has a higher degree of parallelism than traditional Hadoop mechanisms, shortens the waiting time of a running service node, and significantly reduces data access conflicts.
文摘Cloud computing offers numerous web-based services.The adoption of many Cloud applications has been hindered by concerns about data security and privacy.Cloud service providers’access to private information raises more security issues.In addition,Cloud computing is incompatible with several industries,including finance and government.Public-key cryptography is frequently cited as a significant advancement in cryptography.In contrast,the Digital Envelope that will be used combines symmetric and asymmetric methods to secure sensitive data.This study aims to design a Digital Envelope for distributed Cloud-based large data security using public-key cryptography.Through strategic design,the hybrid Envelope model adequately supports enterprises delivering routine customer services via independent multi-sourced entities.Both the Cloud service provider and the consumer benefit from the proposed scheme since it results in more resilient and secure services.The suggested approach employs a secret version of the distributed equation to ensure the highest level of security and confidentiality for large amounts of data.Based on the proposed scheme,a Digital Envelope application is developed which prohibits Cloud service providers from directly accessing insufficient or encrypted data.
文摘The Corona Virus Disease 2019(COVID-19) pandemic has taught us many valuable lessons regarding the importance of our physical and mental health. Even with so many technological advancements, we still lag in developing a system that can fully digitalize the medical data of each individual and make it readily accessible for both the patient and health worker at any point in time. Moreover, there are also no ways for the government to identify the legitimacy of a particular clinic. This study merges modern technology with traditional approaches,thereby highlighting a scenario where artificial intelligence(AI) merges with traditional Chinese medicine(TCM), proposing a way to advance the conventional approaches. The main objective of our research is to provide a one-stop platform for the government, doctors,nurses, and patients to access their data effortlessly. The proposed portal will also check the doctors’ authenticity. Data is one of the most critical assets of an organization, so a breach of data can risk users’ lives. Data security is of primary importance and must be prioritized. The proposed methodology is based on cloud computing technology which assures the security of the data and avoids any kind of breach. The study also accounts for the difficulties encountered in creating such an infrastructure in the cloud and overcomes the hurdles faced during the project, keeping enough room for possible future innovations. To summarize, this study focuses on the digitalization of medical data and suggests some possible ways to achieve it. Moreover, it also focuses on some related aspects like security and potential digitalization difficulties.
基金supported in part by the Fundamental Research Funds for the Central Universities under Grant No.2013RC0114111 Project of China under Grant No.B08004
文摘With increasingly complex website structure and continuously advancing web technologies,accurate user clicks recognition from massive HTTP data,which is critical for web usage mining,becomes more difficult.In this paper,we propose a dependency graph model to describe the relationships between web requests.Based on this model,we design and implement a heuristic parallel algorithm to distinguish user clicks with the assistance of cloud computing technology.We evaluate the proposed algorithm with real massive data.The size of the dataset collected from a mobile core network is 228.7GB.It covers more than three million users.The experiment results demonstrate that the proposed algorithm can achieve higher accuracy than previous methods.
文摘Reversible data hiding techniques are capable of reconstructing the original cover image from stego-images. Recently, many researchers have focused on reversible data hiding to protect intellectual property rights. In this paper, we combine reversible data hiding with the chaotic Henon map as an encryption technique to achieve an acceptable level of confidentiality in cloud computing environments. And, Haar digital wavelet transformation (HDWT) is also applied to convert an image from a spatial domain into a frequency domain. And then the decimal of coefficients and integer of high frequency band are modified for hiding secret bits. Finally, the modified coefficients are inversely transformed to stego-images.
文摘Data outsourcing through cloud storage enables the users to share on-demand resources with cost effective IT services but several security issues arise like confidentiality, integrity and authentication. Each of them plays an important role in the successful achievement of the other. In cloud computing data integrity assurance is one of the major challenges because the user has no control over the security mechanism to protect the data. Data integrity insures that data received are the same as data stored. It is a result of data security but data integrity refers to validity and accuracy of data rather than protect the data. Data security refers to protection of data against unauthorized access, modification or corruption and it is necessary to ensure data integrity. This paper proposed a new approach using Matrix Dialing Method in block level to enhance the performance of both data integrity and data security without using Third Party Auditor (TPA). In this approach, the data are partitioned into number of blocks and each block converted into a square matrix. Determinant factor of each matrix is generated dynamically to ensure data integrity. This model also implements a combination of AES algorithm and SHA-1 algorithm for digital signature generation. Data coloring on digital signature is applied to ensure data security with better performance. The performance analysis using cloud simulator shows that the proposed scheme is highly efficient and secure as it overcomes the limitations of previous approaches of data security using encryption and decryption algorithms and data integrity assurance using TPA due to server computation time and accuracy.
文摘Digital data have become a torrent engulfing every area of business, science and engineering disciplines, gushing into every economy, every organization and every user of digital technology. In the age of big data, deriving values and insights from big data using rich analytics becomes important for achieving competitiveness, success and leadership in every field. The Internet of Things (IoT) is causing the number and types of products to emit data at an unprecedented rate. Heterogeneity, scale, timeliness, complexity, and privacy problems with large data impede progress at all phases of the pipeline that can create value from data issues. With the push of such massive data, we are entering a new era of computing driven by novel and ground breaking research innovation on elastic parallelism, partitioning and scalability. Designing a scalable system for analysing, processing and mining huge real world datasets has become one of the challenging problems facing both systems researchers and data management researchers. In this paper, we will give an overview of computing infrastructure for IoT data processing, focusing on architectural and major challenges of massive data. We will briefly discuss about emerging computing infrastructure and technologies that are promising for improving massive data management.
文摘In light of the coronavirus disease 2019(COVID-19)outbreak caused by the novel coronavirus,companies and institutions have instructed their employees to work from home as a precautionary measure to reduce the risk of contagion.Employees,however,have been exposed to different security risks because of working from home.Moreover,the rapid global spread of COVID-19 has increased the volume of data generated from various sources.Working from home depends mainly on cloud computing(CC)applications that help employees to efficiently accomplish their tasks.The cloud computing environment(CCE)is an unsung hero in the COVID-19 pandemic crisis.It consists of the fast-paced practices for services that reflect the trend of rapidly deployable applications for maintaining data.Despite the increase in the use of CC applications,there is an ongoing research challenge in the domains of CCE concerning data,guaranteeing security,and the availability of CC applications.This paper,to the best of our knowledge,is the first paper that thoroughly explains the impact of the COVID-19 pandemic on CCE.Additionally,this paper also highlights the security risks of working from home during the COVID-19 pandemic.