Deterministic databases are able to reduce coordination costs in a replication.This property has fostered a significant interest in the design of efficient deterministic concurrency control protocols.However,the state...Deterministic databases are able to reduce coordination costs in a replication.This property has fostered a significant interest in the design of efficient deterministic concurrency control protocols.However,the state-of-the-art deterministic concurrency control protocol Aria has three issues.First,it is impractical to configure a suitable batch size when the read-write set is unknown.Second,Aria running in low-concurrency scenarios,e.g.,a single-thread scenario,suffers from the same conflicts as running in high-concurrency scenarios.Third,the single-version schema brings write-after-write conflicts.To address these issues,we propose Gria,an efficient deterministic concurrency control protocol.Gria has the following properties.First,the batch size of Gria is auto-scaling.Second,Gria’s conflict probability in low-concurrency scenarios is lower than that in high-concurrency scenarios.Third,Gria has no write-after-write conflicts by adopting a multi-version structure.To further reduce conflicts,we propose two optimizations:a reordering mechanism as well as a rechecking strategy.The evaluation result on two popular benchmarks shows that Gria outperforms Aria by 13x.展开更多
Benchmarks play a crucial role in database performance evaluation,and have been effectively promoting the development of database management systems.With critical transaction processing requirements of new application...Benchmarks play a crucial role in database performance evaluation,and have been effectively promoting the development of database management systems.With critical transaction processing requirements of new applications,we see an explosion of innovative database technologies for dealing with highly intensive transaction workloads(OLTP)with the obvious characteristics of sharp dynamics,terrificskewness,high contention,or high concurrency(abbr.DSC^2),which can not be well described or evaluated by current standard benchmarks.In this paper,based on the representative SecKill applications,we define a pacakge of workloads simulating intensive transactional processing requirements.And we create a general and flexible benchmark framework PeakBench for evaluating intensive OLTP workloads on databases.We are the first work to have full control on simulating DSC^2,especially for the fine granularity control for contention generation.With a comprehensive set of experiments conducted on popular open sourced DBMSs compared with the other representative OLTP benchmarks,we completely demonstrate the usefulness of PeakBench.展开更多
Many business processes in enterprise applications are both long running and transactional in nature. However, no current transaction model can provide full transaction support for such long running business processes...Many business processes in enterprise applications are both long running and transactional in nature. However, no current transaction model can provide full transaction support for such long running business processes. This paper proposes a new transaction model, the pessimistic predicate/transform (PP/T) model, which can provide full transaction support for long running business processes. A framework was proposed on the enterprise JavaBeans platform to implement the PP/T model. The framework enables application developers to focus on the business logic, with the underlying platform providing the required transactional semantics. The development and maintenance effort are therefore greatly reduced. Simulations show that the model has a sound concurrency management ability for long running business processes.展开更多
Modern database systems desperate for the ability to support highly scalable transactions and efficient queries simultaneously for real-time applications.One solution is to utilize query optimization techniques on the...Modern database systems desperate for the ability to support highly scalable transactions and efficient queries simultaneously for real-time applications.One solution is to utilize query optimization techniques on the on-line transaction processing(OLTP)systems.The materialized view is considered as a panacea to decrease query latency.However,it also involves the significant cost of maintenance which trades away transaction performance.In this paper,we examine the design space and conclude several design features for the implementation of a view on a distributed log-structured merge-tree(LSMtree),which is a well-known structure for improving data write performance.As a result,we develop two incremental view maintenance(IVM)approaches on LSM-tree.One avoids join computation in view maintenance transactions.Another with two optimizations is proposed to decouple the view maintenance with the transaction process.Under the asynchronous update,we also provide consistency queries for views.Experiments on TPC-H benchmark show our methods achieve better performance than straightforward methods on different workloads.展开更多
Logging and replication are commonly used recovery approaches in database systems.To guarantee that the database state is not corrupted due to system crash,database systems rely on a centralized logging method to pers...Logging and replication are commonly used recovery approaches in database systems.To guarantee that the database state is not corrupted due to system crash,database systems rely on a centralized logging method to persist log entries into a stable storage device;to prevent data loss due to device failure,a primary server in the database system periodically replicates its state to backup servers by copying log entries over networks.As the transaction execution in a modern database system is highly parallelized,the centralized logging with a single I/O channel tends to inhibit the scalability of the system.Meanwhile,log entries generated at high speed make a network with limited bandwidth a potential bottleneck for replication.In this paper,we propose an in-memory transaction engine named Plover with parallel logging and speedy replication for primary-backup replication systems.The parallel logging enables concurrent execution of logging by utilizing multiple log buffers associated with multiple stable storages.All log entries in the log buffers maintain a global sequence number(GSN),which ensures a partial order among transactions.The kernel of the speedy replication is an adaptive shipping method,which allows to transfer data increments instead of log entries to backups under heavy workloads.Experimental results using the YCSB and TPC-C benchmarks show that Plover scales well with the increasing number of worker threads and stable storage devices.And our adaptive shipping requires only one fifth network bandwidth of the conventional log shipping.展开更多
Cloud manufacturing is a new manufacturing paradigm which creates an open environment for transactions among the enterprises.Research on transaction modes and regularities in a cloud manufacturing environment is impor...Cloud manufacturing is a new manufacturing paradigm which creates an open environment for transactions among the enterprises.Research on transaction modes and regularities in a cloud manufacturing environment is important for promoting the applications of cloud manufacturing.To this end,we design and implement a simulation platform according to the typical transaction processes of enterprises in the cloud manufacturing environment.In the simulation platform,enterprises are encapsulated into Service Agents,and thus the activities of service agents can be used to describe enterprise behaviors.By defining different rules,simulations for different business models can be conducted.Detailed descriptions of the platform architecture,functions,and key technologies are presented.The feasibility of the simulation platform is verified through a case study.展开更多
基金Key Laboratory Ministry of Industry and Information Technology of China,Shenzhen Science and Technology Program(Nos.GXWD20220817124827001 and JCYJ20210324132406016)National Natural Science Foundation of China(Grant No.61732022)National Science Foundation of Guangdong Province(No.2023A1515030273).
文摘Deterministic databases are able to reduce coordination costs in a replication.This property has fostered a significant interest in the design of efficient deterministic concurrency control protocols.However,the state-of-the-art deterministic concurrency control protocol Aria has three issues.First,it is impractical to configure a suitable batch size when the read-write set is unknown.Second,Aria running in low-concurrency scenarios,e.g.,a single-thread scenario,suffers from the same conflicts as running in high-concurrency scenarios.Third,the single-version schema brings write-after-write conflicts.To address these issues,we propose Gria,an efficient deterministic concurrency control protocol.Gria has the following properties.First,the batch size of Gria is auto-scaling.Second,Gria’s conflict probability in low-concurrency scenarios is lower than that in high-concurrency scenarios.Third,Gria has no write-after-write conflicts by adopting a multi-version structure.To further reduce conflicts,we propose two optimizations:a reordering mechanism as well as a rechecking strategy.The evaluation result on two popular benchmarks shows that Gria outperforms Aria by 13x.
基金We are partially supported by the Key Program of National Natural Science Foundation of China(2018YFB1003402)the National Natural Science Foundation of China(Grant No.61432006).
文摘Benchmarks play a crucial role in database performance evaluation,and have been effectively promoting the development of database management systems.With critical transaction processing requirements of new applications,we see an explosion of innovative database technologies for dealing with highly intensive transaction workloads(OLTP)with the obvious characteristics of sharp dynamics,terrificskewness,high contention,or high concurrency(abbr.DSC^2),which can not be well described or evaluated by current standard benchmarks.In this paper,based on the representative SecKill applications,we define a pacakge of workloads simulating intensive transactional processing requirements.And we create a general and flexible benchmark framework PeakBench for evaluating intensive OLTP workloads on databases.We are the first work to have full control on simulating DSC^2,especially for the fine granularity control for contention generation.With a comprehensive set of experiments conducted on popular open sourced DBMSs compared with the other representative OLTP benchmarks,we completely demonstrate the usefulness of PeakBench.
基金Supported by the National Key Basic Research and Development (973) Program of China (No. 2002CB312005), and the National High-Tech Research and Development (863) Program of China (No. 2003AA115440)
文摘Many business processes in enterprise applications are both long running and transactional in nature. However, no current transaction model can provide full transaction support for such long running business processes. This paper proposes a new transaction model, the pessimistic predicate/transform (PP/T) model, which can provide full transaction support for long running business processes. A framework was proposed on the enterprise JavaBeans platform to implement the PP/T model. The framework enables application developers to focus on the business logic, with the underlying platform providing the required transactional semantics. The development and maintenance effort are therefore greatly reduced. Simulations show that the model has a sound concurrency management ability for long running business processes.
基金This work was partially supported by Youth Foundation of National Science Foundation(61702189)National Science Foundation(61772202).
文摘Modern database systems desperate for the ability to support highly scalable transactions and efficient queries simultaneously for real-time applications.One solution is to utilize query optimization techniques on the on-line transaction processing(OLTP)systems.The materialized view is considered as a panacea to decrease query latency.However,it also involves the significant cost of maintenance which trades away transaction performance.In this paper,we examine the design space and conclude several design features for the implementation of a view on a distributed log-structured merge-tree(LSMtree),which is a well-known structure for improving data write performance.As a result,we develop two incremental view maintenance(IVM)approaches on LSM-tree.One avoids join computation in view maintenance transactions.Another with two optimizations is proposed to decouple the view maintenance with the transaction process.Under the asynchronous update,we also provide consistency queries for views.Experiments on TPC-H benchmark show our methods achieve better performance than straightforward methods on different workloads.
基金This work was partially supported by National Key R&D Program of China(2018YFB1003303)the National Natural Science Foundation of China(Grant Nos.61672232 and 61772202)+1 种基金Youth Foundation of National Science Foundation(61702189)Youth Science and Technology—“Yang Fan”Program of Shanghai(17YF1427800).
文摘Logging and replication are commonly used recovery approaches in database systems.To guarantee that the database state is not corrupted due to system crash,database systems rely on a centralized logging method to persist log entries into a stable storage device;to prevent data loss due to device failure,a primary server in the database system periodically replicates its state to backup servers by copying log entries over networks.As the transaction execution in a modern database system is highly parallelized,the centralized logging with a single I/O channel tends to inhibit the scalability of the system.Meanwhile,log entries generated at high speed make a network with limited bandwidth a potential bottleneck for replication.In this paper,we propose an in-memory transaction engine named Plover with parallel logging and speedy replication for primary-backup replication systems.The parallel logging enables concurrent execution of logging by utilizing multiple log buffers associated with multiple stable storages.All log entries in the log buffers maintain a global sequence number(GSN),which ensures a partial order among transactions.The kernel of the speedy replication is an adaptive shipping method,which allows to transfer data increments instead of log entries to backups under heavy workloads.Experimental results using the YCSB and TPC-C benchmarks show that Plover scales well with the increasing number of worker threads and stable storage devices.And our adaptive shipping requires only one fifth network bandwidth of the conventional log shipping.
基金the National Natural Science Foundation of China(Grant No.61374199)the National High-Tech Research and Development Plan of China under(Grant No.2015AA042101)+1 种基金the Beijing Municipal Natural Science Foundation(Grant No.4142031)the State Key Laboratory of Intelligent Manufacturing,System Technology,Beijing Institute of Electronic System Engineering,Beijing,P.R.China.
文摘Cloud manufacturing is a new manufacturing paradigm which creates an open environment for transactions among the enterprises.Research on transaction modes and regularities in a cloud manufacturing environment is important for promoting the applications of cloud manufacturing.To this end,we design and implement a simulation platform according to the typical transaction processes of enterprises in the cloud manufacturing environment.In the simulation platform,enterprises are encapsulated into Service Agents,and thus the activities of service agents can be used to describe enterprise behaviors.By defining different rules,simulations for different business models can be conducted.Detailed descriptions of the platform architecture,functions,and key technologies are presented.The feasibility of the simulation platform is verified through a case study.