Various organizations store data online rather than on physical servers.As the number of user’s data stored in cloud servers increases,the attack rate to access data from cloud servers also increases.Different resear...Various organizations store data online rather than on physical servers.As the number of user’s data stored in cloud servers increases,the attack rate to access data from cloud servers also increases.Different researchers worked on different algorithms to protect cloud data from replay attacks.None of the papers used a technique that simultaneously detects a full-message and partial-message replay attack.This study presents the development of a TKN(Text,Key and Name)cryptographic algorithm aimed at protecting data from replay attacks.The program employs distinct ways to encrypt plain text[P],a user-defined Key[K],and a Secret Code[N].The novelty of the TKN cryptographic algorithm is that the bit value of each text is linked to another value with the help of the proposed algorithm,and the length of the cipher text obtained is twice the length of the original text.In the scenario that an attacker executes a replay attack on the cloud server,engages in cryptanalysis,or manipulates any data,it will result in automated modification of all associated values inside the backend.This mechanism has the benefit of enhancing the detectability of replay attacks.Nevertheless,the attacker cannot access data not included in any of the papers,regardless of how effective the attack strategy is.At the end of paper,the proposed algorithm’s novelty will be compared with different algorithms,and it will be discussed how far the proposed algorithm is better than all other algorithms.展开更多
File semantic has proven effective in optimizing large scale distributed file system.As a consequence of the elaborate and rich I/O interfaces between upper layer applications and file systems,file system can provide ...File semantic has proven effective in optimizing large scale distributed file system.As a consequence of the elaborate and rich I/O interfaces between upper layer applications and file systems,file system can provide useful and insightful information about semantic.Hence,file semantic mining has become an increasingly important practice in both engineering and research community.Unfortunately,it is a challenge to exploit file semantic knowledge because a variety of factors coulda ffect this information exploration process.Even worse,the challenges are exacerbated due to the intricate interdependency between these factors,and make it difficult to fully exploit the potentially important correlation among various semantic knowledges.This article proposes a file access correlation miming and evaluation reference(FARMER) model,where file is treated as a multivariate vector space,and each item within the vector corresponds a separate factor of the given file.The selection of factor depends on the application,examples of factors are file path,creator and executing program.If one particular factor occurs in both files,its value is non-zero.It is clear that the extent of inter-file relationships can be measured based on the likeness of their factor values in the semantic vectors.Benefit from this model,FARMER represents files as structured vectors of identifiers,and basic vector operations can be leveraged to quantify file correlation between two file vectors.FARMER model leverages linear regression model to estimate the strength of the relationship between file correlation and a set of influencing factors so that the "bad knowledge" can be filtered out.To demonstrate the ability of new FARMER model,FARMER is incorporated into a real large-scale object-based storage system as a case study to dynamically infer file correlations.In addition FARMER-enabled optimize service for metadata prefetching algorithm and object data layout algorithm is implemented.Experimental results show that is FARMER-enabled prefetching algorithm is shown to reduce the metadata operations latency by approximately 30%-40% when compared to a state-of-the-art metadata prefetching algorithm and a commonly used replacement policy.展开更多
To better understand different users' accessing intentions, a novel clustering and supervising method based on accessing path is presented. This method divides users' interest space to express the distribution...To better understand different users' accessing intentions, a novel clustering and supervising method based on accessing path is presented. This method divides users' interest space to express the distribution of users' interests, and directly to instruct the constructing process of web pages indexing for advanced performance.展开更多
In order to improve the management strategy for personnel files in colleges and universities,simplify the complex process of file management,and improve file management security and content preservation of the files.T...In order to improve the management strategy for personnel files in colleges and universities,simplify the complex process of file management,and improve file management security and content preservation of the files.This paper elaborates on the application of Artificial Intelligence(AI)technology in university personnel file management through theoretical analysis based on the understanding of Al technology.展开更多
A file transfer algorithm based on ADP (autonomous decentrafized protocol) was proposed to solve the problem that the ADS (autonomous decentralized system) middleware (NeXUS/Drink) lacks of file transfer functio...A file transfer algorithm based on ADP (autonomous decentrafized protocol) was proposed to solve the problem that the ADS (autonomous decentralized system) middleware (NeXUS/Drink) lacks of file transfer functions for Windows. The algorithm realizes the peer-to-peer file transfer, one-to-N inquiry/multi-response file transfer and one-to-N file distribution in the same data field based on communication patterns provided by the ADP. The peer-to-peer file transfer is implemented through a peer-to-peer communication path, one-to-N inquiry/multi-response file transfer and one-to-N file distribution are implemented through multicast communication. In this algorithm, a file to be transferred is named with a GUID (global unique identification), every data packet is marked with a sequence number, and file-receiving in parallel is implemented by caching DPOs(data processing objects) and multi-thread technologies. The algorithm is applied in a simulation system of the decentralized control platform, and the test results and long time stable tanning prove the feasibility of the algorithm.展开更多
EndNote is a reference management software package produced by Thomson Reuters used to manage bibliographies and references in research essays and articles. EndNote works by grouping citations into "libraries" with ...EndNote is a reference management software package produced by Thomson Reuters used to manage bibliographies and references in research essays and articles. EndNote works by grouping citations into "libraries" with the file extension *.enl and a corresponding *.data folder. Aside from the library function,展开更多
The rapid development of Internet of Things(IoT)technology has made previously unavailable data available,and applications can take advantage of device data for people to visualize,explore,and build complex analyses.A...The rapid development of Internet of Things(IoT)technology has made previously unavailable data available,and applications can take advantage of device data for people to visualize,explore,and build complex analyses.As the size of the network and the number of network users continue to increase,network requests tend to aggregate on a small number of network resources,which results in uneven load on network requests.Real-time,highly reliable network file distribution technology is of great importance in the Internet of Things.This paper studies real-time and highly reliable file distribution technology for large-scale networks.In response to this topic,this paper studies the current file distribution technology,proposes a file distribution model,and proposes a corresponding load balancing method based on the file distribution model.Experiments show that the system has achieved real-time and high reliability of network transmission.展开更多
Deep space communication has its own features such as long propagation delays,heavy noise,asymmetric link rates,and intermittent connectivity in space,therefore TCP/IP protocol cannot perform as well as it does in ter...Deep space communication has its own features such as long propagation delays,heavy noise,asymmetric link rates,and intermittent connectivity in space,therefore TCP/IP protocol cannot perform as well as it does in terrestrial communications.Accordingly,the Consultative Committee for Space Data Systems(CCSDS) developed CCSDS File Delivery Protocol(CFDP),which sets standards of efficient file delivery service capable of transferring files to and from mass memory located in the space segment.In CFDP,four optional acknowledge modes are supported to make the communication more reliable.In this paper,we gave a general introduction of typical communication process in CFDP and analysis of its four Negative Acknowledgement(NAK) modes on the respect of file delivery delay and times of retransmission.We found out that despite the shortest file delivery delay,immediate NAK mode suffers from the problem that frequent retransmission may probably lead to network congestion.Thus,we proposed a new mode,the error counter-based NAK mode.By simulation of the case focused on the link between a deep space probe on Mars and a ter-restrial station on Earth,we concluded that error counter-based NAK mode has successfully reduced the retransmission times at negligible cost of certain amount of file delivery delay.展开更多
1 Geological Setting The Huayuan Pb-Zn ore field in Xiangxi is located in the southeastern margin of the Yangtze block and the mid-segment of the West Hunan-West Hubei metallogenic belt.The exposed stratum are the lower
ZTE Corporation announced on February 12, 2010 that by late 2009 it had applied for more than 1,700 LTE/SAE patents. ZTE is focused on advancing the LTE market with strong innovation that will benefit both consumers a...ZTE Corporation announced on February 12, 2010 that by late 2009 it had applied for more than 1,700 LTE/SAE patents. ZTE is focused on advancing the LTE market with strong innovation that will benefit both consumers and operators with 4G experience and functionality.展开更多
This article firstly described the related definition and technology of Ajax. Combined with Freemarker template engine with Spring and Hibernate framework technology to generate a page view, put forward the FSH the fr...This article firstly described the related definition and technology of Ajax. Combined with Freemarker template engine with Spring and Hibernate framework technology to generate a page view, put forward the FSH the framework of a complete program, and realize upload with progress bar display.展开更多
So far, file access prediction models is mainly based on either the file access frequency or the historical record of the latest access. In this paper, a new file access prediction model called frequency- and recency-...So far, file access prediction models is mainly based on either the file access frequency or the historical record of the latest access. In this paper, a new file access prediction model called frequency- and recency-based successor (FRS) is presented which combines the advantages of the file frequency with the historical record. FRS model has the capability of rapid response to workload changes and can predict future events with greater accuracy than most of other prediction models. To evaluate the performance of FRS mode, the Linux kernel is modified to predict and prefetch upcoming accesses. The experiment shows that FRS can accurately predict approximately 80% of all file access events, while maintaining an immediate successor queue (ISQ) per-file which only requires regular dynamic updates.展开更多
File labeling techniques have a long history in analyzing the anthological trends in computational linguistics.The situation becomes worse in the case of files downloaded into systems from the Internet.Currently,most ...File labeling techniques have a long history in analyzing the anthological trends in computational linguistics.The situation becomes worse in the case of files downloaded into systems from the Internet.Currently,most users either have to change file names manually or leave a meaningless name of the files,which increases the time to search required files and results in redundancy and duplications of user files.Currently,no significant work is done on automated file labeling during the organization of heterogeneous user files.A few attempts have been made in topic modeling.However,one major drawback of current topic modeling approaches is better results.They rely on specific language types and domain similarity of the data.In this research,machine learning approaches have been employed to analyze and extract the information from heterogeneous corpus.A different file labeling technique has also been used to get the meaningful and`cohesive topic of the files.The results show that the proposed methodology can generate relevant and context-sensitive names for heterogeneous data files and provide additional insight into automated file labeling in operating systems.展开更多
In the Big Data era,numerous sources and environments generate massive amounts of data.This enormous amount of data necessitates specialized advanced tools and procedures that effectively evaluate the information and ...In the Big Data era,numerous sources and environments generate massive amounts of data.This enormous amount of data necessitates specialized advanced tools and procedures that effectively evaluate the information and anticipate decisions for future changes.Hadoop is used to process this kind of data.It is known to handle vast volumes of data more efficiently than tiny amounts,which results in inefficiency in the framework.This study proposes a novel solution to the problem by applying the Enhanced Best Fit Merging algorithm(EBFM)that merges files depending on predefined parameters(type and size).Implementing this algorithm will ensure that the maximum amount of the block size and the generated file size will be in the same range.Its primary goal is to dynamically merge files with the stated criteria based on the file type to guarantee the efficacy and efficiency of the established system.This procedure takes place before the files are available for the Hadoop framework.Additionally,the files generated by the system are named with specific keywords to ensure there is no data loss(file overwrite).The proposed approach guarantees the generation of the fewest possible large files,which reduces the input/output memory burden and corresponds to the Hadoop framework’s effectiveness.The findings show that the proposed technique enhances the framework’s performance by approximately 64%while comparing all other potential performance-impairing variables.The proposed approach is implementable in any environment that uses the Hadoop framework,not limited to smart cities,real-time data analysis,etc.展开更多
Working with files and the safety of information has always been relevant, especially in financial institutions where the requirements for the safety of information and security are especially important. And in today...Working with files and the safety of information has always been relevant, especially in financial institutions where the requirements for the safety of information and security are especially important. And in today’s conditions, when an earthquake can destroy the floor of a city in an instant, or when a missile hits an office and all servers turn into scrap metal, the issue of data safety becomes especially important. Also, you can’t put the cost of the software and the convenience of working with files in last place. Especially if an office worker needs to find the necessary information on a client, a financial contract or a company’s financial product in a few seconds. Also, during the operation of computer equipment, failures are possible, and some of them can lead to partial or complete loss of information. In this paper, it is proposed to create another level of abstraction for working with the file system, which will be based on a relational database as a storage of objects and access rights to objects. Also considered are possible protocols for transferring data to other programs that work with files, these can be both small sites and the operating system itself. This article will be especially interesting for financial institutions or companies operating in the banking sector. The purpose of this article is an attempt to introduce another level of abstraction for working with files. A level that is completely abstracted from the storage medium.展开更多
Byte-addressable non-volatile memory(NVM),as a new participant in the storage hierarchy,gives extremely high performance in storage,which forces changes to be made on current filesystem designs.Page cache,once a signi...Byte-addressable non-volatile memory(NVM),as a new participant in the storage hierarchy,gives extremely high performance in storage,which forces changes to be made on current filesystem designs.Page cache,once a significant mechanism filling the performance gap between Dynamic Random Access Memory(DRAM)and block devices,is now a liability that heavily hinders the writing performance of NVM filesystems.Therefore state-of-the-art NVM filesystems leverage the direct access(DAX)technology to bypass the page cache entirely.However,the DRAM still provides higher bandwidth than NVM,which prevents skewed read workloads from benefiting from a higher bandwidth of the DRAM and leads to sub-optimal performance for the system.In this paper,we propose RCache,a readintensive workload-aware page cache for NVM filesystems.Different from traditional caching mechanisms where all reads go through DRAM,RCache uses a tiered page cache design,including assigning DRAM and NVM to hot and cold data separately,and reading data from both sides.To avoid copying data to DRAM in a critical path,RCache migrates data from NVM to DRAM in a background thread.Additionally,RCache manages data in DRAM in a lock-free manner for better latency and scalability.Evaluations on Intel Optane Data Center(DC)Persistent Memory Modules show that,compared with NOVA,RCache achieves 3 times higher bandwidth for read-intensive workloads and introduces little performance loss for write operations.展开更多
基金Deanship of Scientific Research at Majmaah University for supporting this work under Project Number R-2023-811.
文摘Various organizations store data online rather than on physical servers.As the number of user’s data stored in cloud servers increases,the attack rate to access data from cloud servers also increases.Different researchers worked on different algorithms to protect cloud data from replay attacks.None of the papers used a technique that simultaneously detects a full-message and partial-message replay attack.This study presents the development of a TKN(Text,Key and Name)cryptographic algorithm aimed at protecting data from replay attacks.The program employs distinct ways to encrypt plain text[P],a user-defined Key[K],and a Secret Code[N].The novelty of the TKN cryptographic algorithm is that the bit value of each text is linked to another value with the help of the proposed algorithm,and the length of the cipher text obtained is twice the length of the original text.In the scenario that an attacker executes a replay attack on the cloud server,engages in cryptanalysis,or manipulates any data,it will result in automated modification of all associated values inside the backend.This mechanism has the benefit of enhancing the detectability of replay attacks.Nevertheless,the attacker cannot access data not included in any of the papers,regardless of how effective the attack strategy is.At the end of paper,the proposed algorithm’s novelty will be compared with different algorithms,and it will be discussed how far the proposed algorithm is better than all other algorithms.
基金Project supported by the National Basic Research Program of China (Grant Nos. 2004CB318201,2011CB302300)the US National Science Foundation (Grant No. CCF-0621526)+1 种基金the National Natural Science Foundation of China (Grant No. 60703046)HUST-SRF (Grant No.2007Q021B)
文摘File semantic has proven effective in optimizing large scale distributed file system.As a consequence of the elaborate and rich I/O interfaces between upper layer applications and file systems,file system can provide useful and insightful information about semantic.Hence,file semantic mining has become an increasingly important practice in both engineering and research community.Unfortunately,it is a challenge to exploit file semantic knowledge because a variety of factors coulda ffect this information exploration process.Even worse,the challenges are exacerbated due to the intricate interdependency between these factors,and make it difficult to fully exploit the potentially important correlation among various semantic knowledges.This article proposes a file access correlation miming and evaluation reference(FARMER) model,where file is treated as a multivariate vector space,and each item within the vector corresponds a separate factor of the given file.The selection of factor depends on the application,examples of factors are file path,creator and executing program.If one particular factor occurs in both files,its value is non-zero.It is clear that the extent of inter-file relationships can be measured based on the likeness of their factor values in the semantic vectors.Benefit from this model,FARMER represents files as structured vectors of identifiers,and basic vector operations can be leveraged to quantify file correlation between two file vectors.FARMER model leverages linear regression model to estimate the strength of the relationship between file correlation and a set of influencing factors so that the "bad knowledge" can be filtered out.To demonstrate the ability of new FARMER model,FARMER is incorporated into a real large-scale object-based storage system as a case study to dynamically infer file correlations.In addition FARMER-enabled optimize service for metadata prefetching algorithm and object data layout algorithm is implemented.Experimental results show that is FARMER-enabled prefetching algorithm is shown to reduce the metadata operations latency by approximately 30%-40% when compared to a state-of-the-art metadata prefetching algorithm and a commonly used replacement policy.
文摘To better understand different users' accessing intentions, a novel clustering and supervising method based on accessing path is presented. This method divides users' interest space to express the distribution of users' interests, and directly to instruct the constructing process of web pages indexing for advanced performance.
文摘In order to improve the management strategy for personnel files in colleges and universities,simplify the complex process of file management,and improve file management security and content preservation of the files.This paper elaborates on the application of Artificial Intelligence(AI)technology in university personnel file management through theoretical analysis based on the understanding of Al technology.
基金The National Innovation Fund(No.00C262251211336)the Project of State Key Laboratory of Rail Traffic Control and Safety, Beijing Jiaotong University(No.SKL2007K001)
文摘A file transfer algorithm based on ADP (autonomous decentrafized protocol) was proposed to solve the problem that the ADS (autonomous decentralized system) middleware (NeXUS/Drink) lacks of file transfer functions for Windows. The algorithm realizes the peer-to-peer file transfer, one-to-N inquiry/multi-response file transfer and one-to-N file distribution in the same data field based on communication patterns provided by the ADP. The peer-to-peer file transfer is implemented through a peer-to-peer communication path, one-to-N inquiry/multi-response file transfer and one-to-N file distribution are implemented through multicast communication. In this algorithm, a file to be transferred is named with a GUID (global unique identification), every data packet is marked with a sequence number, and file-receiving in parallel is implemented by caching DPOs(data processing objects) and multi-thread technologies. The algorithm is applied in a simulation system of the decentralized control platform, and the test results and long time stable tanning prove the feasibility of the algorithm.
文摘EndNote is a reference management software package produced by Thomson Reuters used to manage bibliographies and references in research essays and articles. EndNote works by grouping citations into "libraries" with the file extension *.enl and a corresponding *.data folder. Aside from the library function,
基金This work was supported by National Key Research&Development Plan of China under Grant 2016QY05X1000National Natural Science Foundation of China under Grant No.61771166CERNET Innovation Project(NGII20170412).
文摘The rapid development of Internet of Things(IoT)technology has made previously unavailable data available,and applications can take advantage of device data for people to visualize,explore,and build complex analyses.As the size of the network and the number of network users continue to increase,network requests tend to aggregate on a small number of network resources,which results in uneven load on network requests.Real-time,highly reliable network file distribution technology is of great importance in the Internet of Things.This paper studies real-time and highly reliable file distribution technology for large-scale networks.In response to this topic,this paper studies the current file distribution technology,proposes a file distribution model,and proposes a corresponding load balancing method based on the file distribution model.Experiments show that the system has achieved real-time and high reliability of network transmission.
文摘Deep space communication has its own features such as long propagation delays,heavy noise,asymmetric link rates,and intermittent connectivity in space,therefore TCP/IP protocol cannot perform as well as it does in terrestrial communications.Accordingly,the Consultative Committee for Space Data Systems(CCSDS) developed CCSDS File Delivery Protocol(CFDP),which sets standards of efficient file delivery service capable of transferring files to and from mass memory located in the space segment.In CFDP,four optional acknowledge modes are supported to make the communication more reliable.In this paper,we gave a general introduction of typical communication process in CFDP and analysis of its four Negative Acknowledgement(NAK) modes on the respect of file delivery delay and times of retransmission.We found out that despite the shortest file delivery delay,immediate NAK mode suffers from the problem that frequent retransmission may probably lead to network congestion.Thus,we proposed a new mode,the error counter-based NAK mode.By simulation of the case focused on the link between a deep space probe on Mars and a ter-restrial station on Earth,we concluded that error counter-based NAK mode has successfully reduced the retransmission times at negligible cost of certain amount of file delivery delay.
基金financially supported jointly by the Monoblock Exploration from China Geological Survey (No. 12120114052201)the Independent Innovation Program for Doctoral Candidates of Central South University (No. 2015zzts069)the Foundation from Key Laboratory of Metallogenic Prediction of Nonferrous Metals and Geological Environment Monitoring, Ministry of Education
文摘1 Geological Setting The Huayuan Pb-Zn ore field in Xiangxi is located in the southeastern margin of the Yangtze block and the mid-segment of the West Hunan-West Hubei metallogenic belt.The exposed stratum are the lower
文摘ZTE Corporation announced on February 12, 2010 that by late 2009 it had applied for more than 1,700 LTE/SAE patents. ZTE is focused on advancing the LTE market with strong innovation that will benefit both consumers and operators with 4G experience and functionality.
文摘This article firstly described the related definition and technology of Ajax. Combined with Freemarker template engine with Spring and Hibernate framework technology to generate a page view, put forward the FSH the framework of a complete program, and realize upload with progress bar display.
基金Supported by Key Technology R&D Project Foundation of Sichuan Province (No.02GG006-018)
文摘So far, file access prediction models is mainly based on either the file access frequency or the historical record of the latest access. In this paper, a new file access prediction model called frequency- and recency-based successor (FRS) is presented which combines the advantages of the file frequency with the historical record. FRS model has the capability of rapid response to workload changes and can predict future events with greater accuracy than most of other prediction models. To evaluate the performance of FRS mode, the Linux kernel is modified to predict and prefetch upcoming accesses. The experiment shows that FRS can accurately predict approximately 80% of all file access events, while maintaining an immediate successor queue (ISQ) per-file which only requires regular dynamic updates.
文摘File labeling techniques have a long history in analyzing the anthological trends in computational linguistics.The situation becomes worse in the case of files downloaded into systems from the Internet.Currently,most users either have to change file names manually or leave a meaningless name of the files,which increases the time to search required files and results in redundancy and duplications of user files.Currently,no significant work is done on automated file labeling during the organization of heterogeneous user files.A few attempts have been made in topic modeling.However,one major drawback of current topic modeling approaches is better results.They rely on specific language types and domain similarity of the data.In this research,machine learning approaches have been employed to analyze and extract the information from heterogeneous corpus.A different file labeling technique has also been used to get the meaningful and`cohesive topic of the files.The results show that the proposed methodology can generate relevant and context-sensitive names for heterogeneous data files and provide additional insight into automated file labeling in operating systems.
基金This research was supported by the Universiti Sains Malaysia(USM)and the ministry of Higher Education Malaysia through Fundamental Research Grant Scheme(FRGS-Grant No:FRGS/1/2020/TK0/USM/02/1).
文摘In the Big Data era,numerous sources and environments generate massive amounts of data.This enormous amount of data necessitates specialized advanced tools and procedures that effectively evaluate the information and anticipate decisions for future changes.Hadoop is used to process this kind of data.It is known to handle vast volumes of data more efficiently than tiny amounts,which results in inefficiency in the framework.This study proposes a novel solution to the problem by applying the Enhanced Best Fit Merging algorithm(EBFM)that merges files depending on predefined parameters(type and size).Implementing this algorithm will ensure that the maximum amount of the block size and the generated file size will be in the same range.Its primary goal is to dynamically merge files with the stated criteria based on the file type to guarantee the efficacy and efficiency of the established system.This procedure takes place before the files are available for the Hadoop framework.Additionally,the files generated by the system are named with specific keywords to ensure there is no data loss(file overwrite).The proposed approach guarantees the generation of the fewest possible large files,which reduces the input/output memory burden and corresponds to the Hadoop framework’s effectiveness.The findings show that the proposed technique enhances the framework’s performance by approximately 64%while comparing all other potential performance-impairing variables.The proposed approach is implementable in any environment that uses the Hadoop framework,not limited to smart cities,real-time data analysis,etc.
文摘Working with files and the safety of information has always been relevant, especially in financial institutions where the requirements for the safety of information and security are especially important. And in today’s conditions, when an earthquake can destroy the floor of a city in an instant, or when a missile hits an office and all servers turn into scrap metal, the issue of data safety becomes especially important. Also, you can’t put the cost of the software and the convenience of working with files in last place. Especially if an office worker needs to find the necessary information on a client, a financial contract or a company’s financial product in a few seconds. Also, during the operation of computer equipment, failures are possible, and some of them can lead to partial or complete loss of information. In this paper, it is proposed to create another level of abstraction for working with the file system, which will be based on a relational database as a storage of objects and access rights to objects. Also considered are possible protocols for transferring data to other programs that work with files, these can be both small sites and the operating system itself. This article will be especially interesting for financial institutions or companies operating in the banking sector. The purpose of this article is an attempt to introduce another level of abstraction for working with files. A level that is completely abstracted from the storage medium.
基金supported by ZTE Industry⁃University⁃Institute Coopera⁃tion Funds under Grant No.HC⁃CN⁃20181128026.
文摘Byte-addressable non-volatile memory(NVM),as a new participant in the storage hierarchy,gives extremely high performance in storage,which forces changes to be made on current filesystem designs.Page cache,once a significant mechanism filling the performance gap between Dynamic Random Access Memory(DRAM)and block devices,is now a liability that heavily hinders the writing performance of NVM filesystems.Therefore state-of-the-art NVM filesystems leverage the direct access(DAX)technology to bypass the page cache entirely.However,the DRAM still provides higher bandwidth than NVM,which prevents skewed read workloads from benefiting from a higher bandwidth of the DRAM and leads to sub-optimal performance for the system.In this paper,we propose RCache,a readintensive workload-aware page cache for NVM filesystems.Different from traditional caching mechanisms where all reads go through DRAM,RCache uses a tiered page cache design,including assigning DRAM and NVM to hot and cold data separately,and reading data from both sides.To avoid copying data to DRAM in a critical path,RCache migrates data from NVM to DRAM in a background thread.Additionally,RCache manages data in DRAM in a lock-free manner for better latency and scalability.Evaluations on Intel Optane Data Center(DC)Persistent Memory Modules show that,compared with NOVA,RCache achieves 3 times higher bandwidth for read-intensive workloads and introduces little performance loss for write operations.