Due to the fact that the register files seriously affectthe performance and area of coarse-gained reconfigurablecryptographic processors, an efficient structure of thedistributed cross-domain register file is proposed...Due to the fact that the register files seriously affectthe performance and area of coarse-gained reconfigurablecryptographic processors, an efficient structure of thedistributed cross-domain register file is proposed to realize acryptographic processor with a high performance and a lowarea cost. In order to meet the demands of high performanceand high flexibility at a low area cost, a union structure withthe multi-ports access structure, i, e., a distributed cross-domain register file, is designed by analyzing the algorithmfeatures of different ciphers. Considering different algorithmrequirements of the global register files and local register files,the circuit design is realized by adopting different designparameters under TSMC ( Taiwan SemiconductorManufacturing Company) 40 nm CMOS(complementary metaloxide semiconductor) technology and compared with othersimilar works. The experimental results show that theproposed distributed cross-domain register structure caneffectively improve the performance of the unit area, of whichthe total performance of block per cycle is improved by17.79% and performance of block per cycle per area isimproved bv 117 %.展开更多
A casting process CAD is put forward to design and draw casting process. The 2D casting process CAD, most of the current systems are developed based on one certain version of the AutoCAD system. However the applicatio...A casting process CAD is put forward to design and draw casting process. The 2D casting process CAD, most of the current systems are developed based on one certain version of the AutoCAD system. However the application of these 2D casting process CAD systems in foundry enterprises are restricted because they have several deficiencies, such as being overly dependent on the AutoCAD system, and some part files based on PDF format can not be opened directly. To overcome these deficiencies, for the first time an innovative 2D casting process CAD system based on PDF and image format file has been proposed, which breaks through the traditional research and application notion of the 2D casting process CAD system based on AutoCAD. Several key technologies of this system such as coordinate transformation, CAD interactive drawing, file storage, PDF and image format files display, and image recognition technologies were described in detail. A practical 2D CAD casting process system named HZCAD2D(PDF) was developed, which is capable of designing and drawing the casting process on the part drawing based on the PDF format directly, without spending time on drawing the part produced by AutoCAD system. Finally, taking two actual castings as examples, the casting processes were drawn using this system, demonstrating that this system can significantly shorten the cycle of casting process designing.展开更多
Case-file backlogs were identified as one of the cause factors affecting the competitiveness of the forensic science laboratory (FSL). Backlogs represent case-files?that remain unprocessed or unreported within a selec...Case-file backlogs were identified as one of the cause factors affecting the competitiveness of the forensic science laboratory (FSL). Backlogs represent case-files?that remain unprocessed or unreported within a selected time interval (year, week or month) which leads to increased customer complaints, rework, cost of analysis, degradation of biological samples, etc. Case-file backlogging was quantified in three consecutive years (2014 to 2016), using the following parameters: case-files received and case-files processed, difference of which gives case-files backlogged. There was a need to define time interval for a case-file to be regarded as backlogged (that is, one week), results of which can translate into backlogged case-files per month or year. A data collection tool was established and used for three work stations (forensic chemistry, biology/DNA and toxicology laboratories). The tool includes starting and ending date for each?time interval, in which the numbers of case-files received and processed were entered followed by computing the backlogs. It was observed that, case-files reported?increased between 2014 and 2016 leading to a decrease in backlogged case-files.?The annual percentage of the case-files backlogged was highest for forensic?toxicology. The highest number of case-files backlogged was observed for forensic?chemistry, followed by forensic biology/DNA. The number of case-files?backlogged per analyst per year was highest in 2014 and dropped continuously?towards 2016, being comparably higher in forensic biology/DNA and chemistry.?Probability density functions (PDFs) and cumulative distribution functions (CDFs)?of backlogs data indicated that a large number of backlogs created in previous?weeks were eliminated. It was concluded that the effect of case-file backlogging on FSL competitiveness can be minimized by continued management effort in backlog elimination.展开更多
Hadoop distributed file system (HDFS) as a popular cloud storage platform, benefiting from its scalable, reliable and low-cost storage capability.However it is mainly designed for batch processing of large files, it’...Hadoop distributed file system (HDFS) as a popular cloud storage platform, benefiting from its scalable, reliable and low-cost storage capability.However it is mainly designed for batch processing of large files, it’s mean that small files cannot be efficiently handled by HDFS. In this paper, we propose a mechanism to store small files in HDFS. In our approach, file size need to be judged before uploading to HDFS. If the file size is less than the size of the block, all correlated small files will be merged into one single file and we will build index for each small file. Furthermore, prefetching and caching mechanism are used to improve the reading efficiency of small files. Meanwhile, for the new small files, we can execute appending operation on the basis of merged file. Contrasting to original HDFS, experimental results show that the storage efficiency of small files is improved.展开更多
File labeling techniques have a long history in analyzing the anthological trends in computational linguistics.The situation becomes worse in the case of files downloaded into systems from the Internet.Currently,most ...File labeling techniques have a long history in analyzing the anthological trends in computational linguistics.The situation becomes worse in the case of files downloaded into systems from the Internet.Currently,most users either have to change file names manually or leave a meaningless name of the files,which increases the time to search required files and results in redundancy and duplications of user files.Currently,no significant work is done on automated file labeling during the organization of heterogeneous user files.A few attempts have been made in topic modeling.However,one major drawback of current topic modeling approaches is better results.They rely on specific language types and domain similarity of the data.In this research,machine learning approaches have been employed to analyze and extract the information from heterogeneous corpus.A different file labeling technique has also been used to get the meaningful and`cohesive topic of the files.The results show that the proposed methodology can generate relevant and context-sensitive names for heterogeneous data files and provide additional insight into automated file labeling in operating systems.展开更多
In the Big Data era,numerous sources and environments generate massive amounts of data.This enormous amount of data necessitates specialized advanced tools and procedures that effectively evaluate the information and ...In the Big Data era,numerous sources and environments generate massive amounts of data.This enormous amount of data necessitates specialized advanced tools and procedures that effectively evaluate the information and anticipate decisions for future changes.Hadoop is used to process this kind of data.It is known to handle vast volumes of data more efficiently than tiny amounts,which results in inefficiency in the framework.This study proposes a novel solution to the problem by applying the Enhanced Best Fit Merging algorithm(EBFM)that merges files depending on predefined parameters(type and size).Implementing this algorithm will ensure that the maximum amount of the block size and the generated file size will be in the same range.Its primary goal is to dynamically merge files with the stated criteria based on the file type to guarantee the efficacy and efficiency of the established system.This procedure takes place before the files are available for the Hadoop framework.Additionally,the files generated by the system are named with specific keywords to ensure there is no data loss(file overwrite).The proposed approach guarantees the generation of the fewest possible large files,which reduces the input/output memory burden and corresponds to the Hadoop framework’s effectiveness.The findings show that the proposed technique enhances the framework’s performance by approximately 64%while comparing all other potential performance-impairing variables.The proposed approach is implementable in any environment that uses the Hadoop framework,not limited to smart cities,real-time data analysis,etc.展开更多
Working with files and the safety of information has always been relevant, especially in financial institutions where the requirements for the safety of information and security are especially important. And in today...Working with files and the safety of information has always been relevant, especially in financial institutions where the requirements for the safety of information and security are especially important. And in today’s conditions, when an earthquake can destroy the floor of a city in an instant, or when a missile hits an office and all servers turn into scrap metal, the issue of data safety becomes especially important. Also, you can’t put the cost of the software and the convenience of working with files in last place. Especially if an office worker needs to find the necessary information on a client, a financial contract or a company’s financial product in a few seconds. Also, during the operation of computer equipment, failures are possible, and some of them can lead to partial or complete loss of information. In this paper, it is proposed to create another level of abstraction for working with the file system, which will be based on a relational database as a storage of objects and access rights to objects. Also considered are possible protocols for transferring data to other programs that work with files, these can be both small sites and the operating system itself. This article will be especially interesting for financial institutions or companies operating in the banking sector. The purpose of this article is an attempt to introduce another level of abstraction for working with files. A level that is completely abstracted from the storage medium.展开更多
Byte-addressable non-volatile memory(NVM),as a new participant in the storage hierarchy,gives extremely high performance in storage,which forces changes to be made on current filesystem designs.Page cache,once a signi...Byte-addressable non-volatile memory(NVM),as a new participant in the storage hierarchy,gives extremely high performance in storage,which forces changes to be made on current filesystem designs.Page cache,once a significant mechanism filling the performance gap between Dynamic Random Access Memory(DRAM)and block devices,is now a liability that heavily hinders the writing performance of NVM filesystems.Therefore state-of-the-art NVM filesystems leverage the direct access(DAX)technology to bypass the page cache entirely.However,the DRAM still provides higher bandwidth than NVM,which prevents skewed read workloads from benefiting from a higher bandwidth of the DRAM and leads to sub-optimal performance for the system.In this paper,we propose RCache,a readintensive workload-aware page cache for NVM filesystems.Different from traditional caching mechanisms where all reads go through DRAM,RCache uses a tiered page cache design,including assigning DRAM and NVM to hot and cold data separately,and reading data from both sides.To avoid copying data to DRAM in a critical path,RCache migrates data from NVM to DRAM in a background thread.Additionally,RCache manages data in DRAM in a lock-free manner for better latency and scalability.Evaluations on Intel Optane Data Center(DC)Persistent Memory Modules show that,compared with NOVA,RCache achieves 3 times higher bandwidth for read-intensive workloads and introduces little performance loss for write operations.展开更多
为解决危大工程中吊装作业安全管理的问题,基于深度学习构建目标检测算法(You Only Look Once version 5,YOLOv5)网络模型,针对进入吊装作业区域内人员的防护装备进行多目标融合检测,并对吊钩在施工过程中的状态进行检测。在原始的检测...为解决危大工程中吊装作业安全管理的问题,基于深度学习构建目标检测算法(You Only Look Once version 5,YOLOv5)网络模型,针对进入吊装作业区域内人员的防护装备进行多目标融合检测,并对吊钩在施工过程中的状态进行检测。在原始的检测网络模型中引入4种注意力机制,并通过5种训练模型的结果对比分析,进而选择卷积块注意力模块(Convolutional Block Attention Module,CBAM)最优模型。优化后的检测模型对安全帽的平均识别精度达86.5%,对反光衣的平均识别精度达83.0%,对吊钩的状态识别精度达92.0%。将训练好的人员检测模型和吊钩检测模型打包成exe执行文件,应用到施工安全管理人员的中控平台,可帮助管理人员更好地判断吊装作业的工作情况,进而及时进行风险管控。展开更多
基金The National Natural Science Foundation of China(No.61176024)
文摘Due to the fact that the register files seriously affectthe performance and area of coarse-gained reconfigurablecryptographic processors, an efficient structure of thedistributed cross-domain register file is proposed to realize acryptographic processor with a high performance and a lowarea cost. In order to meet the demands of high performanceand high flexibility at a low area cost, a union structure withthe multi-ports access structure, i, e., a distributed cross-domain register file, is designed by analyzing the algorithmfeatures of different ciphers. Considering different algorithmrequirements of the global register files and local register files,the circuit design is realized by adopting different designparameters under TSMC ( Taiwan SemiconductorManufacturing Company) 40 nm CMOS(complementary metaloxide semiconductor) technology and compared with othersimilar works. The experimental results show that theproposed distributed cross-domain register structure caneffectively improve the performance of the unit area, of whichthe total performance of block per cycle is improved by17.79% and performance of block per cycle per area isimproved bv 117 %.
基金financially supported by the Program for New Century Excellent Talents in University (No.NCET-09-0396)the National Science&Technology Key Projects of Numerical Control (No.2012ZX04012-011)the Fundamental Research Funds for the Central Universities (2014-IV-016)
文摘A casting process CAD is put forward to design and draw casting process. The 2D casting process CAD, most of the current systems are developed based on one certain version of the AutoCAD system. However the application of these 2D casting process CAD systems in foundry enterprises are restricted because they have several deficiencies, such as being overly dependent on the AutoCAD system, and some part files based on PDF format can not be opened directly. To overcome these deficiencies, for the first time an innovative 2D casting process CAD system based on PDF and image format file has been proposed, which breaks through the traditional research and application notion of the 2D casting process CAD system based on AutoCAD. Several key technologies of this system such as coordinate transformation, CAD interactive drawing, file storage, PDF and image format files display, and image recognition technologies were described in detail. A practical 2D CAD casting process system named HZCAD2D(PDF) was developed, which is capable of designing and drawing the casting process on the part drawing based on the PDF format directly, without spending time on drawing the part produced by AutoCAD system. Finally, taking two actual castings as examples, the casting processes were drawn using this system, demonstrating that this system can significantly shorten the cycle of casting process designing.
文摘Case-file backlogs were identified as one of the cause factors affecting the competitiveness of the forensic science laboratory (FSL). Backlogs represent case-files?that remain unprocessed or unreported within a selected time interval (year, week or month) which leads to increased customer complaints, rework, cost of analysis, degradation of biological samples, etc. Case-file backlogging was quantified in three consecutive years (2014 to 2016), using the following parameters: case-files received and case-files processed, difference of which gives case-files backlogged. There was a need to define time interval for a case-file to be regarded as backlogged (that is, one week), results of which can translate into backlogged case-files per month or year. A data collection tool was established and used for three work stations (forensic chemistry, biology/DNA and toxicology laboratories). The tool includes starting and ending date for each?time interval, in which the numbers of case-files received and processed were entered followed by computing the backlogs. It was observed that, case-files reported?increased between 2014 and 2016 leading to a decrease in backlogged case-files.?The annual percentage of the case-files backlogged was highest for forensic?toxicology. The highest number of case-files backlogged was observed for forensic?chemistry, followed by forensic biology/DNA. The number of case-files?backlogged per analyst per year was highest in 2014 and dropped continuously?towards 2016, being comparably higher in forensic biology/DNA and chemistry.?Probability density functions (PDFs) and cumulative distribution functions (CDFs)?of backlogs data indicated that a large number of backlogs created in previous?weeks were eliminated. It was concluded that the effect of case-file backlogging on FSL competitiveness can be minimized by continued management effort in backlog elimination.
文摘Hadoop distributed file system (HDFS) as a popular cloud storage platform, benefiting from its scalable, reliable and low-cost storage capability.However it is mainly designed for batch processing of large files, it’s mean that small files cannot be efficiently handled by HDFS. In this paper, we propose a mechanism to store small files in HDFS. In our approach, file size need to be judged before uploading to HDFS. If the file size is less than the size of the block, all correlated small files will be merged into one single file and we will build index for each small file. Furthermore, prefetching and caching mechanism are used to improve the reading efficiency of small files. Meanwhile, for the new small files, we can execute appending operation on the basis of merged file. Contrasting to original HDFS, experimental results show that the storage efficiency of small files is improved.
文摘File labeling techniques have a long history in analyzing the anthological trends in computational linguistics.The situation becomes worse in the case of files downloaded into systems from the Internet.Currently,most users either have to change file names manually or leave a meaningless name of the files,which increases the time to search required files and results in redundancy and duplications of user files.Currently,no significant work is done on automated file labeling during the organization of heterogeneous user files.A few attempts have been made in topic modeling.However,one major drawback of current topic modeling approaches is better results.They rely on specific language types and domain similarity of the data.In this research,machine learning approaches have been employed to analyze and extract the information from heterogeneous corpus.A different file labeling technique has also been used to get the meaningful and`cohesive topic of the files.The results show that the proposed methodology can generate relevant and context-sensitive names for heterogeneous data files and provide additional insight into automated file labeling in operating systems.
基金This research was supported by the Universiti Sains Malaysia(USM)and the ministry of Higher Education Malaysia through Fundamental Research Grant Scheme(FRGS-Grant No:FRGS/1/2020/TK0/USM/02/1).
文摘In the Big Data era,numerous sources and environments generate massive amounts of data.This enormous amount of data necessitates specialized advanced tools and procedures that effectively evaluate the information and anticipate decisions for future changes.Hadoop is used to process this kind of data.It is known to handle vast volumes of data more efficiently than tiny amounts,which results in inefficiency in the framework.This study proposes a novel solution to the problem by applying the Enhanced Best Fit Merging algorithm(EBFM)that merges files depending on predefined parameters(type and size).Implementing this algorithm will ensure that the maximum amount of the block size and the generated file size will be in the same range.Its primary goal is to dynamically merge files with the stated criteria based on the file type to guarantee the efficacy and efficiency of the established system.This procedure takes place before the files are available for the Hadoop framework.Additionally,the files generated by the system are named with specific keywords to ensure there is no data loss(file overwrite).The proposed approach guarantees the generation of the fewest possible large files,which reduces the input/output memory burden and corresponds to the Hadoop framework’s effectiveness.The findings show that the proposed technique enhances the framework’s performance by approximately 64%while comparing all other potential performance-impairing variables.The proposed approach is implementable in any environment that uses the Hadoop framework,not limited to smart cities,real-time data analysis,etc.
文摘Working with files and the safety of information has always been relevant, especially in financial institutions where the requirements for the safety of information and security are especially important. And in today’s conditions, when an earthquake can destroy the floor of a city in an instant, or when a missile hits an office and all servers turn into scrap metal, the issue of data safety becomes especially important. Also, you can’t put the cost of the software and the convenience of working with files in last place. Especially if an office worker needs to find the necessary information on a client, a financial contract or a company’s financial product in a few seconds. Also, during the operation of computer equipment, failures are possible, and some of them can lead to partial or complete loss of information. In this paper, it is proposed to create another level of abstraction for working with the file system, which will be based on a relational database as a storage of objects and access rights to objects. Also considered are possible protocols for transferring data to other programs that work with files, these can be both small sites and the operating system itself. This article will be especially interesting for financial institutions or companies operating in the banking sector. The purpose of this article is an attempt to introduce another level of abstraction for working with files. A level that is completely abstracted from the storage medium.
基金supported by ZTE Industry⁃University⁃Institute Coopera⁃tion Funds under Grant No.HC⁃CN⁃20181128026.
文摘Byte-addressable non-volatile memory(NVM),as a new participant in the storage hierarchy,gives extremely high performance in storage,which forces changes to be made on current filesystem designs.Page cache,once a significant mechanism filling the performance gap between Dynamic Random Access Memory(DRAM)and block devices,is now a liability that heavily hinders the writing performance of NVM filesystems.Therefore state-of-the-art NVM filesystems leverage the direct access(DAX)technology to bypass the page cache entirely.However,the DRAM still provides higher bandwidth than NVM,which prevents skewed read workloads from benefiting from a higher bandwidth of the DRAM and leads to sub-optimal performance for the system.In this paper,we propose RCache,a readintensive workload-aware page cache for NVM filesystems.Different from traditional caching mechanisms where all reads go through DRAM,RCache uses a tiered page cache design,including assigning DRAM and NVM to hot and cold data separately,and reading data from both sides.To avoid copying data to DRAM in a critical path,RCache migrates data from NVM to DRAM in a background thread.Additionally,RCache manages data in DRAM in a lock-free manner for better latency and scalability.Evaluations on Intel Optane Data Center(DC)Persistent Memory Modules show that,compared with NOVA,RCache achieves 3 times higher bandwidth for read-intensive workloads and introduces little performance loss for write operations.
文摘为解决危大工程中吊装作业安全管理的问题,基于深度学习构建目标检测算法(You Only Look Once version 5,YOLOv5)网络模型,针对进入吊装作业区域内人员的防护装备进行多目标融合检测,并对吊钩在施工过程中的状态进行检测。在原始的检测网络模型中引入4种注意力机制,并通过5种训练模型的结果对比分析,进而选择卷积块注意力模块(Convolutional Block Attention Module,CBAM)最优模型。优化后的检测模型对安全帽的平均识别精度达86.5%,对反光衣的平均识别精度达83.0%,对吊钩的状态识别精度达92.0%。将训练好的人员检测模型和吊钩检测模型打包成exe执行文件,应用到施工安全管理人员的中控平台,可帮助管理人员更好地判断吊装作业的工作情况,进而及时进行风险管控。