期刊文献+
共找到41篇文章
< 1 2 3 >
每页显示 20 50 100
A review of high performance computing applications in high-speed rail systems
1
作者 Shenyuan Ren Yidong Li 《High-Speed Railway》 2023年第2期92-96,共5页
Further improving the railway innovation capacity and technological strength is the important goal of the 14th Five-Year Plan for railway scientific and technological innovation.It includes promoting the deep integrat... Further improving the railway innovation capacity and technological strength is the important goal of the 14th Five-Year Plan for railway scientific and technological innovation.It includes promoting the deep integration of cutting-edge technologies with the railway systems,strengthening the research and application of intelligent railway technologies,applying green computing technologies and advancing the collaborative sharing of transportation big data.The high-speed rail system tasks need to process huge amounts of data and heavy workload with the requirement of ultra-fast response.Therefore,it is of great necessity to promote computation efficiency by applying High Performance Computing(HPC)to high-speed rail systems.The HPC technique is a great solution for improving the performance,efficiency,and safety of high-speed rail systems.In this review,we introduce and analyze the application research of high performance computing technology in the field of highspeed railways.These HPC applications are cataloged into four broad categories,namely:fault diagnosis,network and communication,management system,and simulations.Moreover,challenges and issues to be addressed are discussed and further directions are suggested. 展开更多
关键词 high performance computing high-speed rail
下载PDF
The Changing Face of High Performance Computing in the United States 被引量:2
2
作者 Ann Haves(Advanced Computing Laboratory Los Alamos National Laboratory Los Alamos. NM 87545, USA) 《Wuhan University Journal of Natural Sciences》 CAS 1996年第Z1期309-311,共3页
TheChangingFaceofHighPerformanceComputingintheUnitedStatesAnnHaves(AdvancedComputingLaboratoryLosAlamosNatio... TheChangingFaceofHighPerformanceComputingintheUnitedStatesAnnHaves(AdvancedComputingLaboratoryLosAlamosNationalLaboratoryLosA... 展开更多
关键词 The Changing Face of high performance computing in the United States
下载PDF
The future is frozen:cryogenic CMOS for high-performance computing
3
作者 R.Saligram A.Raychowdhury Suman Datta 《Chip》 EI 2024年第1期43-54,共12页
Low temperature complementary metal oxide semiconductor(CMOS)or cryogenic CMOS is a promising avenue for the continuation of Moore’s law while serving the needs of high performance computing.With temperature as a con... Low temperature complementary metal oxide semiconductor(CMOS)or cryogenic CMOS is a promising avenue for the continuation of Moore’s law while serving the needs of high performance computing.With temperature as a control“knob”to steepen the subthreshold slope behavior of CMOS devices,the supply voltage of operation can be reduced with no impact on operating speed.With the optimal threshold voltage engineering,the device ON current can be further enhanced,translating to higher performance.In this article,the experimentally calibrated data was adopted to tune the threshold voltage and investigated the power performance area of cryogenic CMOS at device,circuit and system level.We also presented results from measurement and analysis of functional memory chips fabricated in 28 nm bulk CMOS and 22 nm fully depleted silicon on insulator(FDSOI)operating at cryogenic temperature.Finally,the challenges and opportunities in the further development and deployment of such systems were discussed. 展开更多
关键词 Cryogenic CMOS Design technology co-optimization high performance computing Parameter variation Threshold voltage engineering Cryogenic Memories Interconnects
原文传递
HIGH PERFORMANCE SPARSE SOLVER FOR UNSYMMETRICAL LINEAR EQUATIONS WITH OUT-OF-CORE STRATEGIES AND ITS APPLICATION ON MESHLESS METHODS 被引量:1
4
作者 苑维然 陈璞 刘凯欣 《Applied Mathematics and Mechanics(English Edition)》 SCIE EI 2006年第10期1339-1348,共10页
A new direct method for solving unsymmetrical sparse linear systems(USLS) arising from meshless methods was introduced. Computation of certain meshless methods such as meshless local Petrov-Galerkin (MLPG) method ... A new direct method for solving unsymmetrical sparse linear systems(USLS) arising from meshless methods was introduced. Computation of certain meshless methods such as meshless local Petrov-Galerkin (MLPG) method need to solve large USLS. The proposed solution method for unsymmetrical case performs factorization processes symmetrically on the upper and lower triangular portion of matrix, which differs from previous work based on general unsymmetrical process, and attains higher performance. It is shown that the solution algorithm for USLS can be simply derived from the existing approaches for the symmetrical case. The new matrix factorization algorithm in our method can be implemented easily by modifying a standard JKI symmetrical matrix factorization code. Multi-blocked out-of-core strategies were also developed to expand the solution scale. The approach convincingly increases the speed of the solution process, which is demonstrated with the numerical tests. 展开更多
关键词 sparse matrices linear equations meshless methods high performance computation
下载PDF
A NEW HIGH PERFORMANCE SPARSE STATIC SOLVER IN FINITE ELEMENT ANALYSIS WITH LOOP-UNROLLING 被引量:1
5
作者 Chen Pu Sun Shuli 《Acta Mechanica Solida Sinica》 SCIE EI 2005年第3期248-255,共8页
In the previous papers, a high performance sparse static solver with two-level unrolling based on a cell-sparse storage scheme was reported. Although the solver reaches quite a high efficiency for a big percentage of ... In the previous papers, a high performance sparse static solver with two-level unrolling based on a cell-sparse storage scheme was reported. Although the solver reaches quite a high efficiency for a big percentage of finite element analysis benchmark tests, the MFLOPS (million floating operations per second) of LDL^T factorization of benchmark tests vary on a Dell Pentium IV 850 MHz machine from 100 to 456 depending on the average size of the super-equations, i.e., on the average depth of unrolling. In this paper, a new sparse static solver with two-level unrolling that employs the concept of master-equations and searches for an appropriate depths of unrolling is proposed. The new solver provides higher MFLOPS for LDL^T factorization of benchmark tests, and therefore speeds up the solution process. 展开更多
关键词 high performance computing sparse matrix finite element analysis
下载PDF
Parallel Optical Interconnect Technology: Combination of Higher Performance and Lower Energy Consumption
6
作者 Qiao Yaojun Gu Rentao Ji Yuefeng 《China Communications》 SCIE CSCD 2010年第3期99-106,共8页
This paper analyzes the physical potential, computing performance benefi t and power consumption of optical interconnects. Compared with electrical interconnections, optical ones show undoubted advantages based on phy... This paper analyzes the physical potential, computing performance benefi t and power consumption of optical interconnects. Compared with electrical interconnections, optical ones show undoubted advantages based on physical factor analysis. At the same time, since the recent developments drive us to think about whether these optical interconnect technologies with higher bandwidth but higher cost are worthy to be deployed, the computing performance comparison is performed. To meet the increasing demand of large-scale parallel or multi-processor computing tasks, an analytic method to evaluate parallel computing performance ofinterconnect systems is proposed in this paper. Both bandwidth-limit model and full-bandwidth model are under our investigation. Speedup and effi ciency are selected to represent the parallel performance of an interconnect system. Deploying the proposed models, we depict the performance gap between the optical and electrically interconnected systems. Another investigation on power consumption of commercial products showed that if the parallel interconnections are deployed, the unit power consumption will be reduced. Therefore, from the analysis of computing influence and power dissipation, we found that parallel optical interconnect is valuable combination of high performance and low energy consumption. Considering the possible data center under construction, huge power could be saved if parallel optical interconnects technologies are used. 展开更多
关键词 optical interconnects high performance computing power dissipation
下载PDF
Intrusion Detection Using Federated Learning for Computing
7
作者 R.S.Aashmi T.Jaya 《Computer Systems Science & Engineering》 SCIE EI 2023年第5期1295-1308,共14页
The integration of clusters,grids,clouds,edges and other computing platforms result in contemporary technology of jungle computing.This novel technique has the aptitude to tackle high performance computation systems a... The integration of clusters,grids,clouds,edges and other computing platforms result in contemporary technology of jungle computing.This novel technique has the aptitude to tackle high performance computation systems and it manages the usage of all computing platforms at a time.Federated learning is a collaborative machine learning approach without centralized training data.The proposed system effectively detects the intrusion attack without human intervention and subsequently detects anomalous deviations in device communication behavior,potentially caused by malicious adversaries and it can emerge with new and unknown attacks.The main objective is to learn overall behavior of an intruder while performing attacks to the assumed target service.Moreover,the updated system model is send to the centralized server in jungle computing,to detect their pattern.Federated learning greatly helps the machine to study the type of attack from each device and this technique paves a way to complete dominion over all malicious behaviors.In our proposed work,we have implemented an intrusion detection system that has high accuracy,low False Positive Rate(FPR)scalable,and versatile for the jungle computing environment.The execution time taken to complete a round is less than two seconds,with an accuracy rate of 96%. 展开更多
关键词 Jungle computing high performance computation federated learning false positive rate intrusion detection system(IDS)
下载PDF
Parallel Image Processing: Taking Grayscale Conversion Using OpenMP as an Example
8
作者 Bayan AlHumaidan Shahad Alghofaily +2 位作者 Maitha Al Qhahtani Sara Oudah Naya Nagy 《Journal of Computer and Communications》 2024年第2期1-10,共10页
In recent years, the widespread adoption of parallel computing, especially in multi-core processors and high-performance computing environments, ushered in a new era of efficiency and speed. This trend was particularl... In recent years, the widespread adoption of parallel computing, especially in multi-core processors and high-performance computing environments, ushered in a new era of efficiency and speed. This trend was particularly noteworthy in the field of image processing, which witnessed significant advancements. This parallel computing project explored the field of parallel image processing, with a focus on the grayscale conversion of colorful images. Our approach involved integrating OpenMP into our framework for parallelization to execute a critical image processing task: grayscale conversion. By using OpenMP, we strategically enhanced the overall performance of the conversion process by distributing the workload across multiple threads. The primary objectives of our project revolved around optimizing computation time and improving overall efficiency, particularly in the task of grayscale conversion of colorful images. Utilizing OpenMP for concurrent processing across multiple cores significantly reduced execution times through the effective distribution of tasks among these cores. The speedup values for various image sizes highlighted the efficacy of parallel processing, especially for large images. However, a detailed examination revealed a potential decline in parallelization efficiency with an increasing number of cores. This underscored the importance of a carefully optimized parallelization strategy, considering factors like load balancing and minimizing communication overhead. Despite challenges, the overall scalability and efficiency achieved with parallel image processing underscored OpenMP’s effectiveness in accelerating image manipulation tasks. 展开更多
关键词 Parallel computing Image Processing OPENMP Parallel Programming high performance computing GPU (Graphic Processing Unit)
下载PDF
High Throughput Scheduling Algorithms for Input Queued Packet Switches 被引量:2
9
作者 R.Chithra Devi D.Jemi Florinabel Narayanan Prasanth 《Computers, Materials & Continua》 SCIE EI 2022年第1期1527-1540,共14页
The high-performance computing paradigm needs high-speed switching fabrics to meet the heavy traffic generated by their applications.These switching fabrics are efficiently driven by the deployed scheduling algorithms... The high-performance computing paradigm needs high-speed switching fabrics to meet the heavy traffic generated by their applications.These switching fabrics are efficiently driven by the deployed scheduling algorithms.In this paper,we proposed two scheduling algorithms for input queued switches whose operations are based on ranking procedures.At first,we proposed a Simple 2-Bit(S2B)scheme which uses binary ranking procedure and queue size for scheduling the packets.Here,the Virtual Output Queue(VOQ)set with maximum number of empty queues receives higher rank than other VOQ’s.Through simulation,we showed S2B has better throughput performance than Highest Ranking First(HRF)arbitration under uniform,and non-uniform traffic patterns.To further improve the throughput-delay performance,an Enhanced 2-Bit(E2B)approach is proposed.This approach adopts an integer representation for rank,which is the number of empty queues in a VOQ set.The simulation result shows E2B outperforms S2B and HRF scheduling algorithms with maximum throughput-delay performance.Furthermore,the algorithms are simulated under hotspot traffic and E2B proves to be more efficient. 展开更多
关键词 Crossbar switch input queued switch virtual output queue scheduling algorithm high performance computing
下载PDF
Representing Increasing Virtual Machine Security Strategy in Cloud Computing Computations 被引量:1
10
作者 Mohammad Shirzadi 《Electrical Science & Engineering》 2021年第2期7-16,共10页
This paper proposes algorithm for Increasing Virtual Machine Security Strategy in Cloud Computing computations.Imbalance between load and energy has been one of the disadvantages of old methods in providing server and... This paper proposes algorithm for Increasing Virtual Machine Security Strategy in Cloud Computing computations.Imbalance between load and energy has been one of the disadvantages of old methods in providing server and hosting,so that if two virtual severs be active on a host and energy load be more on a host,it would allocated the energy of other hosts(virtual host)to itself to stay steady and this option usually leads to hardware overflow errors and users dissatisfaction.This problem has been removed in methods based on cloud processing but not perfectly,therefore,providing an algorithm not only will implement a suitable security background but also it will suitably divide energy consumption and load balancing among virtual severs.The proposed algorithm is compared with several previously proposed Security Strategy including SC-PSSF,PSSF and DEEAC.Comparisons show that the proposed method offers high performance computing,efficiency and consumes lower energy in the network. 展开更多
关键词 Cloud computing high performance computing AUTOMATION Security SERVER
下载PDF
Multi Granularity Page Size Support for Linux and the Performance Evaluation
11
作者 Naohiko Shimizu School of Engineering, Tokai University,1117 Kitakaname Hiratsuka shi, Kanagawa 259 1292, Japan 《Wuhan University Journal of Natural Sciences》 CAS 2001年第Z1期347-350,共4页
Today the PC class machines are quite popular for HPC area, especially on the problemsthat require the good cost/performance ratios. One of the drawback of these machines is the poormemory throughput performance. And ... Today the PC class machines are quite popular for HPC area, especially on the problemsthat require the good cost/performance ratios. One of the drawback of these machines is the poormemory throughput performance. And one of the reasons of the poor performance is depend on the lack of the mapping capability of the TLB which is a buffer to accelerate the virtual memory access. In this report, I present that the mapping capability and the performance can be improved with the multi granularity TLB feature that some processors have. And I also present that the new TLB handling routine can be incorporated into the demand paging system of Linux. 展开更多
关键词 translation look aride buffer LINUX high performance computing performance evaluation
下载PDF
Towards Auction-Based HPC Computing in the Cloud
12
作者 Moussa Taifi Justin Y. Shi Abdallah Khreishah 《Computer Technology and Application》 2012年第7期499-509,共11页
Cloud computing is expanding widely in the world of IT infrastructure. This is due partly to the cost-saving effect of economies of scale. Fair market conditions can in theory provide a healthy environment to reflect ... Cloud computing is expanding widely in the world of IT infrastructure. This is due partly to the cost-saving effect of economies of scale. Fair market conditions can in theory provide a healthy environment to reflect the most reasonable costs of computations. While fixed cloud pricing provides an attractive low entry barrier for compute-intensive applications, both the consumer and supplier of computing resources can see high efficiency for their investments by participating in auction-based exchanges. There are huge incentives for the cloud provider to offer auctioned resources. However, from the consumer perspective, using these resources is a sparsely discussed challenge. This paper reports a methodology and framework designed to address the challenges of using HPC (High Performance Computing) applications on auction-based cloud clusters. The authors focus on HPC applications and describe a method for determining bid-aware checkpointing intervals. They extend a theoretical model for determining checkpoint intervals using statistical analysis of pricing histories. Also the latest developments in the SpotHPC framework are introduced which aim at facilitating the managed execution of real MPI applications on auction-based cloud environments. The authors use their model to simulate a set of algorithms with different computing and communication densities. The results show the complex interactions between optimal bidding strategies and parallel applications performance. 展开更多
关键词 Auction-based cloud computing fault tolerance cloud HPC high performance computing
下载PDF
High performance computing of DGDFT for tens of thousands of atoms using millions of cores on Sunway TaihuLight 被引量:4
13
作者 Wei Hu Xinming Qin +9 位作者 Qingcai Jiang Junshi Chen Hong An Weile Jia Fang Li Xin Liu Dexun Chen Fangfang Liu Yuwen Zhao Jinlong Yang 《Science Bulletin》 SCIE EI CSCD 2021年第2期111-119,M0003,共10页
High performance computing(HPC)is a powerful tool to accelerate the Kohn–Sham density functional theory(KS-DFT)calculations on modern heterogeneous supercomputers.Here,we describe a massively parallel implementation ... High performance computing(HPC)is a powerful tool to accelerate the Kohn–Sham density functional theory(KS-DFT)calculations on modern heterogeneous supercomputers.Here,we describe a massively parallel implementation of discontinuous Galerkin density functional theory(DGDFT)method on the Sunway Taihu Light supercomputer.The DGDFT method uses the adaptive local basis(ALB)functions generated on-the-fly during the self-consistent field(SCF)iteration to solve the KS equations with high precision comparable to plane-wave basis set.In particular,the DGDFT method adopts a two-level parallelization strategy that deals with various types of data distribution,task scheduling,and data communication schemes,and combines with the master–slave multi-thread heterogeneous parallelism of SW26010 processor,resulting in large-scale HPC KS-DFT calculations on the Sunway Taihu Light supercomputer.We show that the DGDFT method can scale up to 8,519,680 processing cores(131,072 core groups)on the Sunway Taihu Light supercomputer for studying the electronic structures of twodimensional(2 D)metallic graphene systems that contain tens of thousands of carbon atoms. 展开更多
关键词 Density functional theory Tens of thousands of atoms high performance computing Sunway TaihuLight
原文传递
ONFS: a hierarchical hybrid file system based on memory, SSD, and HDD for high performance computers 被引量:1
14
作者 Xin LIU Yu-tong LU +3 位作者 Jie YU Peng-fei WANG Jie-ting WU Ying LU 《Frontiers of Information Technology & Electronic Engineering》 SCIE EI CSCD 2017年第12期1940-1971,共32页
With supercomputers developing towards exascale, the number of compute cores increases dramatically, making more complex and larger-scale applications possible. The input/output (I/O) requirements of large-scale app... With supercomputers developing towards exascale, the number of compute cores increases dramatically, making more complex and larger-scale applications possible. The input/output (I/O) requirements of large-scale applications, workflow applications, and their checkpointing include substantial bandwidth and an extremely low latency, posing a serious challenge to high performance computing (HPC) storage systems. Current hard disk drive (HDD) based underlying storage systems are becoming more and more incompetent to meet the requirements of next-generation exascale supercomputers. To rise to the challenge, we propose a hierarchical hybrid storage system, on-line and near-line file system (ONFS). It leverages dynamic random access memory (DRAM) and solid state drive (SSD) in compute nodes, and HDD in storage servers to build a three-level storage system in a unified namespace. It supports portable operating system interface (POSIX) semantics, and provides high bandwidth, low latency, and huge storage capacity. In this paper, we present the technical details on distributed metadata management, the strategy of memory borrow and return, data consistency, parallel access control, and mechanisms guiding downward and upward migration in ONFS. We implement an ONFS prototype on the TH-1A supercomputer, and conduct experiments to test its I/O performance and scalability. The results show that the bandwidths of single-thread and multi-thread 'read'/'write' are 6-fold and 5-fold better than HDD-based Lustre, respectively. The I/O bandwidth of data-intensive applications in ONFS can be 6.35 timcs that in Lustre. 展开更多
关键词 high performance computing Hierarchical hybrid storage system Distributed metadata management Data migration
原文传递
Self-deployed execution environment for high performance computing
15
作者 Mingtian SHAO Kai LU Wenzhe ZHANG 《Frontiers of Information Technology & Electronic Engineering》 SCIE EI CSCD 2022年第6期845-857,共13页
Traditional high performance computing(HPC)systems provide a standard preset environment to support scientific computation.However,HPC development needs to provide support for more and more diverse applications,such a... Traditional high performance computing(HPC)systems provide a standard preset environment to support scientific computation.However,HPC development needs to provide support for more and more diverse applications,such as artificial intelligence and big data.The standard preset environment can no longer meet these diverse requirements.If users still run these emerging applications on HPC systems,they need to manually maintain the specific dependencies(libraries,environment variables,and so on)of their applications.This increases the development and deployment burden for users.Moreover,the multi-user mode brings about privacy problems among users.Containers like Docker and Singularity can encapsulate the job’s execution environment,but in a highly customized HPC system,cross-environment application deployment of Docker and Singularity is limited.The introduction of container images also imposes a maintenance burden on system administrators.Facing the above-mentioned problems,in this paper we propose a self-deployed execution environment(SDEE)for HPC.SDEE combines the advantages of traditional virtualization and modern containers.SDEE provides an isolated and customizable environment(similar to a virtual machine)to the user.The user is the root user in this environment.The user develops and debugs the application and deploys its special dependencies in this environment.Then the user can load the job to compute nodes directly through the traditional HPC job management system.The job and its dependencies are analyzed,packaged,deployed,and executed automatically.This process enables transparent and rapid job deployment,which not only reduces the burden on users,but also protects user privacy.Experiments show that the overhead introduced by SDEE is negligible and lower than those of both Docker and Singularity. 展开更多
关键词 Execution environment high performance computing LIGHT-WEIGHT ISOLATION OVERLAY
原文传递
Monte Carlo performance study of virtual high performance computing cluster over cloud
16
作者 Libing Zhu Ze Xi +5 位作者 Peng Cong Gongyi Yu Yuan Liu Xincheng Xiang Wei Xu Xiangang Wang 《Radiation Medicine and Protection》 2022年第3期108-114,共7页
Objective:As a high computation cost discipline,nuclear science and engineering still relies heavily on traditional high performance computing(HPC)clusters.However,the usage of traditional HPC for nuclear science and ... Objective:As a high computation cost discipline,nuclear science and engineering still relies heavily on traditional high performance computing(HPC)clusters.However,the usage of traditional HPC for nuclear science and engineering has been limited due to the poor flexibility,the software compatibility and the poor user interfaces.Virtualized/virtual HPC(vHPC)can mimic an HPC by using a cloud computing platform.In this work,we designed and developed a vHPC system for employment in nuclear engineering.Methods:The system is tested using the computation of the numberπby Monte Carlo and an X-ray digital imaging system simulation.The performance of the vHPC system is compared with that of the traditional HPCs.Results:As the number of the simulated particles increases,the virtual cluster computing time grows propor-tionally.The time used for the simulation of the X-ray imaging was about 21.1 h over a 12 kernels virtual server.Experimental results show that the performance of virtual cluster computing and the actual physical machine is almost the same.Conclusions:From these tests,it is concluded that vHPC is a good alternative for employing in nuclear engineering.The proposed vHPC in this paper will make HPC flexible and easy to deploy. 展开更多
关键词 Cloud computing VIRTUALIZATION high performance computing(HPC) Virtual HPC Monte Carlo
原文传递
Smart Cities in Europe and the ALMA Logistics Project 被引量:2
17
作者 Didier El Baz Julien Bourgeois 《ZTE Communications》 2015年第4期10-15,共6页
In this paper, a brief survey of smart citiy projects in Europe is presented. This survey shows the extent of transport and logistics in smart cities. We concentrate on a smart city project we have been working on tha... In this paper, a brief survey of smart citiy projects in Europe is presented. This survey shows the extent of transport and logistics in smart cities. We concentrate on a smart city project we have been working on that is related to A Logistic Mobile Application (ALMA). The application is based on Internet of Things and combines a communication infrastructure and a High Performance Computing infrastructure in order to deliver mobile logistic services with high quality of service and adaptation to the dynamic nature of logistic operations. 展开更多
关键词 smart cities Internet of Things LOGISTICS combinatorial optimization high performance computing
下载PDF
SCStore: Managing Scientific Computing Packages for Hybrid System with Containers 被引量:5
18
作者 Wusheng Zhang Jiao Lin +2 位作者 Weiping Xu Haohuan Fu Guangwen Yang 《Tsinghua Science and Technology》 SCIE EI CAS CSCD 2017年第6期675-681,共7页
Managing software packages in a scientific computing environment is a challenging task, especially in the case of heterogeneous systems. It is error prone when installing and updating software packages in a sophistica... Managing software packages in a scientific computing environment is a challenging task, especially in the case of heterogeneous systems. It is error prone when installing and updating software packages in a sophisticated computing environment. Testing and performance evaluation in an on-the-fly manner is also a troublesome task for a production system. In this paper, we discuss a package management scheme based on containers. The newly developed method can ease the maintenance complexity and reduce human mistakes. We can benefit from the self-containing and isolation features of container technologies for maintaining the software packages among intricately connected clusters. By deploying the Super Computing application Strore(SCStore) over the WAN connected world-largest clusters, it proved that it can greatly reduce the effort for maintaining the consistency of software environment and bring benefit to achieve automation. 展开更多
关键词 high performance computing package management container hybrid system
原文传递
High Order Accurate Direct Arbitrary-Lagrangian-Eulerian ADER-MOOD Finite Volume Schemes for Non-Conservative Hyperbolic Systems with Stiff Source Terms 被引量:1
19
作者 Walter Boscheri Raphael Loubere 《Communications in Computational Physics》 SCIE 2017年第1期271-312,共42页
In this paper we present a 2D/3D high order accurate finite volume scheme in the context of direct Arbitrary-Lagrangian-Eulerian algorithms for general hyperbolic systems of partial differential equations with non-con... In this paper we present a 2D/3D high order accurate finite volume scheme in the context of direct Arbitrary-Lagrangian-Eulerian algorithms for general hyperbolic systems of partial differential equations with non-conservative products and stiff source terms.This scheme is constructed with a single stencil polynomial reconstruction operator,a one-step space-time ADER integration which is suitably designed for dealing even with stiff sources,a nodal solver with relaxation to determine the mesh motion,a path-conservative integration technique for the treatment of non-conservative products and an a posteriori stabilization procedure derived from the so-called Multidimensional Optimal Order Detection(MOOD)paradigm.In this work we consider the seven equation Baer-Nunziato model of compressible multi-phase flows as a representative model involving non-conservative products as well as relaxation source terms which are allowed to become stiff.The new scheme is validated against a set of test cases on 2D/3D unstructured moving meshes on parallel machines and the high order of accuracy achieved by the method is demonstrated by performing a numerical convergence study.Classical Riemann problems and explosion problems with exact solutions are simulated in 2D and 3D.The overall numerical code is also profiled to provide an estimate of the computational cost required by each component of the whole algorithm. 展开更多
关键词 Direct Arbitrary-Lagrangian-Eulerian a posteriori MOOD stabilization Baer-Nunziato model stiff source terms non-conservative products unstructured mesh ADER high order of accuracy in space and time high performance computing(HPC) hyperbolic conservation laws
原文传递
A spatial decomposition approach for accelerating buffer analysis of vector data 被引量:1
20
作者 Li Xiaohua Guo Mingqiang Qi Xinhong 《High Technology Letters》 EI CAS 2020年第4期455-459,共5页
Parallel vector buffer analysis approaches can be classified into 2 types:algorithm-oriented parallel strategy and the data-oriented parallel strategy.These methods do not take its applicability on the existing geogra... Parallel vector buffer analysis approaches can be classified into 2 types:algorithm-oriented parallel strategy and the data-oriented parallel strategy.These methods do not take its applicability on the existing geographic information systems(GIS)platforms into consideration.In order to address the problem,a spatial decomposition approach for accelerating buffer analysis of vector data is proposed.The relationship between the number of vertices of each feature and the buffer analysis computing time is analyzed to generate computational intensity transformation functions(CITFs).Then,computational intensity grids(CIGs)of polyline and polygon are constructed based on the relative CITFs.Using the corresponding CIGs,a spatial decomposition method for parallel buffer analysis is developed.Based on the computational intensity of the features and the sub-domains generated in the decomposition,the features are averagely assigned within the sub-domains into parallel buffer analysis tasks for load balance.Compared with typical regular domain decomposition methods,the new approach accomplishes greater balanced decomposition of computational intensity for parallel buffer analysis and achieves near-linear speedups. 展开更多
关键词 high performance spatial computing buffer analysis parallel computing load balancing vector data
下载PDF
上一页 1 2 3 下一页 到第
使用帮助 返回顶部