Brain science accelerates the study of intelligence and behavior,contributes fundamental insights into human cognition,and offers prospective treatments for brain disease.Faced with the challenges posed by imaging tec...Brain science accelerates the study of intelligence and behavior,contributes fundamental insights into human cognition,and offers prospective treatments for brain disease.Faced with the challenges posed by imaging technologies and deep learning computational models,big data and high-performance computing(HPC)play essential roles in studying brain function,brain diseases,and large-scale brain models or connectomes.We review the driving forces behind big data and HPC methods applied to brain science,including deep learning,powerful data analysis capabilities,and computational performance solutions,each of which can be used to improve diagnostic accuracy and research output.This work reinforces predictions that big data and HPC will continue to improve brain science by making ultrahigh-performance analysis possible,by improving data standardization and sharing,and by providing new neuromorphic insights.展开更多
Cloud computing has been considered as the next-generation computing platform with the potential to address the data and computing challenges in geosciences.However,only a limited number of geoscientists have been ada...Cloud computing has been considered as the next-generation computing platform with the potential to address the data and computing challenges in geosciences.However,only a limited number of geoscientists have been adapting this platform for their scientific research mainly due to two barriers:1)selecting an appropriate cloud platform for a specific application could be challenging,as various cloud services are available and 2)existing general cloud platforms are not designed to support geoscience applications,algorithms and models.To tackle such barriers,this research aims to design a hybrid cloud computing(HCC)platform that can utilize and integrate the computing resources across different organizations to build a unified geospatial cloud computing platform.This platform can manage different types of underlying cloud infrastructure(e.g.,private or public clouds),and enables geoscientists to test and leverage the cloud capabilities through a web interface.Additionally,the platform also provides different geospatial cloud services,such as workflow as a service,on the top of common cloud services(e.g.,infrastructure as a service)provided by general cloud platforms.Therefore,geoscientists can easily create a model workflow by recruiting the needed models for a geospatial application or task on the fly.A HCC prototype is developed and dust storm simulation is used to demonstrate the capability and feasibility of such platform in facilitating geosciences by leveraging across-organization computing and model resources.展开更多
With the convergence of high-performance computing(HPC),big data and artificial intelligence(AI),the HPC community is pushing for"triple use"systems to expedite scientific discoveries.However,supporting thes...With the convergence of high-performance computing(HPC),big data and artificial intelligence(AI),the HPC community is pushing for"triple use"systems to expedite scientific discoveries.However,supporting these converged applications on HPC systems presents formidable challenges in terms of storage and data management due to the explosive growth of scientific data and the fundamental differences in I/O characteristics among HPC,big data and AI workloads.In this paper,we discuss the driving force behind the converging trend,highlight three data management challenges,and summarize our efforts in addressing these data management challenges on a typical HPC system at the parallel file system,data management middleware,and user application levels.As HPC systems are approaching the border of exascale computing,this paper sheds light on how to enable application-driven data management as a preliminary step toward the deep convergence of exascale computing ecosystems,big data,and AI.展开更多
With the significant advancement in emerging processor, memory, and networking technologies, exascale systems will become available in the next few years (2020 2022). As the exascale systems begin to be deployed and...With the significant advancement in emerging processor, memory, and networking technologies, exascale systems will become available in the next few years (2020 2022). As the exascale systems begin to be deployed and used, there will be a continuous demand to run next-generation applications with finer granularity, finer time-steps, and increased data sizes. Based on historical trends, next-generation applications will require post-exascale systems during 2025-2035. In this study, we focus on the networking and communication challenges for post-exascale systems. Firstly, we present an envisioned architecture for post-exascale systems. Secondly, the challenges are summarized from different perspectives: heterogeneous networking technologies, high-performance eonmmnication and synchronization protocols, integrated support with accelerators and field-programmable gate arrays, fault-tolerance and quality-of-service support, energy-aware communication schemes and protocols, software- defined networking, and scalable communication protocols with heterogeneous memory and storage. Thirdly, we present the challenges in designing efficient programming model support for high-performance computing, big data, and deep learning on these systems. Finally, we emphasize the critical need for co-designing runtime with upper layers on these systems to achieve the maximum performance and scalability.展开更多
基金supported by the National Natural Science Foundation of China(Grant No.31771466)the National Key R&D Program of China(Grant Nos.2018YFB0203903,2016YFC0503607,and 2016YFB0200300)+3 种基金the Transformation Project in Scientific and Technological Achievements of Qinghai,China(Grant No.2016-SF-127)the Special Project of Informatization of Chinese Academy of Sciences,China(Grant No.XXH13504-08)the Strategic Pilot Science and Technology Project of Chinese Academy of Sciences,China(Grant No.XDA12010000)the 100-Talents Program of Chinese Academy of Sciences,China(awarded to BN)
文摘Brain science accelerates the study of intelligence and behavior,contributes fundamental insights into human cognition,and offers prospective treatments for brain disease.Faced with the challenges posed by imaging technologies and deep learning computational models,big data and high-performance computing(HPC)play essential roles in studying brain function,brain diseases,and large-scale brain models or connectomes.We review the driving forces behind big data and HPC methods applied to brain science,including deep learning,powerful data analysis capabilities,and computational performance solutions,each of which can be used to improve diagnostic accuracy and research output.This work reinforces predictions that big data and HPC will continue to improve brain science by making ultrahigh-performance analysis possible,by improving data standardization and sharing,and by providing new neuromorphic insights.
文摘Cloud computing has been considered as the next-generation computing platform with the potential to address the data and computing challenges in geosciences.However,only a limited number of geoscientists have been adapting this platform for their scientific research mainly due to two barriers:1)selecting an appropriate cloud platform for a specific application could be challenging,as various cloud services are available and 2)existing general cloud platforms are not designed to support geoscience applications,algorithms and models.To tackle such barriers,this research aims to design a hybrid cloud computing(HCC)platform that can utilize and integrate the computing resources across different organizations to build a unified geospatial cloud computing platform.This platform can manage different types of underlying cloud infrastructure(e.g.,private or public clouds),and enables geoscientists to test and leverage the cloud capabilities through a web interface.Additionally,the platform also provides different geospatial cloud services,such as workflow as a service,on the top of common cloud services(e.g.,infrastructure as a service)provided by general cloud platforms.Therefore,geoscientists can easily create a model workflow by recruiting the needed models for a geospatial application or task on the fly.A HCC prototype is developed and dust storm simulation is used to demonstrate the capability and feasibility of such platform in facilitating geosciences by leveraging across-organization computing and model resources.
基金This work is supported by the National Key Research and Development Program of China under Grant No.2016YFB1000302the National Natural Science Foundation of China under Grant Nos.U1611261 and 61872392the Program for Guangdong Introducing Innovative and Entrepreneurial Teams under Grant No.2016ZT06D211.
文摘With the convergence of high-performance computing(HPC),big data and artificial intelligence(AI),the HPC community is pushing for"triple use"systems to expedite scientific discoveries.However,supporting these converged applications on HPC systems presents formidable challenges in terms of storage and data management due to the explosive growth of scientific data and the fundamental differences in I/O characteristics among HPC,big data and AI workloads.In this paper,we discuss the driving force behind the converging trend,highlight three data management challenges,and summarize our efforts in addressing these data management challenges on a typical HPC system at the parallel file system,data management middleware,and user application levels.As HPC systems are approaching the border of exascale computing,this paper sheds light on how to enable application-driven data management as a preliminary step toward the deep convergence of exascale computing ecosystems,big data,and AI.
基金Project supported by the National Science Foundation of the USA(Nos.IIS-1447804 and CNS-1513120)
文摘With the significant advancement in emerging processor, memory, and networking technologies, exascale systems will become available in the next few years (2020 2022). As the exascale systems begin to be deployed and used, there will be a continuous demand to run next-generation applications with finer granularity, finer time-steps, and increased data sizes. Based on historical trends, next-generation applications will require post-exascale systems during 2025-2035. In this study, we focus on the networking and communication challenges for post-exascale systems. Firstly, we present an envisioned architecture for post-exascale systems. Secondly, the challenges are summarized from different perspectives: heterogeneous networking technologies, high-performance eonmmnication and synchronization protocols, integrated support with accelerators and field-programmable gate arrays, fault-tolerance and quality-of-service support, energy-aware communication schemes and protocols, software- defined networking, and scalable communication protocols with heterogeneous memory and storage. Thirdly, we present the challenges in designing efficient programming model support for high-performance computing, big data, and deep learning on these systems. Finally, we emphasize the critical need for co-designing runtime with upper layers on these systems to achieve the maximum performance and scalability.