期刊文献+
共找到828篇文章
< 1 2 42 >
每页显示 20 50 100
Federated Feature Concatenate Method for Heterogeneous Computing in Federated Learning 被引量:1
1
作者 Wu-Chun Chung Yung-Chin Chang +2 位作者 Ching-Hsien Hsu Chih-Hung Chang Che-Lun Hung 《Computers, Materials & Continua》 SCIE EI 2023年第4期351-371,共21页
Federated learning is an emerging machine learning techniquethat enables clients to collaboratively train a deep learning model withoutuploading raw data to the aggregation server. Each client may be equippedwith diff... Federated learning is an emerging machine learning techniquethat enables clients to collaboratively train a deep learning model withoutuploading raw data to the aggregation server. Each client may be equippedwith different computing resources for model training. The client equippedwith a lower computing capability requires more time for model training,resulting in a prolonged training time in federated learning. Moreover, it mayfail to train the entire model because of the out-of-memory issue. This studyaims to tackle these problems and propose the federated feature concatenate(FedFC) method for federated learning considering heterogeneous clients.FedFC leverages the model splitting and feature concatenate for offloadinga portion of the training loads from clients to the aggregation server. Eachclient in FedFC can collaboratively train a model with different cutting layers.Therefore, the specific features learned in the deeper layer of the serversidemodel are more identical for the data class classification. Accordingly,FedFC can reduce the computation loading for the resource-constrainedclient and accelerate the convergence time. The performance effectiveness isverified by considering different dataset scenarios, such as data and classimbalance for the participant clients in the experiments. The performanceimpacts of different cutting layers are evaluated during the model training.The experimental results show that the co-adapted features have a criticalimpact on the adequate classification of the deep learning model. Overall,FedFC not only shortens the convergence time, but also improves the bestaccuracy by up to 5.9% and 14.5% when compared to conventional federatedlearning and splitfed, respectively. In conclusion, the proposed approach isfeasible and effective for heterogeneous clients in federated learning. 展开更多
关键词 Federated learning deep learning artificial intelligence heterogeneous computing
下载PDF
A new heuristic for task scheduling in heterogeneous computing environment
2
作者 Ehsan Ullah MUNIR Jian-zhong LI +2 位作者 Sheng-fei SHI Zhao-nian ZOU Qaisar RASOOL 《Journal of Zhejiang University-Science A(Applied Physics & Engineering)》 SCIE EI CAS CSCD 2008年第12期1715-1723,共9页
Heterogeneous computing (HC) environment utilizes diverse resources with different computational capabilities to solve computing-intensive applications having diverse computational requirements and constraints. The ta... Heterogeneous computing (HC) environment utilizes diverse resources with different computational capabilities to solve computing-intensive applications having diverse computational requirements and constraints. The task assignment problem in HC environment can be formally defined as for a given set of tasks and machines, assigning tasks to machines to achieve the minimum makespan. In this paper we propose a new task scheduling heuristic, high standard deviation first (HSTDF), which considers the standard deviation of the expected execution time of a task as a selection criterion. Standard deviation of the ex- pected execution time of a task represents the amount of variation in task execution time on different machines. Our conclusion is that tasks having high standard deviation must be assigned first for scheduling. A large number of experiments were carried out to check the effectiveness of the proposed heuristic in different scenarios, and the comparison with the existing heuristics (Max-min, Sufferage, Segmented Min-average, Segmented Min-min, and Segmented Max-min) clearly reveals that the proposed heuristic outperforms all existing heuristics in terms of average makespan. 展开更多
关键词 heterogeneous computing Task scheduling Greedy heuristics High standard deviation first (HSTDF) heuristic
下载PDF
Scalable Data Assignment Algorithm of Parallel I/O for Distributed Heterogeneous Computing
3
作者 ZENG Biqing CHEN Zhigang TAN Lu XIONG Ce 《通讯和计算机(中英文版)》 2005年第3期51-55,共5页
关键词 分布式数据库 I/O 数据处理 多功能计算机
下载PDF
Distributed Matching Theory-Based Task Re-Allocating for Heterogeneous Multi-UAV Edge Computing
4
作者 Yangang Wang Xianglin Wei +3 位作者 Hai Wang Yongyang Hu Kuang Zhao Jianhua Fan 《China Communications》 SCIE CSCD 2024年第1期260-278,共19页
Many efforts have been devoted to efficient task scheduling in Multi-Unmanned Aerial Vehicle(UAV)edge computing.However,the heterogeneity of UAV computation resource,and the task re-allocating between UAVs have not be... Many efforts have been devoted to efficient task scheduling in Multi-Unmanned Aerial Vehicle(UAV)edge computing.However,the heterogeneity of UAV computation resource,and the task re-allocating between UAVs have not been fully considered yet.Moreover,most existing works neglect the fact that a task can only be executed on the UAV equipped with its desired service function(SF).In this backdrop,this paper formulates the task scheduling problem as a multi-objective task scheduling problem,which aims at maximizing the task execution success ratio while minimizing the average weighted sum of all tasks’completion time and energy consumption.Optimizing three coupled goals in a realtime manner with the dynamic arrival of tasks hinders us from adopting existing methods,like machine learning-based solutions that require a long training time and tremendous pre-knowledge about the task arrival process,or heuristic-based ones that usually incur a long decision-making time.To tackle this problem in a distributed manner,we establish a matching theory framework,in which three conflicting goals are treated as the preferences of tasks,SFs and UAVs.Then,a Distributed Matching Theory-based Re-allocating(DiMaToRe)algorithm is put forward.We formally proved that a stable matching can be achieved by our proposal.Extensive simulation results show that Di Ma To Re algorithm outperforms benchmark algorithms under diverse parameter settings and has good robustness. 展开更多
关键词 edge computing HETEROGENEITY matching theory service function unmanned aerial vehicle
下载PDF
FPGA Accelerators for Computing Interatomic Potential-Based Molecular Dynamics Simulation for Gold Nanoparticles:Exploring Different Communication Protocols
5
作者 Ankitkumar Patel Srivathsan Vasudevan Satya Bulusu 《Computers, Materials & Continua》 SCIE EI 2024年第9期3803-3818,共16页
Molecular Dynamics(MD)simulation for computing Interatomic Potential(IAP)is a very important High-Performance Computing(HPC)application.MD simulation on particles of experimental relevance takes huge computation time,... Molecular Dynamics(MD)simulation for computing Interatomic Potential(IAP)is a very important High-Performance Computing(HPC)application.MD simulation on particles of experimental relevance takes huge computation time,despite using an expensive high-end server.Heterogeneous computing,a combination of the Field Programmable Gate Array(FPGA)and a computer,is proposed as a solution to compute MD simulation efficiently.In such heterogeneous computation,communication between FPGA and Computer is necessary.One such MD simulation,explained in the paper,is the(Artificial Neural Network)ANN-based IAP computation of gold(Au_(147)&Au_(309))nanoparticles.MD simulation calculates the forces between atoms and the total energy of the chemical system.This work proposes the novel design and implementation of an ANN IAP-based MD simulation for Au_(147)&Au_(309) using communication protocols,such as Universal Asynchronous Receiver-Transmitter(UART)and Ethernet,for communication between the FPGA and the host computer.To improve the latency of MD simulation through heterogeneous computing,Universal Asynchronous Receiver-Transmitter(UART)and Ethernet communication protocols were explored to conduct MD simulation of 50,000 cycles.In this study,computation times of 17.54 and 18.70 h were achieved with UART and Ethernet,respectively,compared to the conventional server time of 29 h for Au_(147) nanoparticles.The results pave the way for the development of a Lab-on-a-chip application. 展开更多
关键词 Ethernet hardware accelerator heterogeneous computing interatomic potential(IAP) MDsimulation peripheral component interconnect express(PCIe) UART
下载PDF
Optimization Task Scheduling Using Cooperation Search Algorithm for Heterogeneous Cloud Computing Systems 被引量:1
6
作者 Ahmed Y.Hamed M.Kh.Elnahary +1 位作者 Faisal S.Alsubaei Hamdy H.El-Sayed 《Computers, Materials & Continua》 SCIE EI 2023年第1期2133-2148,共16页
Cloud computing has taken over the high-performance distributed computing area,and it currently provides on-demand services and resource polling over the web.As a result of constantly changing user service demand,the ... Cloud computing has taken over the high-performance distributed computing area,and it currently provides on-demand services and resource polling over the web.As a result of constantly changing user service demand,the task scheduling problem has emerged as a critical analytical topic in cloud computing.The primary goal of scheduling tasks is to distribute tasks to available processors to construct the shortest possible schedule without breaching precedence restrictions.Assignments and schedules of tasks substantially influence system operation in a heterogeneous multiprocessor system.The diverse processes inside the heuristic-based task scheduling method will result in varying makespan in the heterogeneous computing system.As a result,an intelligent scheduling algorithm should efficiently determine the priority of every subtask based on the resources necessary to lower the makespan.This research introduced a novel efficient scheduling task method in cloud computing systems based on the cooperation search algorithm to tackle an essential task and schedule a heterogeneous cloud computing problem.The basic idea of thismethod is to use the advantages of meta-heuristic algorithms to get the optimal solution.We assess our algorithm’s performance by running it through three scenarios with varying numbers of tasks.The findings demonstrate that the suggested technique beats existingmethods NewGenetic Algorithm(NGA),Genetic Algorithm(GA),Whale Optimization Algorithm(WOA),Gravitational Search Algorithm(GSA),and Hybrid Heuristic and Genetic(HHG)by 7.9%,2.1%,8.8%,7.7%,3.4%respectively according to makespan. 展开更多
关键词 heterogeneous processors cooperation search algorithm task scheduling cloud computing
下载PDF
SCIRD: Revealing Infection of Malicious Software in Edge Computing-Enabled IoT Networks
7
作者 Jiehao Ye Wen Cheng +3 位作者 Xiaolong Liu Wenyi Zhu Xuan’ang Wu Shigen Shen 《Computers, Materials & Continua》 SCIE EI 2024年第5期2743-2769,共27页
The Internet of Things(IoT)has characteristics such as node mobility,node heterogeneity,link heterogeneity,and topology heterogeneity.In the face of the IoT characteristics and the explosive growth of IoT nodes,which ... The Internet of Things(IoT)has characteristics such as node mobility,node heterogeneity,link heterogeneity,and topology heterogeneity.In the face of the IoT characteristics and the explosive growth of IoT nodes,which brings about large-scale data processing requirements,edge computing architecture has become an emerging network architecture to support IoT applications due to its ability to provide powerful computing capabilities and good service functions.However,the defense mechanism of Edge Computing-enabled IoT Nodes(ECIoTNs)is still weak due to their limited resources,so that they are susceptible to malicious software spread,which can compromise data confidentiality and network service availability.Facing this situation,we put forward an epidemiology-based susceptible-curb-infectious-removed-dead(SCIRD)model.Then,we analyze the dynamics of ECIoTNs with different infection levels under different initial conditions to obtain the dynamic differential equations.Additionally,we establish the presence of equilibrium states in the SCIRD model.Furthermore,we conduct an analysis of the model’s stability and examine the conditions under which malicious software will either spread or disappear within Edge Computing-enabled IoT(ECIoT)networks.Lastly,we validate the efficacy and superiority of the SCIRD model through MATLAB simulations.These research findings offer a theoretical foundation for suppressing the propagation of malicious software in ECIoT networks.The experimental results indicate that the theoretical SCIRD model has instructive significance,deeply revealing the principles of malicious software propagation in ECIoT networks.This study solves a challenging security problem of ECIoT networks by determining the malicious software propagation threshold,which lays the foundation for buildingmore secure and reliable ECIoT networks. 展开更多
关键词 Edge computing Internet of Things malicious software propagation model HETEROGENEITY
下载PDF
Network simulation task partition method in heterogeneous computing environment 被引量:1
8
作者 Xiaofeng Wang Wei Zhu Yueming Dai 《International Journal of Modeling, Simulation, and Scientific Computing》 EI 2014年第3期136-152,共17页
To reduce the running time of network simulation in heterogeneous computing environment,a network simulation task partition method,named LBPHCE,is put forward.In this method,the network simulation task is partitioned ... To reduce the running time of network simulation in heterogeneous computing environment,a network simulation task partition method,named LBPHCE,is put forward.In this method,the network simulation task is partitioned in comprehensive consideration of the load balance of both routing computing simulation and packet forwarding simulation.First,through benchmark experiments,the computation ability and routing simulation ability of each simulation machine are measured in the heterogeneous computing environment.Second,based on the computation ability of each simulation machine,the network simulation task is initially partitioned to meet the load balance of packet forwarding simulation in the heterogeneous computing environment,and then according to the routing computation ability,the scale of each partition is fine-tuned to satisfy the balance of the routing computing simulation,meanwhile the load balance of packet forwarding simulation is guaranteed.Experiments based on PDNS indicate that,compared to traditional uniform partition method,the LBPHCE method can reduce the total simulation running time by 26.3%in average,and compared to the liner partition method,it can reduce the running time by 18.3%in average. 展开更多
关键词 Networks simulation distributed simulation heterogeneous computing environments task partition
原文传递
Energy-Optimal and Delay-Bounded Computation Offloading in Mobile Edge Computing with Heterogeneous Clouds 被引量:24
9
作者 Tianchu Zhao Sheng Zhou +3 位作者 Linqi Song Zhiyuan Jiang Xueying Guo Zhisheng Niu 《China Communications》 SCIE CSCD 2020年第5期191-210,共20页
By Mobile Edge Computing(MEC), computation-intensive tasks are offloaded from mobile devices to cloud servers, and thus the energy consumption of mobile devices can be notably reduced. In this paper, we study task off... By Mobile Edge Computing(MEC), computation-intensive tasks are offloaded from mobile devices to cloud servers, and thus the energy consumption of mobile devices can be notably reduced. In this paper, we study task offloading in multi-user MEC systems with heterogeneous clouds, including edge clouds and remote clouds. Tasks are forwarded from mobile devices to edge clouds via wireless channels, and they can be further forwarded to remote clouds via the Internet. Our objective is to minimize the total energy consumption of multiple mobile devices, subject to bounded-delay requirements of tasks. Based on dynamic programming, we propose an algorithm that minimizes the energy consumption, by jointly allocating bandwidth and computational resources to mobile devices. The algorithm is of pseudo-polynomial complexity. To further reduce the complexity, we propose an approximation algorithm with energy discretization, and its total energy consumption is proved to be within a bounded gap from the optimum. Simulation results show that, nearly 82.7% energy of mobile devices can be saved by task offloading compared with mobile device execution. 展开更多
关键词 mobile edge computing heterogeneous clouds energy saving delay bounds dynamic programming
下载PDF
Joint Resource Allocation Using Evolutionary Algorithms in Heterogeneous Mobile Cloud Computing Networks 被引量:10
10
作者 Weiwei Xia Lianfeng Shen 《China Communications》 SCIE CSCD 2018年第8期189-204,共16页
The problem of joint radio and cloud resources allocation is studied for heterogeneous mobile cloud computing networks. The objective of the proposed joint resource allocation schemes is to maximize the total utility ... The problem of joint radio and cloud resources allocation is studied for heterogeneous mobile cloud computing networks. The objective of the proposed joint resource allocation schemes is to maximize the total utility of users as well as satisfy the required quality of service(QoS) such as the end-to-end response latency experienced by each user. We formulate the problem of joint resource allocation as a combinatorial optimization problem. Three evolutionary approaches are considered to solve the problem: genetic algorithm(GA), ant colony optimization with genetic algorithm(ACO-GA), and quantum genetic algorithm(QGA). To decrease the time complexity, we propose a mapping process between the resource allocation matrix and the chromosome of GA, ACO-GA, and QGA, search the available radio and cloud resource pairs based on the resource availability matrixes for ACOGA, and encode the difference value between the allocated resources and the minimum resource requirement for QGA. Extensive simulation results show that our proposed methods greatly outperform the existing algorithms in terms of running time, the accuracy of final results, the total utility, resource utilization and the end-to-end response latency guaranteeing. 展开更多
关键词 heterogeneous mobile cloud computing networks resource allocation genetic algorithm ant colony optimization quantum genetic algorithm
下载PDF
System Support for Parallel Computing on Heterogeneous Networks of Workstations 被引量:2
11
作者 Xiaodong Zhang(High Performance Computing and Software Laboratory University of Texas at San Antonio San Antonio, Texas 78249, U .S .A.) 《Wuhan University Journal of Natural Sciences》 CAS 1996年第Z1期362-370,共9页
Abstract In this paper, we introduce several on-going research projects to support parallel and distribut,ed computing on heterogeneous networks of workstations (NOW) in the High Performance Computing and Software Lah... Abstract In this paper, we introduce several on-going research projects to support parallel and distribut,ed computing on heterogeneous networks of workstations (NOW) in the High Performance Computing and Software Lahoratory at the University of Texas at San Antonio. The projects at aiming at addressing three technical issues. First, the factors of heterogeneity and time-sharing effects make traditional performance models/metrics for homogeneous computing performance measurement and evaluation not. suitable for bet.erogeneous computing. We develop practical models and metrics which quantify. the heterogeneity of networks and characterize the performance effects. Second, in order to perform parallel computation effectively, special system support is necessary. We are developing system schemes for heterogeneity management, process scheduling and efficient communications. Finally, to provide insight into system performance, we are developing two types of supporting tools : a graphical instrumentation monitor to aid users in investigating performance problems and in determining the most effective way of exploiting the NOW systems, and a trace-driven simulator to test and compare different system management and scheduling schemes. 展开更多
关键词 PARALLEL SUPPORT SYSTEM heterogeneous computing
下载PDF
Programming for scientific computing on peta-scale heterogeneous parallel systems 被引量:1
12
作者 杨灿群 吴强 +2 位作者 唐滔 王锋 薛京灵 《Journal of Central South University》 SCIE EI CAS 2013年第5期1189-1203,共15页
Peta-scale high-perfomlance computing systems are increasingly built with heterogeneous CPU and GPU nodes to achieve higher power efficiency and computation throughput. While providing unprecedented capabilities to co... Peta-scale high-perfomlance computing systems are increasingly built with heterogeneous CPU and GPU nodes to achieve higher power efficiency and computation throughput. While providing unprecedented capabilities to conduct computational experiments of historic significance, these systems are presently difficult to program. The users, who are domain experts rather than computer experts, prefer to use programming models closer to their domains (e.g., physics and biology) rather than MPI and OpenME This has led the development of domain-specific programming that provides domain-specific programming interfaces but abstracts away some performance-critical architecture details. Based on experience in designing large-scale computing systems, a hybrid programming framework for scientific computing on heterogeneous architectures is proposed in this work. Its design philosophy is to provide a collaborative mechanism for domain experts and computer experts so that both domain-specific knowledge and performance-critical architecture details can be adequately exploited. Two real-world scientific applications have been evaluated on TH-IA, a peta-scale CPU-GPU heterogeneous system that is currently the 5th fastest supercomputer in the world. The experimental results show that the proposed framework is well suited for developing large-scale scientific computing applications on peta-scale heterogeneous CPU/GPU systems. 展开更多
关键词 heterogeneous parallel system programming framework scientific computing GPU computing molecular dynamic
下载PDF
An Improved Model for Computing-Intensive Tasks on Heterogeneous Workstations
13
作者 邬延辉 陆鑫达 《Journal of Shanghai Jiaotong university(Science)》 EI 2004年第2期6-9,15,共5页
An improved algorithm, which solves cooperative concurrent computing tasks using the idle cycles of a number of high performance heterogeneous workstations interconnected through a high-speed network, was proposed. In... An improved algorithm, which solves cooperative concurrent computing tasks using the idle cycles of a number of high performance heterogeneous workstations interconnected through a high-speed network, was proposed. In order to get better parallel computation performance, this paper gave a model and an algorithm of task scheduling among heterogeneous workstations, in which the costs of loading data, computing, communication and collecting results are considered. Using this efficient algorithm, an optimal subset of heterogeneous workstations with the shortest parallel executing time of tasks can be selected. 展开更多
关键词 heterogeneous parallel computing cooperative concurrent computing scheduling
下载PDF
Auction-based profit maximization offloading in mobile edge computing
14
作者 Ruyan Wang Chunyan Zang +2 位作者 Peng He Yaping Cui Dapeng Wu 《Digital Communications and Networks》 SCIE CSCD 2023年第2期545-556,共12页
Offloading Mobile Devices(MDs)computation tasks to Edge Nodes(ENs)is a promising solution to overcome computation and energy resources limitations of MDs.However,there exists an unreasonable profit allocation problem ... Offloading Mobile Devices(MDs)computation tasks to Edge Nodes(ENs)is a promising solution to overcome computation and energy resources limitations of MDs.However,there exists an unreasonable profit allocation problem between MDs and ENs caused by the excessive concern on MD profit.In this paper,we propose an auction-based computation offloading algorithm,inspiring ENs to provide high-quality service by maximizing the profit of ENs.Firstly,a novel cooperation auction framework is designed to avoid overall profit damage of ENs,which is derived from the high computation delay at the overloaded ENs.Thereafter,the bidding willingness of each MD in every round of auction is determined to ensure MD rationality.Furthermore,we put forward a payment rule for the pre-selected winner to effectively guarantee auction truthfulness.Finally,the auction-based profit maximization offloading algorithm is proposed,and the MD is allowed to occupy the computation and spectrum resources of the EN for offloading if it wins the auction.Numerical results verify the performance of the proposed algorithm.Compared with the VA algorithm,the ENs profit is increased by 23.8%,and the task discard ratio is decreased by 7.5%. 展开更多
关键词 Mobile edge computing Computation offloading heterogeneous network Auction pricing
下载PDF
eMD:基于异构计算的大规模分子动力学模拟软件
15
作者 徐顺 张宝花 +1 位作者 刘倩 金钟 《数据与计算发展前沿》 CSCD 2024年第1期21-34,共14页
【目的】异构计算已经成为高性能计算的重要组成部分,GPU异构计算可显著提速计算密集型的分子动力学模拟应用,本文介绍自研分子动力学模拟软件eMD的系统设计及其异构计算应用。【方法】首先介绍eMD软件的目标定位,包括应用功能和计算性... 【目的】异构计算已经成为高性能计算的重要组成部分,GPU异构计算可显著提速计算密集型的分子动力学模拟应用,本文介绍自研分子动力学模拟软件eMD的系统设计及其异构计算应用。【方法】首先介绍eMD软件的目标定位,包括应用功能和计算性能两方面;然后介绍软件概要设计,包括框架、模块和接口等组成部分;重点围绕面向异构计算的软件架构设计和移植优化技术进行阐述。【结果】eMD软件系统基于GPU异构计算可实现大规模体系模拟,同时提供特色的分子动力学模拟算法和模型。【结论】eMD将充分发挥GPU异构计算算力,以提升分子动力学模拟应用效率,助力分子建模理论方法的创新应用和分子科学问题的研究。 展开更多
关键词 分子动力学 GPU异构计算 并行计算 国产超算
下载PDF
异构边缘云架构下的多任务卸载算法 被引量:1
16
作者 尼俊红 臧云 《哈尔滨工程大学学报》 EI CAS CSCD 北大核心 2024年第4期800-807,共8页
为在资源有限的终端设备上运行计算密集型与时延敏感型应用,同时降低系统时延和能耗,构建边缘云异构网络模型。本文提出了一种H-PSOGA多任务卸载优化算法,并通过无人机、路边单元、车辆等边缘设备以及边缘云服务器进行多任务计算卸载。... 为在资源有限的终端设备上运行计算密集型与时延敏感型应用,同时降低系统时延和能耗,构建边缘云异构网络模型。本文提出了一种H-PSOGA多任务卸载优化算法,并通过无人机、路边单元、车辆等边缘设备以及边缘云服务器进行多任务计算卸载。该算法以先串行再并行的方式将粒子群和遗传算法结合在一起,通过适应度值排序、种群选择、多点交叉、反向变异等操作,利用遗传算法对粒子群进行优选,弥补粒子群算法早熟收敛、陷入局部最优的缺陷。6种标准测试函数的测试分析以及与基线方案进行仿真对比的结果表明:在用户数较多时,混合优化算法的系统平均开销可降低26%~43%,可以有效提高收敛精度。 展开更多
关键词 移动边缘计算 异构网络 边缘节点 任务卸载 粒子群算法 遗传算法 多目标优化 标准测试函数
下载PDF
基于云边协同的通用板级自动测试系统方案设计
17
作者 林连冬 于化男 +2 位作者 朱贺 蓝润泽 陈滨 《黑龙江大学自然科学学报》 CAS 2024年第4期485-495,共11页
针对当前数字电路自动测试领域,对通用板级自动测试系统进行了设计,完成了系统方案的设计开发和测试验证。为了解决通用板级自动测试系统的测试数据量大、本地存储容量不足、测试向量与被测试目标无法自动化匹配、无法进行批量快速测试... 针对当前数字电路自动测试领域,对通用板级自动测试系统进行了设计,完成了系统方案的设计开发和测试验证。为了解决通用板级自动测试系统的测试数据量大、本地存储容量不足、测试向量与被测试目标无法自动化匹配、无法进行批量快速测试等问题,采用云边协同架构设计了通用板级自动测试系统,避免了大量数据传输和数据集中式处理,使得云端计算资源能够集中解决关键数据处理任务。为了提高板级测试平台的通用性,使平台适配不同接口的电路板,同时具有高速网络数据处理能力,系统选用精简命令集处理器和现场可编程门阵列(Advanced RISC machines,Filed programmable gate array,ARM+FPGA)的异构计算平台作为边缘设备。为了提高系统的测试效率,采用模块化的设计思想,系统硬件测试平台设计了多总线分布式结构。系统采用扇出导向(Fan-out-oriented,FAN)算法生成测试向量,并基于浏览器/服务器(Brower/Server,B/S)架构设计了用户操作界面,用户可通过浏览器进行测试操作,完成板级自动故障测试,并自动生成故障诊断报告。实验结果表明,对两个待测板卡进行测试验证,通用板级自动测试系统可以自动识别目标板卡,自动匹配测试向量,进行自动测试,生成测试报表。基于云边协同的通用板级自动测试系统提高了测试效率,满足数字电路板卡出厂前批量测试的需求,具有实际应用价值。 展开更多
关键词 云边协同 通用板级自动测试 FAN算法 异构计算平台
下载PDF
基于数据-模型混合驱动的电力系统机电暂态快速仿真方法
18
作者 王鑫 杨珂 +3 位作者 黄文琦 马云飞 耿光超 江全元 《中国电机工程学报》 EI CSCD 北大核心 2024年第8期2955-2964,I0002,共11页
数据驱动建模方法改变了发电机传统的建模范式,导致传统的机电暂态时域仿真方法无法直接应用于新范式下的电力系统。为此,该文提出一种基于数据-模型混合驱动的机电暂态时域仿真(data and physics driven time domain simulation,DPD-T... 数据驱动建模方法改变了发电机传统的建模范式,导致传统的机电暂态时域仿真方法无法直接应用于新范式下的电力系统。为此,该文提出一种基于数据-模型混合驱动的机电暂态时域仿真(data and physics driven time domain simulation,DPD-TDS)算法。算法中发电机状态变量与节点注入电流通过数据驱动模型推理计算,并通过网络方程完成节点电压计算,两者交替求解完成仿真。算法提出一种混合驱动范式下的网络代数方程组预处理方法,用以改善仿真的收敛性;算法设计一种中央处理器单元-神经网络处理器单元(central processing unit-neural network processing unit,CPU-NPU)异构计算框架以加速仿真,CPU进行机理模型的微分代数方程求解;NPU作协处理器完成数据驱动模型的前向推理。最后在IEEE-39和Polish-2383系统中将部分或全部发电机替换为数据驱动模型进行验证,仿真结果表明,所提出的仿真算法收敛性好,计算速度快,结果准确。 展开更多
关键词 机电暂态 时域仿真 数据-模型混合驱动 收敛性 CPU-NPU异构运算
下载PDF
基于矩阵乘积态的有限纠缠量子傅里叶变换模拟
19
作者 刘晓楠 廉德萌 +1 位作者 杜帅岐 刘正煜 《计算机科学》 CSCD 北大核心 2024年第9期80-86,共7页
与经典计算不同,在量子计算中量子比特可以处于叠加态,多个量子比特之间还可以形成纠缠态。表示n个量子比特组成的量子态需要存储2^(n)个振幅,这种指数级的存储开销使得大规模的量子模拟难以进行。然而当量子态的纠缠程度有限时,使用矩... 与经典计算不同,在量子计算中量子比特可以处于叠加态,多个量子比特之间还可以形成纠缠态。表示n个量子比特组成的量子态需要存储2^(n)个振幅,这种指数级的存储开销使得大规模的量子模拟难以进行。然而当量子态的纠缠程度有限时,使用矩阵乘积态表示量子态仅需要线性的空间复杂度,可以扩大模拟的规模。使用HIP-Clang语言,基于CPU+DCU的异构编程模型,使用矩阵乘积态表示量子态,对量子傅里叶变换进行模拟。结合矩阵乘积态的特点,对量子傅里叶变换线路进行分析,减少模拟实现时不必要的张量缩并运算与正交化构建。对模拟过程中的张量缩并进行分析,使用TTGT算法完成张量缩并运算,同时利用DCU的并行处理能力来提高效率。对模拟结果进行分析,分别通过振幅误差与半经典Draper量子加法器的结果验证了模拟的正确性。对模拟规模进行分析,当量子态的纠缠熵最大时,使用16 GB的内存空间最多只能模拟24位的量子态,而当量子态内部纠缠程度较低时,可以对上百位的量子态进行量子傅里叶变换模拟。 展开更多
关键词 量子模拟 量子傅里叶变换 矩阵乘积态 异构计算 DCU HIP-Clang
下载PDF
基于启发式算法的计算机异构大数据跨源调度方法
20
作者 朱晓丽 高鹏 《新乡学院学报》 2024年第6期23-27,共5页
为了提高计算机异构大数据跨源调度性能,设计了一种基于启发式算法的计算机异构大数据跨源调度方法。将计算机异构大数据跨源调度任务划分为若干子任务,利用max-min思想和min-max思想,构建了跨源调度时间负载均衡模型。基于计算机异构... 为了提高计算机异构大数据跨源调度性能,设计了一种基于启发式算法的计算机异构大数据跨源调度方法。将计算机异构大数据跨源调度任务划分为若干子任务,利用max-min思想和min-max思想,构建了跨源调度时间负载均衡模型。基于计算机异构大数据跨源调度顺序的灵活性,利用启发式算法的最优策略,选择最优调度任务。根据计算机异构大数据的状态,计算出异构大数据传输控制协议连接的吞吐量,将待调度异构大数据的质量都作为计算机数据层的子流权重,通过处理所有待调度的子流,实现计算机异构大数据跨源调度。实验结果表明,文中设计的方法可以将跨源操作的级别条件提高至10级,跨源调度计算机异构大数据的利用率超过97%,加速比大于85%,计算机异构大数据跨源调度性能明显提升。 展开更多
关键词 启发式算法 计算机 调度任务 跨源调度 负载均衡 异构大数据
下载PDF
上一页 1 2 42 下一页 到第
使用帮助 返回顶部