240 nm AlGaN-based micro-LEDs with different sizes are designed and fabricated.Then,the external quantum efficiency(EQE)and light extraction efficiency(LEE)are systematically investigated by comparing size and edge ef...240 nm AlGaN-based micro-LEDs with different sizes are designed and fabricated.Then,the external quantum efficiency(EQE)and light extraction efficiency(LEE)are systematically investigated by comparing size and edge effects.Here,it is revealed that the peak optical output power increases by 81.83%with the size shrinking from 50.0 to 25.0μm.Thereinto,the LEE increases by 26.21%and the LEE enhancement mainly comes from the sidewall light extraction.Most notably,transversemagnetic(TM)mode light intensifies faster as the size shrinks due to the tilted mesa side-wall and Al reflector design.However,when it turns to 12.5μm sized micro-LEDs,the output power is lower than 25.0μm sized ones.The underlying mechanism is that even though protected by SiO2 passivation,the edge effect which leads to current leakage and Shockley-Read-Hall(SRH)recombination deteriorates rapidly with the size further shrinking.Moreover,the ratio of the p-contact area to mesa area is much lower,which deteriorates the p-type current spreading at the mesa edge.These findings show a role of thumb for the design of high efficiency micro-LEDs with wavelength below 250 nm,which will pave the way for wide applications of deep ultraviolet(DUV)micro-LEDs.展开更多
By pushing computation,cache,and network control to the edge,mobile edge computing(MEC)is expected to play a leading role in fifth generation(5G)and future sixth generation(6G).Nevertheless,facing ubiquitous fast-grow...By pushing computation,cache,and network control to the edge,mobile edge computing(MEC)is expected to play a leading role in fifth generation(5G)and future sixth generation(6G).Nevertheless,facing ubiquitous fast-growing computational demands,it is impossible for a single MEC paradigm to effectively support high-quality intelligent services at end user equipments(UEs).To address this issue,we propose an air-ground collaborative MEC(AGCMEC)architecture in this article.The proposed AGCMEC integrates all potentially available MEC servers within air and ground in the envisioned 6G,by a variety of collaborative ways to provide computation services at their best for UEs.Firstly,we introduce the AGC-MEC architecture and elaborate three typical use cases.Then,we discuss four main challenges in the AGC-MEC as well as their potential solutions.Next,we conduct a case study of collaborative service placement for AGC-MEC to validate the effectiveness of the proposed collaborative service placement strategy.Finally,we highlight several potential research directions of the AGC-MEC.展开更多
The increasing popularity of the metaverse has led to a growing interest and market size in spatial computing from both academia and industry.Developing portable and accurate imaging and depth sensing systems is cruci...The increasing popularity of the metaverse has led to a growing interest and market size in spatial computing from both academia and industry.Developing portable and accurate imaging and depth sensing systems is crucial for advancing next-generation virtual reality devices.This work demonstrates an intelligent,lightweight,and compact edge-enhanced depth perception system that utilizes a binocular meta-lens for spatial computing.The miniaturized system comprises a binocular meta-lens,a 532 nm filter,and a CMOS sensor.For disparity computation,we propose a stereo-matching neural network with a novel H-Module.The H-Module incorporates an attention mechanism into the Siamese network.The symmetric architecture,with cross-pixel interaction and cross-view interaction,enables a more comprehensive analysis of contextual information in stereo images.Based on spatial intensity discontinuity,the edge enhancement eliminates illposed regions in the image where ambiguous depth predictions may occur due to a lack of texture.With the assistance of deep learning,our edge-enhanced system provides prompt responses in less than 0.15 seconds.This edge-enhanced depth perception meta-lens imaging system will significantly contribute to accurate 3D scene modeling,machine vision,autonomous driving,and robotics development.展开更多
In vehicle edge computing(VEC),asynchronous federated learning(AFL)is used,where the edge receives a local model and updates the global model,effectively reducing the global aggregation latency.Due to different amount...In vehicle edge computing(VEC),asynchronous federated learning(AFL)is used,where the edge receives a local model and updates the global model,effectively reducing the global aggregation latency.Due to different amounts of local data,computing capabilities and locations of the vehicles,renewing the global model with same weight is inappropriate.The above factors will affect the local calculation time and upload time of the local model,and the vehicle may also be affected by Byzantine attacks,leading to the deterioration of the vehicle data.However,based on deep reinforcement learning(DRL),we can consider these factors comprehensively to eliminate vehicles with poor performance as much as possible and exclude vehicles that have suffered Byzantine attacks before AFL.At the same time,when aggregating AFL,we can focus on those vehicles with better performance to improve the accuracy and safety of the system.In this paper,we proposed a vehicle selection scheme based on DRL in VEC.In this scheme,vehicle’s mobility,channel conditions with temporal variations,computational resources with temporal variations,different data amount,transmission channel status of vehicles as well as Byzantine attacks were taken into account.Simulation results show that the proposed scheme effectively improves the safety and accuracy of the global model.展开更多
With the popularity of online learning in educational settings, knowledge tracing(KT) plays an increasingly significant role. The task of KT is to help students learn more effectively by predicting their next mastery ...With the popularity of online learning in educational settings, knowledge tracing(KT) plays an increasingly significant role. The task of KT is to help students learn more effectively by predicting their next mastery of knowledge based on their historical exercise sequences. Nowadays, many related works have emerged in this field, such as Bayesian knowledge tracing and deep knowledge tracing methods. Despite the progress that has been made in KT, existing techniques still have the following limitations: 1) Previous studies address KT by only exploring the observational sparsity data distribution, and the counterfactual data distribution has been largely ignored. 2) Current works designed for KT only consider either the entity relationships between questions and concepts, or the relations between two concepts, and none of them investigates the relations among students, questions, and concepts, simultaneously, leading to inaccurate student modeling. To address the above limitations,we propose a graph counterfactual augmentation method for knowledge tracing. Concretely, to consider the multiple relationships among different entities, we first uniform students, questions, and concepts in graphs, and then leverage a heterogeneous graph convolutional network to conduct representation learning.To model the counterfactual world, we conduct counterfactual transformations on students’ learning graphs by changing the corresponding treatments and then exploit the counterfactual outcomes in a contrastive learning framework. We conduct extensive experiments on three real-world datasets, and the experimental results demonstrate the superiority of our proposed Graph CA method compared with several state-of-the-art baselines.展开更多
Security issues in cloud networks and edge computing have become very common. This research focuses on analyzing such issues and developing the best solutions. A detailed literature review has been conducted in this r...Security issues in cloud networks and edge computing have become very common. This research focuses on analyzing such issues and developing the best solutions. A detailed literature review has been conducted in this regard. The findings have shown that many challenges are linked to edge computing, such as privacy concerns, security breaches, high costs, low efficiency, etc. Therefore, there is a need to implement proper security measures to overcome these issues. Using emerging trends, like machine learning, encryption, artificial intelligence, real-time monitoring, etc., can help mitigate security issues. They can also develop a secure and safe future in cloud computing. It was concluded that the security implications of edge computing can easily be covered with the help of new technologies and techniques.展开更多
In this paper,we consider mobile edge computing(MEC)networks against proactive eavesdropping.To maximize the transmission rate,IRS assisted UAV communications are applied.We take the joint design of the trajectory of ...In this paper,we consider mobile edge computing(MEC)networks against proactive eavesdropping.To maximize the transmission rate,IRS assisted UAV communications are applied.We take the joint design of the trajectory of UAV,the transmitting beamforming of users,and the phase shift matrix of IRS.The original problem is strong non-convex and difficult to solve.We first propose two basic modes of the proactive eavesdropper,and obtain the closed-form solution for the boundary conditions of the two modes.Then we transform the original problem into an equivalent one and propose an alternating optimization(AO)based method to obtain a local optimal solution.The convergence of the algorithm is illustrated by numerical results.Further,we propose a zero forcing(ZF)based method as sub-optimal solution,and the simulation section shows that the proposed two schemes could obtain better performance compared with traditional schemes.展开更多
Edge technology aims to bring cloud resources(specifically,the computation,storage,and network)to the closed proximity of the edge devices,i.e.,smart devices where the data are produced and consumed.Embedding computin...Edge technology aims to bring cloud resources(specifically,the computation,storage,and network)to the closed proximity of the edge devices,i.e.,smart devices where the data are produced and consumed.Embedding computing and application in edge devices lead to emerging of two new concepts in edge technology:edge computing and edge analytics.Edge analytics uses some techniques or algorithms to analyse the data generated by the edge devices.With the emerging of edge analytics,the edge devices have become a complete set.Currently,edge analytics is unable to provide full support to the analytic techniques.The edge devices cannot execute advanced and sophisticated analytic algorithms following various constraints such as limited power supply,small memory size,limited resources,etc.This article aims to provide a detailed discussion on edge analytics.The key contributions of the paper are as follows-a clear explanation to distinguish between the three concepts of edge technology:edge devices,edge computing,and edge analytics,along with their issues.In addition,the article discusses the implementation of edge analytics to solve many problems and applications in various areas such as retail,agriculture,industry,and healthcare.Moreover,the research papers of the state-of-the-art edge analytics are rigorously reviewed in this article to explore the existing issues,emerging challenges,research opportunities and their directions,and applications.展开更多
We investigate the behavior of edge modes in the presence of different edge terminations and long-range(LR)hopping.Here,we mainly focus on such model crystals with two different types of structures(type I:“…-P-Q-P-Q...We investigate the behavior of edge modes in the presence of different edge terminations and long-range(LR)hopping.Here,we mainly focus on such model crystals with two different types of structures(type I:“…-P-Q-P-Q-…”and type II:“…=P-Q=P-Q=…”),where P and Q represent crystal lines(CLs),while the symbols“-”and“=”denote the distance between the nearest neighbor(NN)CLs.Based on the lattice model Hamiltonian with LR hopping,the existence of edge modes is determined analytically by using the transfer matrix method(TMM)when different edge terminals are taken into consideration.Our findings are consistent with the numerical results obtained by the exact diagonalization method.We also notice that edge modes can exhibit different behaviors under different edge terminals.Our result is helpful in solving novel edge modes in honeycomb crystalline graphene and transition metal dichalcogenides with different edge terminals.展开更多
In a network environment composed of different types of computing centers that can be divided into different layers(clod,edge layer,and others),the interconnection between them offers the possibility of peer-to-peer t...In a network environment composed of different types of computing centers that can be divided into different layers(clod,edge layer,and others),the interconnection between them offers the possibility of peer-to-peer task offloading.For many resource-constrained devices,the computation of many types of tasks is not feasible because they cannot support such computations as they do not have enough available memory and processing capacity.In this scenario,it is worth considering transferring these tasks to resource-rich platforms,such as Edge Data Centers or remote cloud servers.For different reasons,it is more exciting and appropriate to download various tasks to specific download destinations depending on the properties and state of the environment and the nature of the functions.At the same time,establishing an optimal offloading policy,which ensures that all tasks are executed within the required latency and avoids excessive workload on specific computing centers is not easy.This study presents two alternatives to solve the offloading decision paradigm by introducing two well-known algorithms,Graph Neural Networks(GNN)and Deep Q-Network(DQN).It applies the alternatives on a well-known Edge Computing simulator called PureEdgeSimand compares them with the two defaultmethods,Trade-Off and Round Robin.Experiments showed that variants offer a slight improvement in task success rate and workload distribution.In terms of energy efficiency,they provided similar results.Finally,the success rates of different computing centers are tested,and the lack of capacity of remote cloud servers to respond to applications in real-time is demonstrated.These novel ways of finding a download strategy in a local networking environment are unique as they emulate the state and structure of the environment innovatively,considering the quality of its connections and constant updates.The download score defined in this research is a crucial feature for determining the quality of a download path in the GNN training process and has not previously been proposed.Simultaneously,the suitability of Reinforcement Learning(RL)techniques is demonstrated due to the dynamism of the network environment,considering all the key factors that affect the decision to offload a given task,including the actual state of all devices.展开更多
The rapid development of emerging technologies,such as edge intelligence and digital twins,have added momentum towards the development of the Industrial Internet of Things(IIo T).However,the massive amount of data gen...The rapid development of emerging technologies,such as edge intelligence and digital twins,have added momentum towards the development of the Industrial Internet of Things(IIo T).However,the massive amount of data generated by the IIo T,coupled with heterogeneous computation capacity across IIo T devices,and users’data privacy concerns,have posed challenges towards achieving industrial edge intelligence(IEI).To achieve IEI,in this paper,we propose a semi-federated learning framework where a portion of the data with higher privacy is kept locally and a portion of the less private data can be potentially uploaded to the edge server.In addition,we leverage digital twins to overcome the problem of computation capacity heterogeneity of IIo T devices through the mapping of physical entities.We formulate a synchronization latency minimization problem which jointly optimizes edge association and the proportion of uploaded nonprivate data.As the joint problem is NP-hard and combinatorial and taking into account the reality of largescale device training,we develop a multi-agent hybrid action deep reinforcement learning(DRL)algorithm to find the optimal solution.Simulation results show that our proposed DRL algorithm can reduce latency and have a better convergence performance for semi-federated learning compared to benchmark algorithms.展开更多
We construct a one-dimensional quasiperiodic quantum walk to investigate the localization–delocalization transition.The inverse participation ratio and Lyapunov exponent are employed as two indexes to determine the m...We construct a one-dimensional quasiperiodic quantum walk to investigate the localization–delocalization transition.The inverse participation ratio and Lyapunov exponent are employed as two indexes to determine the mobility edge, a critical energy to distinguish the energy regions of extended and localized states. The analytical solution of mobility edge is obtained by the Lyapunov exponents in global theory, and the consistency of the two indexes is confirmed. We further study the dynamic characteristics of the quantum walk and show that the probabilities are localized to some specific lattice sites with time evolution. This phenomenon is explained by the effective potential of the Hamiltonian which corresponds to the phase in the coin operator of the quantum walk.展开更多
[Objective]Real-time monitoring of cow ruminant behavior is of paramount importance for promptly obtaining relevant information about cow health and predicting cow diseases.Currently,various strategies have been propo...[Objective]Real-time monitoring of cow ruminant behavior is of paramount importance for promptly obtaining relevant information about cow health and predicting cow diseases.Currently,various strategies have been proposed for monitoring cow ruminant behavior,including video surveillance,sound recognition,and sensor monitoring methods.How‐ever,the application of edge device gives rise to the issue of inadequate real-time performance.To reduce the volume of data transmission and cloud computing workload while achieving real-time monitoring of dairy cow rumination behavior,a real-time monitoring method was proposed for cow ruminant behavior based on edge computing.[Methods]Autono‐mously designed edge devices were utilized to collect and process six-axis acceleration signals from cows in real-time.Based on these six-axis data,two distinct strategies,federated edge intelligence and split edge intelligence,were investigat‐ed for the real-time recognition of cow ruminant behavior.Focused on the real-time recognition method for cow ruminant behavior leveraging federated edge intelligence,the CA-MobileNet v3 network was proposed by enhancing the MobileNet v3 network with a collaborative attention mechanism.Additionally,a federated edge intelligence model was designed uti‐lizing the CA-MobileNet v3 network and the FedAvg federated aggregation algorithm.In the study on split edge intelli‐gence,a split edge intelligence model named MobileNet-LSTM was designed by integrating the MobileNet v3 network with a fusion collaborative attention mechanism and the Bi-LSTM network.[Results and Discussions]Through compara‐tive experiments with MobileNet v3 and MobileNet-LSTM,the federated edge intelligence model based on CA-Mo‐bileNet v3 achieved an average Precision rate,Recall rate,F1-Score,Specificity,and Accuracy of 97.1%,97.9%,97.5%,98.3%,and 98.2%,respectively,yielding the best recognition performance.[Conclusions]It is provided a real-time and effective method for monitoring cow ruminant behavior,and the proposed federated edge intelligence model can be ap‐plied in practical settings.展开更多
In mega-constellation Communication Systems, efficient routing algorithms and data transmission technologies are employed to ensure fast and reliable data transfer. However, the limited computational resources of sate...In mega-constellation Communication Systems, efficient routing algorithms and data transmission technologies are employed to ensure fast and reliable data transfer. However, the limited computational resources of satellites necessitate the use of edge computing to enhance secure communication.While edge computing reduces the burden on cloud computing, it introduces security and reliability challenges in open satellite communication channels. To address these challenges, we propose a blockchain architecture specifically designed for edge computing in mega-constellation communication systems. This architecture narrows down the consensus scope of the blockchain to meet the requirements of edge computing while ensuring comprehensive log storage across the network. Additionally, we introduce a reputation management mechanism for nodes within the blockchain, evaluating their trustworthiness, workload, and efficiency. Nodes with higher reputation scores are selected to participate in tasks and are appropriately incentivized. Simulation results demonstrate that our approach achieves a task result reliability of 95% while improving computational speed.展开更多
Users and edge servers are not fullymutually trusted inmobile edge computing(MEC),and hence blockchain can be introduced to provide trustableMEC.In blockchain-basedMEC,each edge server functions as a node in bothMEC a...Users and edge servers are not fullymutually trusted inmobile edge computing(MEC),and hence blockchain can be introduced to provide trustableMEC.In blockchain-basedMEC,each edge server functions as a node in bothMEC and blockchain,processing users’tasks and then uploading the task related information to the blockchain.That is,each edge server runs both users’offloaded tasks and blockchain tasks simultaneously.Note that there is a trade-off between the resource allocation for MEC and blockchain tasks.Therefore,the allocation of the resources of edge servers to the blockchain and theMEC is crucial for the processing delay of blockchain-based MEC.Most of the existing research tackles the problem of resource allocation in either blockchain or MEC,which leads to unfavorable performance of the blockchain-based MEC system.In this paper,we study how to allocate the computing resources of edge servers to the MEC and blockchain tasks with the aimtominimize the total systemprocessing delay.For the problem,we propose a computing resource Allocation algorithmfor Blockchain-based MEC(ABM)which utilizes the Slater’s condition,Karush-Kuhn-Tucker(KKT)conditions,partial derivatives of the Lagrangian function and subgradient projection method to obtain the solution.Simulation results show that ABM converges and effectively reduces the processing delay of blockchain-based MEC.展开更多
In this paper,we present a comprehensive system model for Industrial Internet of Things(IIoT)networks empowered by Non-Orthogonal Multiple Access(NOMA)and Mobile Edge Computing(MEC)technologies.The network comprises e...In this paper,we present a comprehensive system model for Industrial Internet of Things(IIoT)networks empowered by Non-Orthogonal Multiple Access(NOMA)and Mobile Edge Computing(MEC)technologies.The network comprises essential components such as base stations,edge servers,and numerous IIoT devices characterized by limited energy and computing capacities.The central challenge addressed is the optimization of resource allocation and task distribution while adhering to stringent queueing delay constraints and minimizing overall energy consumption.The system operates in discrete time slots and employs a quasi-static approach,with a specific focus on the complexities of task partitioning and the management of constrained resources within the IIoT context.This study makes valuable contributions to the field by enhancing the understanding of resourceefficient management and task allocation,particularly relevant in real-time industrial applications.Experimental results indicate that our proposed algorithmsignificantly outperforms existing approaches,reducing queue backlog by 45.32% and 17.25% compared to SMRA and ACRA while achieving a 27.31% and 74.12% improvement in Qn O.Moreover,the algorithmeffectively balances complexity and network performance,as demonstratedwhen reducing the number of devices in each group(Ng)from 200 to 50,resulting in a 97.21% reduction in complexity with only a 7.35% increase in energy consumption.This research offers a practical solution for optimizing IIoT networks in real-time industrial settings.展开更多
It is still challenging to fully integrate computing in memory chip as edge learning devices.In recent work published on Science,a fully-integrated chip based on neuromorphic memristors was developed for edge learning...It is still challenging to fully integrate computing in memory chip as edge learning devices.In recent work published on Science,a fully-integrated chip based on neuromorphic memristors was developed for edge learning as artificial neural networks with functionality of synapses,dendrites,and somas.A crossbar-array memristor chip facilitated edge learning including hardware realization,learning algorithm,and cycle-parallel sign-and threshold-based learning(STELLAR)scheme.The motion control and demonstration platforms were executed to improve the edge learning ability for adapting to new scenarios.展开更多
基金This work was supported by National Key R&D Program of China(2022YFB3605103)the National Natural Science Foundation of China(62204241,U22A2084,62121005,and 61827813)+3 种基金the Natural Science Foundation of Jilin Province(20230101345JC,20230101360JC,and 20230101107JC)the Youth Innovation Promotion Association of CAS(2023223)the Young Elite Scientist Sponsorship Program By CAST(YESS20200182)the CAS Talents Program(E30122E4M0).
文摘240 nm AlGaN-based micro-LEDs with different sizes are designed and fabricated.Then,the external quantum efficiency(EQE)and light extraction efficiency(LEE)are systematically investigated by comparing size and edge effects.Here,it is revealed that the peak optical output power increases by 81.83%with the size shrinking from 50.0 to 25.0μm.Thereinto,the LEE increases by 26.21%and the LEE enhancement mainly comes from the sidewall light extraction.Most notably,transversemagnetic(TM)mode light intensifies faster as the size shrinks due to the tilted mesa side-wall and Al reflector design.However,when it turns to 12.5μm sized micro-LEDs,the output power is lower than 25.0μm sized ones.The underlying mechanism is that even though protected by SiO2 passivation,the edge effect which leads to current leakage and Shockley-Read-Hall(SRH)recombination deteriorates rapidly with the size further shrinking.Moreover,the ratio of the p-contact area to mesa area is much lower,which deteriorates the p-type current spreading at the mesa edge.These findings show a role of thumb for the design of high efficiency micro-LEDs with wavelength below 250 nm,which will pave the way for wide applications of deep ultraviolet(DUV)micro-LEDs.
基金supported in part by the National Natural Science Foundation of China under Grant 62171465,62072303,62272223,U22A2031。
文摘By pushing computation,cache,and network control to the edge,mobile edge computing(MEC)is expected to play a leading role in fifth generation(5G)and future sixth generation(6G).Nevertheless,facing ubiquitous fast-growing computational demands,it is impossible for a single MEC paradigm to effectively support high-quality intelligent services at end user equipments(UEs).To address this issue,we propose an air-ground collaborative MEC(AGCMEC)architecture in this article.The proposed AGCMEC integrates all potentially available MEC servers within air and ground in the envisioned 6G,by a variety of collaborative ways to provide computation services at their best for UEs.Firstly,we introduce the AGC-MEC architecture and elaborate three typical use cases.Then,we discuss four main challenges in the AGC-MEC as well as their potential solutions.Next,we conduct a case study of collaborative service placement for AGC-MEC to validate the effectiveness of the proposed collaborative service placement strategy.Finally,we highlight several potential research directions of the AGC-MEC.
基金supports from the Research Grants Council of the Hong Kong Special Administrative Region,China[Project No.C5031-22GCityU11310522+3 种基金CityU11300123]the Department of Science and Technology of Guangdong Province[Project No.2020B1515120073]City University of Hong Kong[Project No.9610628]JST CREST(Grant No.JPMJCR1904).
文摘The increasing popularity of the metaverse has led to a growing interest and market size in spatial computing from both academia and industry.Developing portable and accurate imaging and depth sensing systems is crucial for advancing next-generation virtual reality devices.This work demonstrates an intelligent,lightweight,and compact edge-enhanced depth perception system that utilizes a binocular meta-lens for spatial computing.The miniaturized system comprises a binocular meta-lens,a 532 nm filter,and a CMOS sensor.For disparity computation,we propose a stereo-matching neural network with a novel H-Module.The H-Module incorporates an attention mechanism into the Siamese network.The symmetric architecture,with cross-pixel interaction and cross-view interaction,enables a more comprehensive analysis of contextual information in stereo images.Based on spatial intensity discontinuity,the edge enhancement eliminates illposed regions in the image where ambiguous depth predictions may occur due to a lack of texture.With the assistance of deep learning,our edge-enhanced system provides prompt responses in less than 0.15 seconds.This edge-enhanced depth perception meta-lens imaging system will significantly contribute to accurate 3D scene modeling,machine vision,autonomous driving,and robotics development.
基金supported in part by the National Natural Science Foundation of China(No.61701197)in part by the National Key Research and Development Program of China(No.2021YFA1000500(4))in part by the 111 Project(No.B23008).
文摘In vehicle edge computing(VEC),asynchronous federated learning(AFL)is used,where the edge receives a local model and updates the global model,effectively reducing the global aggregation latency.Due to different amounts of local data,computing capabilities and locations of the vehicles,renewing the global model with same weight is inappropriate.The above factors will affect the local calculation time and upload time of the local model,and the vehicle may also be affected by Byzantine attacks,leading to the deterioration of the vehicle data.However,based on deep reinforcement learning(DRL),we can consider these factors comprehensively to eliminate vehicles with poor performance as much as possible and exclude vehicles that have suffered Byzantine attacks before AFL.At the same time,when aggregating AFL,we can focus on those vehicles with better performance to improve the accuracy and safety of the system.In this paper,we proposed a vehicle selection scheme based on DRL in VEC.In this scheme,vehicle’s mobility,channel conditions with temporal variations,computational resources with temporal variations,different data amount,transmission channel status of vehicles as well as Byzantine attacks were taken into account.Simulation results show that the proposed scheme effectively improves the safety and accuracy of the global model.
基金supported by the Natural Science Foundation of China (62372277)the Natural Science Foundation of Shandong Province (ZR2022MF257, ZR2022MF295)Humanities and Social Sciences Fund of the Ministry of Education (21YJC630157)。
文摘With the popularity of online learning in educational settings, knowledge tracing(KT) plays an increasingly significant role. The task of KT is to help students learn more effectively by predicting their next mastery of knowledge based on their historical exercise sequences. Nowadays, many related works have emerged in this field, such as Bayesian knowledge tracing and deep knowledge tracing methods. Despite the progress that has been made in KT, existing techniques still have the following limitations: 1) Previous studies address KT by only exploring the observational sparsity data distribution, and the counterfactual data distribution has been largely ignored. 2) Current works designed for KT only consider either the entity relationships between questions and concepts, or the relations between two concepts, and none of them investigates the relations among students, questions, and concepts, simultaneously, leading to inaccurate student modeling. To address the above limitations,we propose a graph counterfactual augmentation method for knowledge tracing. Concretely, to consider the multiple relationships among different entities, we first uniform students, questions, and concepts in graphs, and then leverage a heterogeneous graph convolutional network to conduct representation learning.To model the counterfactual world, we conduct counterfactual transformations on students’ learning graphs by changing the corresponding treatments and then exploit the counterfactual outcomes in a contrastive learning framework. We conduct extensive experiments on three real-world datasets, and the experimental results demonstrate the superiority of our proposed Graph CA method compared with several state-of-the-art baselines.
文摘Security issues in cloud networks and edge computing have become very common. This research focuses on analyzing such issues and developing the best solutions. A detailed literature review has been conducted in this regard. The findings have shown that many challenges are linked to edge computing, such as privacy concerns, security breaches, high costs, low efficiency, etc. Therefore, there is a need to implement proper security measures to overcome these issues. Using emerging trends, like machine learning, encryption, artificial intelligence, real-time monitoring, etc., can help mitigate security issues. They can also develop a secure and safe future in cloud computing. It was concluded that the security implications of edge computing can easily be covered with the help of new technologies and techniques.
基金This work was supported by the Key Scientific and Technological Project of Henan Province(Grant Number 222102210212)Doctoral Research Start Project of Henan Institute of Technology(Grant Number KQ2005)Key Research Projects of Colleges and Universities in Henan Province(Grant Number 23B510006).
文摘In this paper,we consider mobile edge computing(MEC)networks against proactive eavesdropping.To maximize the transmission rate,IRS assisted UAV communications are applied.We take the joint design of the trajectory of UAV,the transmitting beamforming of users,and the phase shift matrix of IRS.The original problem is strong non-convex and difficult to solve.We first propose two basic modes of the proactive eavesdropper,and obtain the closed-form solution for the boundary conditions of the two modes.Then we transform the original problem into an equivalent one and propose an alternating optimization(AO)based method to obtain a local optimal solution.The convergence of the algorithm is illustrated by numerical results.Further,we propose a zero forcing(ZF)based method as sub-optimal solution,and the simulation section shows that the proposed two schemes could obtain better performance compared with traditional schemes.
文摘Edge technology aims to bring cloud resources(specifically,the computation,storage,and network)to the closed proximity of the edge devices,i.e.,smart devices where the data are produced and consumed.Embedding computing and application in edge devices lead to emerging of two new concepts in edge technology:edge computing and edge analytics.Edge analytics uses some techniques or algorithms to analyse the data generated by the edge devices.With the emerging of edge analytics,the edge devices have become a complete set.Currently,edge analytics is unable to provide full support to the analytic techniques.The edge devices cannot execute advanced and sophisticated analytic algorithms following various constraints such as limited power supply,small memory size,limited resources,etc.This article aims to provide a detailed discussion on edge analytics.The key contributions of the paper are as follows-a clear explanation to distinguish between the three concepts of edge technology:edge devices,edge computing,and edge analytics,along with their issues.In addition,the article discusses the implementation of edge analytics to solve many problems and applications in various areas such as retail,agriculture,industry,and healthcare.Moreover,the research papers of the state-of-the-art edge analytics are rigorously reviewed in this article to explore the existing issues,emerging challenges,research opportunities and their directions,and applications.
基金supported by the National Natural Science Foundation of China(Grant No.11847061)Domestic Visiting Program for Young and Middle-aged Teachers in Shanghai Universities.
文摘We investigate the behavior of edge modes in the presence of different edge terminations and long-range(LR)hopping.Here,we mainly focus on such model crystals with two different types of structures(type I:“…-P-Q-P-Q-…”and type II:“…=P-Q=P-Q=…”),where P and Q represent crystal lines(CLs),while the symbols“-”and“=”denote the distance between the nearest neighbor(NN)CLs.Based on the lattice model Hamiltonian with LR hopping,the existence of edge modes is determined analytically by using the transfer matrix method(TMM)when different edge terminals are taken into consideration.Our findings are consistent with the numerical results obtained by the exact diagonalization method.We also notice that edge modes can exhibit different behaviors under different edge terminals.Our result is helpful in solving novel edge modes in honeycomb crystalline graphene and transition metal dichalcogenides with different edge terminals.
基金funding from TECNALIA,Basque Research and Technology Alliance(BRTA)supported by the project aOptimization of Deep Learning algorithms for Edge IoT devices for sensorization and control in Buildings and Infrastructures(EMBED)funded by the Gipuzkoa Provincial Council and approved under the 2023 call of the Guipuzcoan Network of Science,Technology and Innovation Program with File Number 2023-CIEN-000051-01.
文摘In a network environment composed of different types of computing centers that can be divided into different layers(clod,edge layer,and others),the interconnection between them offers the possibility of peer-to-peer task offloading.For many resource-constrained devices,the computation of many types of tasks is not feasible because they cannot support such computations as they do not have enough available memory and processing capacity.In this scenario,it is worth considering transferring these tasks to resource-rich platforms,such as Edge Data Centers or remote cloud servers.For different reasons,it is more exciting and appropriate to download various tasks to specific download destinations depending on the properties and state of the environment and the nature of the functions.At the same time,establishing an optimal offloading policy,which ensures that all tasks are executed within the required latency and avoids excessive workload on specific computing centers is not easy.This study presents two alternatives to solve the offloading decision paradigm by introducing two well-known algorithms,Graph Neural Networks(GNN)and Deep Q-Network(DQN).It applies the alternatives on a well-known Edge Computing simulator called PureEdgeSimand compares them with the two defaultmethods,Trade-Off and Round Robin.Experiments showed that variants offer a slight improvement in task success rate and workload distribution.In terms of energy efficiency,they provided similar results.Finally,the success rates of different computing centers are tested,and the lack of capacity of remote cloud servers to respond to applications in real-time is demonstrated.These novel ways of finding a download strategy in a local networking environment are unique as they emulate the state and structure of the environment innovatively,considering the quality of its connections and constant updates.The download score defined in this research is a crucial feature for determining the quality of a download path in the GNN training process and has not previously been proposed.Simultaneously,the suitability of Reinforcement Learning(RL)techniques is demonstrated due to the dynamism of the network environment,considering all the key factors that affect the decision to offload a given task,including the actual state of all devices.
基金supported in part by the National Nature Science Foundation of China under Grant 62001168in part by the Foundation and Application Research Grant of Guangzhou under Grant 202102020515。
文摘The rapid development of emerging technologies,such as edge intelligence and digital twins,have added momentum towards the development of the Industrial Internet of Things(IIo T).However,the massive amount of data generated by the IIo T,coupled with heterogeneous computation capacity across IIo T devices,and users’data privacy concerns,have posed challenges towards achieving industrial edge intelligence(IEI).To achieve IEI,in this paper,we propose a semi-federated learning framework where a portion of the data with higher privacy is kept locally and a portion of the less private data can be potentially uploaded to the edge server.In addition,we leverage digital twins to overcome the problem of computation capacity heterogeneity of IIo T devices through the mapping of physical entities.We formulate a synchronization latency minimization problem which jointly optimizes edge association and the proportion of uploaded nonprivate data.As the joint problem is NP-hard and combinatorial and taking into account the reality of largescale device training,we develop a multi-agent hybrid action deep reinforcement learning(DRL)algorithm to find the optimal solution.Simulation results show that our proposed DRL algorithm can reduce latency and have a better convergence performance for semi-federated learning compared to benchmark algorithms.
文摘We construct a one-dimensional quasiperiodic quantum walk to investigate the localization–delocalization transition.The inverse participation ratio and Lyapunov exponent are employed as two indexes to determine the mobility edge, a critical energy to distinguish the energy regions of extended and localized states. The analytical solution of mobility edge is obtained by the Lyapunov exponents in global theory, and the consistency of the two indexes is confirmed. We further study the dynamic characteristics of the quantum walk and show that the probabilities are localized to some specific lattice sites with time evolution. This phenomenon is explained by the effective potential of the Hamiltonian which corresponds to the phase in the coin operator of the quantum walk.
文摘[Objective]Real-time monitoring of cow ruminant behavior is of paramount importance for promptly obtaining relevant information about cow health and predicting cow diseases.Currently,various strategies have been proposed for monitoring cow ruminant behavior,including video surveillance,sound recognition,and sensor monitoring methods.How‐ever,the application of edge device gives rise to the issue of inadequate real-time performance.To reduce the volume of data transmission and cloud computing workload while achieving real-time monitoring of dairy cow rumination behavior,a real-time monitoring method was proposed for cow ruminant behavior based on edge computing.[Methods]Autono‐mously designed edge devices were utilized to collect and process six-axis acceleration signals from cows in real-time.Based on these six-axis data,two distinct strategies,federated edge intelligence and split edge intelligence,were investigat‐ed for the real-time recognition of cow ruminant behavior.Focused on the real-time recognition method for cow ruminant behavior leveraging federated edge intelligence,the CA-MobileNet v3 network was proposed by enhancing the MobileNet v3 network with a collaborative attention mechanism.Additionally,a federated edge intelligence model was designed uti‐lizing the CA-MobileNet v3 network and the FedAvg federated aggregation algorithm.In the study on split edge intelli‐gence,a split edge intelligence model named MobileNet-LSTM was designed by integrating the MobileNet v3 network with a fusion collaborative attention mechanism and the Bi-LSTM network.[Results and Discussions]Through compara‐tive experiments with MobileNet v3 and MobileNet-LSTM,the federated edge intelligence model based on CA-Mo‐bileNet v3 achieved an average Precision rate,Recall rate,F1-Score,Specificity,and Accuracy of 97.1%,97.9%,97.5%,98.3%,and 98.2%,respectively,yielding the best recognition performance.[Conclusions]It is provided a real-time and effective method for monitoring cow ruminant behavior,and the proposed federated edge intelligence model can be ap‐plied in practical settings.
基金supported in part by the National Natural Science Foundation of China under Grant No.U2268204,62172061 and 61871422National Key R&D Program of China under Grant No.2020YFB1711800 and 2020YFB1707900+2 种基金the Science and Technology Project of Sichuan Province under Grant No.2023ZHCG0014,2023ZHCG0011,2022YFG0155,2022YFG0157,2021GFW019,2021YFG0152,2021YFG0025,2020YFG0322Central Universities of Southwest Minzu University under Grant No.ZYN2022032,2023NYXXS034the State Scholarship Fund of the China Scholarship Council under Grant No.202008510081。
文摘In mega-constellation Communication Systems, efficient routing algorithms and data transmission technologies are employed to ensure fast and reliable data transfer. However, the limited computational resources of satellites necessitate the use of edge computing to enhance secure communication.While edge computing reduces the burden on cloud computing, it introduces security and reliability challenges in open satellite communication channels. To address these challenges, we propose a blockchain architecture specifically designed for edge computing in mega-constellation communication systems. This architecture narrows down the consensus scope of the blockchain to meet the requirements of edge computing while ensuring comprehensive log storage across the network. Additionally, we introduce a reputation management mechanism for nodes within the blockchain, evaluating their trustworthiness, workload, and efficiency. Nodes with higher reputation scores are selected to participate in tasks and are appropriately incentivized. Simulation results demonstrate that our approach achieves a task result reliability of 95% while improving computational speed.
基金supported by the Key Research and Development Project in Anhui Province of China(Grant No.202304a05020059)the Fundamental Research Funds for the Central Universities of China(Grant No.PA2023GDSK0055)the Project of Anhui Province Economic and Information Bureau(Grant No.JB20099).
文摘Users and edge servers are not fullymutually trusted inmobile edge computing(MEC),and hence blockchain can be introduced to provide trustableMEC.In blockchain-basedMEC,each edge server functions as a node in bothMEC and blockchain,processing users’tasks and then uploading the task related information to the blockchain.That is,each edge server runs both users’offloaded tasks and blockchain tasks simultaneously.Note that there is a trade-off between the resource allocation for MEC and blockchain tasks.Therefore,the allocation of the resources of edge servers to the blockchain and theMEC is crucial for the processing delay of blockchain-based MEC.Most of the existing research tackles the problem of resource allocation in either blockchain or MEC,which leads to unfavorable performance of the blockchain-based MEC system.In this paper,we study how to allocate the computing resources of edge servers to the MEC and blockchain tasks with the aimtominimize the total systemprocessing delay.For the problem,we propose a computing resource Allocation algorithmfor Blockchain-based MEC(ABM)which utilizes the Slater’s condition,Karush-Kuhn-Tucker(KKT)conditions,partial derivatives of the Lagrangian function and subgradient projection method to obtain the solution.Simulation results show that ABM converges and effectively reduces the processing delay of blockchain-based MEC.
基金the Deanship of Scientific Research at King Khalid University for funding this work through large group research project under Grant Number RGP2/474/44.
文摘In this paper,we present a comprehensive system model for Industrial Internet of Things(IIoT)networks empowered by Non-Orthogonal Multiple Access(NOMA)and Mobile Edge Computing(MEC)technologies.The network comprises essential components such as base stations,edge servers,and numerous IIoT devices characterized by limited energy and computing capacities.The central challenge addressed is the optimization of resource allocation and task distribution while adhering to stringent queueing delay constraints and minimizing overall energy consumption.The system operates in discrete time slots and employs a quasi-static approach,with a specific focus on the complexities of task partitioning and the management of constrained resources within the IIoT context.This study makes valuable contributions to the field by enhancing the understanding of resourceefficient management and task allocation,particularly relevant in real-time industrial applications.Experimental results indicate that our proposed algorithmsignificantly outperforms existing approaches,reducing queue backlog by 45.32% and 17.25% compared to SMRA and ACRA while achieving a 27.31% and 74.12% improvement in Qn O.Moreover,the algorithmeffectively balances complexity and network performance,as demonstratedwhen reducing the number of devices in each group(Ng)from 200 to 50,resulting in a 97.21% reduction in complexity with only a 7.35% increase in energy consumption.This research offers a practical solution for optimizing IIoT networks in real-time industrial settings.
基金funding support from the National Natural Science Foundation of China(52172205).
文摘It is still challenging to fully integrate computing in memory chip as edge learning devices.In recent work published on Science,a fully-integrated chip based on neuromorphic memristors was developed for edge learning as artificial neural networks with functionality of synapses,dendrites,and somas.A crossbar-array memristor chip facilitated edge learning including hardware realization,learning algorithm,and cycle-parallel sign-and threshold-based learning(STELLAR)scheme.The motion control and demonstration platforms were executed to improve the edge learning ability for adapting to new scenarios.