Distributed fiber optic sensors(DFOSs)possess the capability to measure strain and temperature variations over long distances,demonstrating outstanding potential for monitoring underground infrastructure.This study pr...Distributed fiber optic sensors(DFOSs)possess the capability to measure strain and temperature variations over long distances,demonstrating outstanding potential for monitoring underground infrastructure.This study presents a state-of-the-art review of the DFOS applications for monitoring and assessing the deformation behavior of typical tunnel infrastructure,including bored tunnels,conventional tunnels,as well as immersed and cut-and-cover tunnels.DFOS systems based on Brillouin and Rayleigh scattering principles are both considered.When implementing DFOS monitoring,the fiber optic cable can be primarily installed along transverse and longitudinal directions to(1)measure distributed strains by continuously adhering the fiber to the structure’s surface or embedding it in the lining,or(2)measure point displacements by spot-anchoring it on the lining surface.There are four critical aspects of DFOS monitoring,including proper selection of the sensing fiber,selection of the measuring principle for the specific application,design of an effective sensor layout,and establishment of robust field sensor instrumentation.These four issues are comprehensively discussed,and practical suggestions are provided for the implementation of DFOS in tunnel infrastructure monitoring.展开更多
In this paper, platoons of autonomous vehicles operating in urban road networks are considered. From a methodological point of view, the problem of interest consists of formally characterizing vehicle state trajectory...In this paper, platoons of autonomous vehicles operating in urban road networks are considered. From a methodological point of view, the problem of interest consists of formally characterizing vehicle state trajectory tubes by means of routing decisions complying with traffic congestion criteria. To this end, a novel distributed control architecture is conceived by taking advantage of two methodologies: deep reinforcement learning and model predictive control. On one hand, the routing decisions are obtained by using a distributed reinforcement learning algorithm that exploits available traffic data at each road junction. On the other hand, a bank of model predictive controllers is in charge of computing the more adequate control action for each involved vehicle. Such tasks are here combined into a single framework:the deep reinforcement learning output(action) is translated into a set-point to be tracked by the model predictive controller;conversely, the current vehicle position, resulting from the application of the control move, is exploited by the deep reinforcement learning unit for improving its reliability. The main novelty of the proposed solution lies in its hybrid nature: on one hand it fully exploits deep reinforcement learning capabilities for decisionmaking purposes;on the other hand, time-varying hard constraints are always satisfied during the dynamical platoon evolution imposed by the computed routing decisions. To efficiently evaluate the performance of the proposed control architecture, a co-design procedure, involving the SUMO and MATLAB platforms, is implemented so that complex operating environments can be used, and the information coming from road maps(links,junctions, obstacles, semaphores, etc.) and vehicle state trajectories can be shared and exchanged. Finally by considering as operating scenario a real entire city block and a platoon of eleven vehicles described by double-integrator models, several simulations have been performed with the aim to put in light the main f eatures of the proposed approach. Moreover, it is important to underline that in different operating scenarios the proposed reinforcement learning scheme is capable of significantly reducing traffic congestion phenomena when compared with well-reputed competitors.展开更多
In this study,a novel residential virtual power plant(RVPP)scheduling method that leverages a gate recurrent unit(GRU)-integrated deep reinforcement learning(DRL)algorithm is proposed.In the proposed scheme,the GRU-in...In this study,a novel residential virtual power plant(RVPP)scheduling method that leverages a gate recurrent unit(GRU)-integrated deep reinforcement learning(DRL)algorithm is proposed.In the proposed scheme,the GRU-integrated DRL algorithm guides the RVPP to participate effectively in both the day-ahead and real-time markets,lowering the electricity purchase costs and consumption risks for end-users.The Lagrangian relaxation technique is introduced to transform the constrained Markov decision process(CMDP)into an unconstrained optimization problem,which guarantees that the constraints are strictly satisfied without determining the penalty coefficients.Furthermore,to enhance the scalability of the constrained soft actor-critic(CSAC)-based RVPP scheduling approach,a fully distributed scheduling architecture was designed to enable plug-and-play in the residential distributed energy resources(RDER).Case studies performed on the constructed RVPP scenario validated the performance of the proposed methodology in enhancing the responsiveness of the RDER to power tariffs,balancing the supply and demand of the power grid,and ensuring customer comfort.展开更多
This paper is concerned with distributed Nash equi librium seeking strategies under quantized communication. In the proposed seeking strategy, a projection operator is synthesized with a gradient search method to achi...This paper is concerned with distributed Nash equi librium seeking strategies under quantized communication. In the proposed seeking strategy, a projection operator is synthesized with a gradient search method to achieve the optimization o players' objective functions while restricting their actions within required non-empty, convex and compact domains. In addition, a leader-following consensus protocol, in which quantized informa tion flows are utilized, is employed for information sharing among players. More specifically, logarithmic quantizers and uniform quantizers are investigated under both undirected and connected communication graphs and strongly connected digraphs, respec tively. Through Lyapunov stability analysis, it is shown that play ers' actions can be steered to a neighborhood of the Nash equilib rium with logarithmic and uniform quantizers, and the quanti fied convergence error depends on the parameter of the quan tizer for both undirected and directed cases. A numerical exam ple is given to verify the theoretical results.展开更多
In the past two decades,extensive and in-depth research has been conducted on Time Series InSAR technology with the advancement of high-performance SAR satellites and the accumulation of big SAR data.The introduction ...In the past two decades,extensive and in-depth research has been conducted on Time Series InSAR technology with the advancement of high-performance SAR satellites and the accumulation of big SAR data.The introduction of distributed scatterers in Distributed Scatterers InSAR(DS-InSAR)has significantly expanded the application scenarios of InSAR geodetic measurement by increasing the number of measurement points.This study traces the history of DS-InSAR,presents the definition and characteristics of distributed scatterers,and focuses on exploring the relationships and distinctions among proposed algorithms in two crucial steps:statistically homogeneous pixel selection and phase optimization.Additionally,the latest research progress in this field is tracked and the possible development direction in the future is discussed.Through simulation experiments and two real InSAR case studies,the proposed algorithms are compared and verified,and the advantages of DS-InSAR in deformation measurement practice are demonstrated.This work not only offers insights into current trends and focal points for theoretical research on DS-InSAR but also provides practical cases and guidance for applied research.展开更多
This paper addresses a multicircular circumnavigation control for UAVs with desired angular spacing around a nonstationary target.By defining a coordinated error relative to neighboring angular spacing,under the premi...This paper addresses a multicircular circumnavigation control for UAVs with desired angular spacing around a nonstationary target.By defining a coordinated error relative to neighboring angular spacing,under the premise that target information is perfectly accessible by all nodes,a centralized circular enclosing control strategy is derived for multiple UAVs connected by an undirected graph to allow for formation behaviors concerning the moving target.Besides,to avoid the requirement of target’s states being accessible for each UAV,fixed-time distributed observers are introduced to acquire the state estimates in a fixed-time sense,and the upper boundary of settling time can be determined offline irrespective of initial properties,greatly releasing the burdensome communication traffic.Then,with the aid of fixed-time distributed observers,a distributed circular circumnavigation controller is derived to force all UAVs to collaboratively evolve along the preset circles while keeping a desired angular spacing.It is inferred from Lyapunov stability that all errors are demonstrated to be convergent.Simulations are offered to verify the utility of proposed protocol.展开更多
Recognizing human activity(HAR)from data in a smartphone sensor plays an important role in the field of health to prevent chronic diseases.Daily and weekly physical activities are recorded on the smartphone and tell t...Recognizing human activity(HAR)from data in a smartphone sensor plays an important role in the field of health to prevent chronic diseases.Daily and weekly physical activities are recorded on the smartphone and tell the user whether he is moving well or not.Typically,smartphones and their associated sensing devices operate in distributed and unstable environments.Therefore,collecting their data and extracting useful information is a significant challenge.In this context,the aimof this paper is twofold:The first is to analyze human behavior based on the recognition of physical activities.Using the results of physical activity detection and classification,the second part aims to develop a health recommendation system to notify smartphone users about their healthy physical behavior related to their physical activities.This system is based on the calculation of calories burned by each user during physical activities.In this way,conclusions can be drawn about a person’s physical behavior by estimating the number of calories burned after evaluating data collected daily or even weekly following a series of physical workouts.To identify and classify human behavior our methodology is based on artificial intelligence models specifically deep learning techniques like Long Short-Term Memory(LSTM),stacked LSTM,and bidirectional LSTM.Since human activity data contains both spatial and temporal information,we proposed,in this paper,to use of an architecture allowing the extraction of the two types of information simultaneously.While Convolutional Neural Networks(CNN)has an architecture designed for spatial information,our idea is to combine CNN with LSTM to increase classification accuracy by taking into consideration the extraction of both spatial and temporal data.The results obtained achieved an accuracy of 96%.On the other side,the data learned by these algorithms is prone to error and uncertainty.To overcome this constraint and improve performance(96%),we proposed to use the fusion mechanisms.The last combines deep learning classifiers tomodel non-accurate and ambiguous data to obtain synthetic information to aid in decision-making.The Voting and Dempster-Shafer(DS)approaches are employed.The results showed that fused classifiers based on DS theory outperformed individual classifiers(96%)with the highest accuracy level of 98%.Also,the findings disclosed that participants engaging in physical activities are healthy,showcasing a disparity in the distribution of physical activities between men and women.展开更多
The scale and complexity of big data are growing continuously,posing severe challenges to traditional data processing methods,especially in the field of clustering analysis.To address this issue,this paper introduces ...The scale and complexity of big data are growing continuously,posing severe challenges to traditional data processing methods,especially in the field of clustering analysis.To address this issue,this paper introduces a new method named Big Data Tensor Multi-Cluster Distributed Incremental Update(BDTMCDIncreUpdate),which combines distributed computing,storage technology,and incremental update techniques to provide an efficient and effective means for clustering analysis.Firstly,the original dataset is divided into multiple subblocks,and distributed computing resources are utilized to process the sub-blocks in parallel,enhancing efficiency.Then,initial clustering is performed on each sub-block using tensor-based multi-clustering techniques to obtain preliminary results.When new data arrives,incremental update technology is employed to update the core tensor and factor matrix,ensuring that the clustering model can adapt to changes in data.Finally,by combining the updated core tensor and factor matrix with historical computational results,refined clustering results are obtained,achieving real-time adaptation to dynamic data.Through experimental simulation on the Aminer dataset,the BDTMCDIncreUpdate method has demonstrated outstanding performance in terms of accuracy(ACC)and normalized mutual information(NMI)metrics,achieving an accuracy rate of 90%and an NMI score of 0.85,which outperforms existing methods such as TClusInitUpdate and TKLClusUpdate in most scenarios.Therefore,the BDTMCDIncreUpdate method offers an innovative solution to the field of big data analysis,integrating distributed computing,incremental updates,and tensor-based multi-clustering techniques.It not only improves the efficiency and scalability in processing large-scale high-dimensional datasets but also has been validated for its effectiveness and accuracy through experiments.This method shows great potential in real-world applications where dynamic data growth is common,and it is of significant importance for advancing the development of data analysis technology.展开更多
We develop a policy of observer-based dynamic event-triggered state feedback control for distributed parameter systems over a mobile sensor-plus-actuator network.It is assumed that the mobile sensing devices that prov...We develop a policy of observer-based dynamic event-triggered state feedback control for distributed parameter systems over a mobile sensor-plus-actuator network.It is assumed that the mobile sensing devices that provide spatially averaged state measurements can be used to improve state estimation in the network.For the purpose of decreasing the update frequency of controller and unnecessary sampled data transmission, an efficient dynamic event-triggered control policy is constructed.In an event-triggered system, when an error signal exceeds a specified time-varying threshold, it indicates the occurrence of a typical event.The global asymptotic stability of the event-triggered closed-loop system and the boundedness of the minimum inter-event time can be guaranteed.Based on the linear quadratic optimal regulator, the actuator selects the optimal displacement only when an event occurs.A simulation example is finally used to verify that the effectiveness of such a control strategy can enhance the system performance.展开更多
The use of privacy-enhanced facial recognition has increased in response to growing concerns about data securityand privacy in the digital age. This trend is spurred by rising demand for face recognition technology in...The use of privacy-enhanced facial recognition has increased in response to growing concerns about data securityand privacy in the digital age. This trend is spurred by rising demand for face recognition technology in a varietyof industries, including access control, law enforcement, surveillance, and internet communication. However,the growing usage of face recognition technology has created serious concerns about data monitoring and userprivacy preferences, especially in context-aware systems. In response to these problems, this study provides a novelframework that integrates sophisticated approaches such as Generative Adversarial Networks (GANs), Blockchain,and distributed computing to solve privacy concerns while maintaining exact face recognition. The framework’spainstaking design and execution strive to strike a compromise between precise face recognition and protectingpersonal data integrity in an increasingly interconnected environment. Using cutting-edge tools like Dlib for faceanalysis,Ray Cluster for distributed computing, and Blockchain for decentralized identity verification, the proposedsystem provides scalable and secure facial analysis while protecting user privacy. The study’s contributions includethe creation of a sustainable and scalable solution for privacy-aware face recognition, the implementation of flexibleprivacy computing approaches based on Blockchain networks, and the demonstration of higher performanceover previous methods. Specifically, the proposed StyleGAN model has an outstanding accuracy rate of 93.84%while processing high-resolution images from the CelebA-HQ dataset, beating other evaluated models such asProgressive GAN 90.27%, CycleGAN 89.80%, and MGAN 80.80%. With improvements in accuracy, speed, andprivacy protection, the framework has great promise for practical use in a variety of fields that need face recognitiontechnology. This study paves the way for future research in privacy-enhanced face recognition systems, emphasizingthe significance of using cutting-edge technology to meet rising privacy issues in digital identity.展开更多
Due to the restricted satellite payloads in LEO mega-constellation networks(LMCNs),remote sensing image analysis,online learning and other big data services desirably need onboard distributed processing(OBDP).In exist...Due to the restricted satellite payloads in LEO mega-constellation networks(LMCNs),remote sensing image analysis,online learning and other big data services desirably need onboard distributed processing(OBDP).In existing technologies,the efficiency of big data applications(BDAs)in distributed systems hinges on the stable-state and low-latency links between worker nodes.However,LMCNs with high-dynamic nodes and long-distance links can not provide the above conditions,which makes the performance of OBDP hard to be intuitively measured.To bridge this gap,a multidimensional simulation platform is indispensable that can simulate the network environment of LMCNs and put BDAs in it for performance testing.Using STK's APIs and parallel computing framework,we achieve real-time simulation for thousands of satellite nodes,which are mapped as application nodes through software defined network(SDN)and container technologies.We elaborate the architecture and mechanism of the simulation platform,and take the Starlink and Hadoop as realistic examples for simulations.The results indicate that LMCNs have dynamic end-to-end latency which fluctuates periodically with the constellation movement.Compared to ground data center networks(GDCNs),LMCNs deteriorate the computing and storage job throughput,which can be alleviated by the utilization of erasure codes and data flow scheduling of worker nodes.展开更多
The flow shop scheduling problem is important for the manufacturing industry.Effective flow shop scheduling can bring great benefits to the industry.However,there are few types of research on Distributed Hybrid Flow S...The flow shop scheduling problem is important for the manufacturing industry.Effective flow shop scheduling can bring great benefits to the industry.However,there are few types of research on Distributed Hybrid Flow Shop Problems(DHFSP)by learning assisted meta-heuristics.This work addresses a DHFSP with minimizing the maximum completion time(Makespan).First,a mathematical model is developed for the concerned DHFSP.Second,four Q-learning-assisted meta-heuristics,e.g.,genetic algorithm(GA),artificial bee colony algorithm(ABC),particle swarm optimization(PSO),and differential evolution(DE),are proposed.According to the nature of DHFSP,six local search operations are designed for finding high-quality solutions in local space.Instead of randomselection,Q-learning assists meta-heuristics in choosing the appropriate local search operations during iterations.Finally,based on 60 cases,comprehensive numerical experiments are conducted to assess the effectiveness of the proposed algorithms.The experimental results and discussions prove that using Q-learning to select appropriate local search operations is more effective than the random strategy.To verify the competitiveness of the Q-learning assistedmeta-heuristics,they are compared with the improved iterated greedy algorithm(IIG),which is also for solving DHFSP.The Friedman test is executed on the results by five algorithms.It is concluded that the performance of four Q-learning-assisted meta-heuristics are better than IIG,and the Q-learning-assisted PSO shows the best competitiveness.展开更多
This paper is aimed at the distributed fault estimation issue associated with the potential loss of actuator efficiency for a type of discrete-time nonlinear systems with sensor saturation.For the distributed estimati...This paper is aimed at the distributed fault estimation issue associated with the potential loss of actuator efficiency for a type of discrete-time nonlinear systems with sensor saturation.For the distributed estimation structure under consideration,an estimation center is not necessary,and the estimator derives its information from itself and neighboring nodes,which fuses the state vector and the measurement vector.In an effort to cut down data conflicts in communication networks,the stochastic communication protocol(SCP)is employed so that the output signals from sensors can be selected.Additionally,a recursive security estimator scheme is created since attackers randomly inject malicious signals into the selected data.On this basis,sufficient conditions for a fault estimator with less conservatism are presented which ensure an upper bound of the estimation error covariance and the mean-square exponential boundedness of the estimating error.Finally,a numerical example is used to show the reliability and effectiveness of the considered distributed estimation algorithm.展开更多
This paper proposes a novel approach for identifying distributed dynamic loads in the time domain.Using polynomial andmodal analysis,the load is transformed intomodal space for coefficient identification.This allows t...This paper proposes a novel approach for identifying distributed dynamic loads in the time domain.Using polynomial andmodal analysis,the load is transformed intomodal space for coefficient identification.This allows the distributed dynamic load with a two-dimensional form in terms of time and space to be simultaneously identified in the form of modal force,thereby achieving dimensionality reduction.The Impulse-based Force Estimation Algorithm is proposed to identify dynamic loads in the time domain.Firstly,the algorithm establishes a recursion scheme based on convolution integral,enabling it to identify loads with a long history and rapidly changing forms over time.Secondly,the algorithm introduces moving mean and polynomial fitting to detrend,enhancing its applicability in load estimation.The aforementioned methodology successfully accomplishes the reconstruction of distributed,instead of centralized,dynamic loads on the continuum in the time domain by utilizing acceleration response.To validate the effectiveness of the method,computational and experimental verification were conducted.展开更多
Quantum error correction is a crucial technology for realizing quantum computers.These computers achieve faulttolerant quantum computing by detecting and correcting errors using decoding algorithms.Quantum error corre...Quantum error correction is a crucial technology for realizing quantum computers.These computers achieve faulttolerant quantum computing by detecting and correcting errors using decoding algorithms.Quantum error correction using neural network-based machine learning methods is a promising approach that is adapted to physical systems without the need to build noise models.In this paper,we use a distributed decoding strategy,which effectively alleviates the problem of exponential growth of the training set required for neural networks as the code distance of quantum error-correcting codes increases.Our decoding algorithm is based on renormalization group decoding and recurrent neural network decoder.The recurrent neural network is trained through the ResNet architecture to improve its decoding accuracy.Then we test the decoding performance of our distributed strategy decoder,recurrent neural network decoder,and the classic minimum weight perfect matching(MWPM)decoder for rotated surface codes with different code distances under the circuit noise model,the thresholds of these three decoders are about 0.0052,0.0051,and 0.0049,respectively.Our results demonstrate that the distributed strategy decoder outperforms the other two decoders,achieving approximately a 5%improvement in decoding efficiency compared to the MWPM decoder and approximately a 2%improvement compared to the recurrent neural network decoder.展开更多
The security of Federated Learning(FL)/Distributed Machine Learning(DML)is gravely threatened by data poisoning attacks,which destroy the usability of the model by contaminating training samples,so such attacks are ca...The security of Federated Learning(FL)/Distributed Machine Learning(DML)is gravely threatened by data poisoning attacks,which destroy the usability of the model by contaminating training samples,so such attacks are called causative availability indiscriminate attacks.Facing the problem that existing data sanitization methods are hard to apply to real-time applications due to their tedious process and heavy computations,we propose a new supervised batch detection method for poison,which can fleetly sanitize the training dataset before the local model training.We design a training dataset generation method that helps to enhance accuracy and uses data complexity features to train a detection model,which will be used in an efficient batch hierarchical detection process.Our model stockpiles knowledge about poison,which can be expanded by retraining to adapt to new attacks.Being neither attack-specific nor scenario-specific,our method is applicable to FL/DML or other online or offline scenarios.展开更多
False data injection attack(FDIA)is an attack that affects the stability of grid cyber-physical system(GCPS)by evading the detecting mechanism of bad data.Existing FDIA detection methods usually employ complex neural ...False data injection attack(FDIA)is an attack that affects the stability of grid cyber-physical system(GCPS)by evading the detecting mechanism of bad data.Existing FDIA detection methods usually employ complex neural networkmodels to detect FDIA attacks.However,they overlook the fact that FDIA attack samples at public-private network edges are extremely sparse,making it difficult for neural network models to obtain sufficient samples to construct a robust detection model.To address this problem,this paper designs an efficient sample generative adversarial model of FDIA attack in public-private network edge,which can effectively bypass the detectionmodel to threaten the power grid system.A generative adversarial network(GAN)framework is first constructed by combining residual networks(ResNet)with fully connected networks(FCN).Then,a sparse adversarial learning model is built by integrating the time-aligned data and normal data,which is used to learn the distribution characteristics between normal data and attack data through iterative confrontation.Furthermore,we introduce a Gaussian hybrid distributionmatrix by aggregating the network structure of attack data characteristics and normal data characteristics,which can connect and calculate FDIA data with normal characteristics.Finally,efficient FDIA attack samples can be sequentially generated through interactive adversarial learning.Extensive simulation experiments are conducted with IEEE 14-bus and IEEE 118-bus system data,and the results demonstrate that the generated attack samples of the proposed model can present superior performance compared to state-of-the-art models in terms of attack strength,robustness,and covert capability.展开更多
In this paper, the optimal variational generalized Nash equilibrium(v-GNE) seeking problem in merely monotone games with linearly coupled cost functions is investigated, in which the feasible strategy domain of each a...In this paper, the optimal variational generalized Nash equilibrium(v-GNE) seeking problem in merely monotone games with linearly coupled cost functions is investigated, in which the feasible strategy domain of each agent is coupled through an affine constraint. A distributed algorithm based on the hybrid steepest descent method is first proposed to seek the optimal v-GNE. Then, an accelerated algorithm with relaxation is proposed and analyzed, which has the potential to further improve the convergence speed to the optimal v-GNE. Some sufficient conditions in both algorithms are obtained to ensure the global convergence towards the optimal v-GNE. To illustrate the performance of the algorithms, numerical simulation is conducted based on a networked Nash-Cournot game with bounded market capacities.展开更多
Fiber-optic distributed strain sensing(FO-DSS)has been successful in monitoring strain changes along horizontal wellbores in hydraulically fractured reservoirs.However,the mechanism driving the various FO-DSS response...Fiber-optic distributed strain sensing(FO-DSS)has been successful in monitoring strain changes along horizontal wellbores in hydraulically fractured reservoirs.However,the mechanism driving the various FO-DSS responses associated with near-wellbore hydraulic fracture properties is still unclear.To address this knowledge gap,we use coupled wellbore-reservoir-geomechanics simulations to study measured strain-change behavior and infer hydraulic fracture characteristics.The crossflow among fractures is captured through explicit modeling of the transient wellbore flow.In addition,local grid refinement is applied to accurately capture strain changes along the fiber.A Base Case model was designed with four fractures of varying properties,simulating strain change signals when the production well is shut-in for 10 d after 240 d of production and reopened for 2 d.Strain-pressure plots for different fracture clusters were used to gain insights into inferring fracture properties using DSS data.When comparing the model with and without the wellbore,distinct strain change signals were observed,emphasizing the importance of incorporating the wellbore in FO-DSS modeling.The effects of fracture spacing and matrix permeability on strain change signals were thoroughly investigated.The results of our numerical study can improve the understanding of the relation between DSS signals and fracture hydraulic properties,thus maximizing the value of the dataset for fracture diagnostics and characterization.展开更多
The Internet of Things(IoT)is a growing technology that allows the sharing of data with other devices across wireless networks.Specifically,IoT systems are vulnerable to cyberattacks due to its opennes The proposed wo...The Internet of Things(IoT)is a growing technology that allows the sharing of data with other devices across wireless networks.Specifically,IoT systems are vulnerable to cyberattacks due to its opennes The proposed work intends to implement a new security framework for detecting the most specific and harmful intrusions in IoT networks.In this framework,a Covariance Linear Learning Embedding Selection(CL2ES)methodology is used at first to extract the features highly associated with the IoT intrusions.Then,the Kernel Distributed Bayes Classifier(KDBC)is created to forecast attacks based on the probability distribution value precisely.In addition,a unique Mongolian Gazellas Optimization(MGO)algorithm is used to optimize the weight value for the learning of the classifier.The effectiveness of the proposed CL2ES-KDBC framework has been assessed using several IoT cyber-attack datasets,The obtained results are then compared with current classification methods regarding accuracy(97%),precision(96.5%),and other factors.Computational analysis of the CL2ES-KDBC system on IoT intrusion datasets is performed,which provides valuable insight into its performance,efficiency,and suitability for securing IoT networks.展开更多
基金funding support from Rijkswaterstaat,the Netherlands,and European Union’s Horizon 2020 Research and Innovation Programme(Project SAFE-10-T under Grant No.723254)China Scholarship Council,and National Natural Science Foundation of China(Grant No.42225702).
文摘Distributed fiber optic sensors(DFOSs)possess the capability to measure strain and temperature variations over long distances,demonstrating outstanding potential for monitoring underground infrastructure.This study presents a state-of-the-art review of the DFOS applications for monitoring and assessing the deformation behavior of typical tunnel infrastructure,including bored tunnels,conventional tunnels,as well as immersed and cut-and-cover tunnels.DFOS systems based on Brillouin and Rayleigh scattering principles are both considered.When implementing DFOS monitoring,the fiber optic cable can be primarily installed along transverse and longitudinal directions to(1)measure distributed strains by continuously adhering the fiber to the structure’s surface or embedding it in the lining,or(2)measure point displacements by spot-anchoring it on the lining surface.There are four critical aspects of DFOS monitoring,including proper selection of the sensing fiber,selection of the measuring principle for the specific application,design of an effective sensor layout,and establishment of robust field sensor instrumentation.These four issues are comprehensively discussed,and practical suggestions are provided for the implementation of DFOS in tunnel infrastructure monitoring.
文摘In this paper, platoons of autonomous vehicles operating in urban road networks are considered. From a methodological point of view, the problem of interest consists of formally characterizing vehicle state trajectory tubes by means of routing decisions complying with traffic congestion criteria. To this end, a novel distributed control architecture is conceived by taking advantage of two methodologies: deep reinforcement learning and model predictive control. On one hand, the routing decisions are obtained by using a distributed reinforcement learning algorithm that exploits available traffic data at each road junction. On the other hand, a bank of model predictive controllers is in charge of computing the more adequate control action for each involved vehicle. Such tasks are here combined into a single framework:the deep reinforcement learning output(action) is translated into a set-point to be tracked by the model predictive controller;conversely, the current vehicle position, resulting from the application of the control move, is exploited by the deep reinforcement learning unit for improving its reliability. The main novelty of the proposed solution lies in its hybrid nature: on one hand it fully exploits deep reinforcement learning capabilities for decisionmaking purposes;on the other hand, time-varying hard constraints are always satisfied during the dynamical platoon evolution imposed by the computed routing decisions. To efficiently evaluate the performance of the proposed control architecture, a co-design procedure, involving the SUMO and MATLAB platforms, is implemented so that complex operating environments can be used, and the information coming from road maps(links,junctions, obstacles, semaphores, etc.) and vehicle state trajectories can be shared and exchanged. Finally by considering as operating scenario a real entire city block and a platoon of eleven vehicles described by double-integrator models, several simulations have been performed with the aim to put in light the main f eatures of the proposed approach. Moreover, it is important to underline that in different operating scenarios the proposed reinforcement learning scheme is capable of significantly reducing traffic congestion phenomena when compared with well-reputed competitors.
基金supported by the Sichuan Science and Technology Program(grant number 2022YFG0123).
文摘In this study,a novel residential virtual power plant(RVPP)scheduling method that leverages a gate recurrent unit(GRU)-integrated deep reinforcement learning(DRL)algorithm is proposed.In the proposed scheme,the GRU-integrated DRL algorithm guides the RVPP to participate effectively in both the day-ahead and real-time markets,lowering the electricity purchase costs and consumption risks for end-users.The Lagrangian relaxation technique is introduced to transform the constrained Markov decision process(CMDP)into an unconstrained optimization problem,which guarantees that the constraints are strictly satisfied without determining the penalty coefficients.Furthermore,to enhance the scalability of the constrained soft actor-critic(CSAC)-based RVPP scheduling approach,a fully distributed scheduling architecture was designed to enable plug-and-play in the residential distributed energy resources(RDER).Case studies performed on the constructed RVPP scenario validated the performance of the proposed methodology in enhancing the responsiveness of the RDER to power tariffs,balancing the supply and demand of the power grid,and ensuring customer comfort.
基金supported by the National Natural Science Foundation of China (NSFC)(62222308, 62173181, 62073171, 62221004)the Natural Science Foundation of Jiangsu Province (BK20200744, BK20220139)+3 种基金Jiangsu Specially-Appointed Professor (RK043STP19001)the Young Elite Scientists Sponsorship Program by CAST (2021QNRC001)1311 Talent Plan of Nanjing University of Posts and Telecommunicationsthe Fundamental Research Funds for the Central Universities (30920032203)。
文摘This paper is concerned with distributed Nash equi librium seeking strategies under quantized communication. In the proposed seeking strategy, a projection operator is synthesized with a gradient search method to achieve the optimization o players' objective functions while restricting their actions within required non-empty, convex and compact domains. In addition, a leader-following consensus protocol, in which quantized informa tion flows are utilized, is employed for information sharing among players. More specifically, logarithmic quantizers and uniform quantizers are investigated under both undirected and connected communication graphs and strongly connected digraphs, respec tively. Through Lyapunov stability analysis, it is shown that play ers' actions can be steered to a neighborhood of the Nash equilib rium with logarithmic and uniform quantizers, and the quanti fied convergence error depends on the parameter of the quan tizer for both undirected and directed cases. A numerical exam ple is given to verify the theoretical results.
基金National Natural Science Foundation of China(No.42374013)National Key Research and Development Program of China(Nos.2019YFC1509201,2021YFB3900604-03)。
文摘In the past two decades,extensive and in-depth research has been conducted on Time Series InSAR technology with the advancement of high-performance SAR satellites and the accumulation of big SAR data.The introduction of distributed scatterers in Distributed Scatterers InSAR(DS-InSAR)has significantly expanded the application scenarios of InSAR geodetic measurement by increasing the number of measurement points.This study traces the history of DS-InSAR,presents the definition and characteristics of distributed scatterers,and focuses on exploring the relationships and distinctions among proposed algorithms in two crucial steps:statistically homogeneous pixel selection and phase optimization.Additionally,the latest research progress in this field is tracked and the possible development direction in the future is discussed.Through simulation experiments and two real InSAR case studies,the proposed algorithms are compared and verified,and the advantages of DS-InSAR in deformation measurement practice are demonstrated.This work not only offers insights into current trends and focal points for theoretical research on DS-InSAR but also provides practical cases and guidance for applied research.
基金supported in part by the National Natural Science Foundation of China under Grant Nos.62173312,61922037,61873115,and 61803348in part by the National Major Scientific Instruments Development Project under Grant 61927807+6 种基金in part by the State Key Laboratory of Deep Buried Target Damage under Grant No.DXMBJJ2019-02in part by the Scientific and Technological Innovation Programs of Higher Education Institutions in Shanxi under Grant 2020L0266in part by the Shanxi Province Science Foundation for Youths under Grant No.201701D221123in part by the Youth Academic North University of China under Grant No.QX201803in part by the Program for the Innovative Talents of Higher Education Institutions of Shanxiin part by the Shanxi“1331Project”Key Subjects Construction under Grant 1331KSCin part by the Supported by Shanxi Province Science Foundation for Excellent Youths。
文摘This paper addresses a multicircular circumnavigation control for UAVs with desired angular spacing around a nonstationary target.By defining a coordinated error relative to neighboring angular spacing,under the premise that target information is perfectly accessible by all nodes,a centralized circular enclosing control strategy is derived for multiple UAVs connected by an undirected graph to allow for formation behaviors concerning the moving target.Besides,to avoid the requirement of target’s states being accessible for each UAV,fixed-time distributed observers are introduced to acquire the state estimates in a fixed-time sense,and the upper boundary of settling time can be determined offline irrespective of initial properties,greatly releasing the burdensome communication traffic.Then,with the aid of fixed-time distributed observers,a distributed circular circumnavigation controller is derived to force all UAVs to collaboratively evolve along the preset circles while keeping a desired angular spacing.It is inferred from Lyapunov stability that all errors are demonstrated to be convergent.Simulations are offered to verify the utility of proposed protocol.
基金the Deputyship for Research&Innovation,Ministry of Education in Saudi Arabia for funding this research work through the Project Number 223202.
文摘Recognizing human activity(HAR)from data in a smartphone sensor plays an important role in the field of health to prevent chronic diseases.Daily and weekly physical activities are recorded on the smartphone and tell the user whether he is moving well or not.Typically,smartphones and their associated sensing devices operate in distributed and unstable environments.Therefore,collecting their data and extracting useful information is a significant challenge.In this context,the aimof this paper is twofold:The first is to analyze human behavior based on the recognition of physical activities.Using the results of physical activity detection and classification,the second part aims to develop a health recommendation system to notify smartphone users about their healthy physical behavior related to their physical activities.This system is based on the calculation of calories burned by each user during physical activities.In this way,conclusions can be drawn about a person’s physical behavior by estimating the number of calories burned after evaluating data collected daily or even weekly following a series of physical workouts.To identify and classify human behavior our methodology is based on artificial intelligence models specifically deep learning techniques like Long Short-Term Memory(LSTM),stacked LSTM,and bidirectional LSTM.Since human activity data contains both spatial and temporal information,we proposed,in this paper,to use of an architecture allowing the extraction of the two types of information simultaneously.While Convolutional Neural Networks(CNN)has an architecture designed for spatial information,our idea is to combine CNN with LSTM to increase classification accuracy by taking into consideration the extraction of both spatial and temporal data.The results obtained achieved an accuracy of 96%.On the other side,the data learned by these algorithms is prone to error and uncertainty.To overcome this constraint and improve performance(96%),we proposed to use the fusion mechanisms.The last combines deep learning classifiers tomodel non-accurate and ambiguous data to obtain synthetic information to aid in decision-making.The Voting and Dempster-Shafer(DS)approaches are employed.The results showed that fused classifiers based on DS theory outperformed individual classifiers(96%)with the highest accuracy level of 98%.Also,the findings disclosed that participants engaging in physical activities are healthy,showcasing a disparity in the distribution of physical activities between men and women.
基金sponsored by the National Natural Science Foundation of China(Nos.61972208,62102194 and 62102196)National Natural Science Foundation of China(Youth Project)(No.62302237)+3 种基金Six Talent Peaks Project of Jiangsu Province(No.RJFW-111),China Postdoctoral Science Foundation Project(No.2018M640509)Postgraduate Research and Practice Innovation Program of Jiangsu Province(Nos.KYCX22_1019,KYCX23_1087,KYCX22_1027,KYCX23_1087,SJCX24_0339 and SJCX24_0346)Innovative Training Program for College Students of Nanjing University of Posts and Telecommunications(No.XZD2019116)Nanjing University of Posts and Telecommunications College Students Innovation Training Program(Nos.XZD2019116,XYB2019331).
文摘The scale and complexity of big data are growing continuously,posing severe challenges to traditional data processing methods,especially in the field of clustering analysis.To address this issue,this paper introduces a new method named Big Data Tensor Multi-Cluster Distributed Incremental Update(BDTMCDIncreUpdate),which combines distributed computing,storage technology,and incremental update techniques to provide an efficient and effective means for clustering analysis.Firstly,the original dataset is divided into multiple subblocks,and distributed computing resources are utilized to process the sub-blocks in parallel,enhancing efficiency.Then,initial clustering is performed on each sub-block using tensor-based multi-clustering techniques to obtain preliminary results.When new data arrives,incremental update technology is employed to update the core tensor and factor matrix,ensuring that the clustering model can adapt to changes in data.Finally,by combining the updated core tensor and factor matrix with historical computational results,refined clustering results are obtained,achieving real-time adaptation to dynamic data.Through experimental simulation on the Aminer dataset,the BDTMCDIncreUpdate method has demonstrated outstanding performance in terms of accuracy(ACC)and normalized mutual information(NMI)metrics,achieving an accuracy rate of 90%and an NMI score of 0.85,which outperforms existing methods such as TClusInitUpdate and TKLClusUpdate in most scenarios.Therefore,the BDTMCDIncreUpdate method offers an innovative solution to the field of big data analysis,integrating distributed computing,incremental updates,and tensor-based multi-clustering techniques.It not only improves the efficiency and scalability in processing large-scale high-dimensional datasets but also has been validated for its effectiveness and accuracy through experiments.This method shows great potential in real-world applications where dynamic data growth is common,and it is of significant importance for advancing the development of data analysis technology.
基金Project supported by the National Natural Science Foundation of China (Grant No.62073045)。
文摘We develop a policy of observer-based dynamic event-triggered state feedback control for distributed parameter systems over a mobile sensor-plus-actuator network.It is assumed that the mobile sensing devices that provide spatially averaged state measurements can be used to improve state estimation in the network.For the purpose of decreasing the update frequency of controller and unnecessary sampled data transmission, an efficient dynamic event-triggered control policy is constructed.In an event-triggered system, when an error signal exceeds a specified time-varying threshold, it indicates the occurrence of a typical event.The global asymptotic stability of the event-triggered closed-loop system and the boundedness of the minimum inter-event time can be guaranteed.Based on the linear quadratic optimal regulator, the actuator selects the optimal displacement only when an event occurs.A simulation example is finally used to verify that the effectiveness of such a control strategy can enhance the system performance.
文摘The use of privacy-enhanced facial recognition has increased in response to growing concerns about data securityand privacy in the digital age. This trend is spurred by rising demand for face recognition technology in a varietyof industries, including access control, law enforcement, surveillance, and internet communication. However,the growing usage of face recognition technology has created serious concerns about data monitoring and userprivacy preferences, especially in context-aware systems. In response to these problems, this study provides a novelframework that integrates sophisticated approaches such as Generative Adversarial Networks (GANs), Blockchain,and distributed computing to solve privacy concerns while maintaining exact face recognition. The framework’spainstaking design and execution strive to strike a compromise between precise face recognition and protectingpersonal data integrity in an increasingly interconnected environment. Using cutting-edge tools like Dlib for faceanalysis,Ray Cluster for distributed computing, and Blockchain for decentralized identity verification, the proposedsystem provides scalable and secure facial analysis while protecting user privacy. The study’s contributions includethe creation of a sustainable and scalable solution for privacy-aware face recognition, the implementation of flexibleprivacy computing approaches based on Blockchain networks, and the demonstration of higher performanceover previous methods. Specifically, the proposed StyleGAN model has an outstanding accuracy rate of 93.84%while processing high-resolution images from the CelebA-HQ dataset, beating other evaluated models such asProgressive GAN 90.27%, CycleGAN 89.80%, and MGAN 80.80%. With improvements in accuracy, speed, andprivacy protection, the framework has great promise for practical use in a variety of fields that need face recognitiontechnology. This study paves the way for future research in privacy-enhanced face recognition systems, emphasizingthe significance of using cutting-edge technology to meet rising privacy issues in digital identity.
基金supported by National Natural Sciences Foundation of China(No.62271165,62027802,62201307)the Guangdong Basic and Applied Basic Research Foundation(No.2023A1515030297)+2 种基金the Shenzhen Science and Technology Program ZDSYS20210623091808025Stable Support Plan Program GXWD20231129102638002the Major Key Project of PCL(No.PCL2024A01)。
文摘Due to the restricted satellite payloads in LEO mega-constellation networks(LMCNs),remote sensing image analysis,online learning and other big data services desirably need onboard distributed processing(OBDP).In existing technologies,the efficiency of big data applications(BDAs)in distributed systems hinges on the stable-state and low-latency links between worker nodes.However,LMCNs with high-dynamic nodes and long-distance links can not provide the above conditions,which makes the performance of OBDP hard to be intuitively measured.To bridge this gap,a multidimensional simulation platform is indispensable that can simulate the network environment of LMCNs and put BDAs in it for performance testing.Using STK's APIs and parallel computing framework,we achieve real-time simulation for thousands of satellite nodes,which are mapped as application nodes through software defined network(SDN)and container technologies.We elaborate the architecture and mechanism of the simulation platform,and take the Starlink and Hadoop as realistic examples for simulations.The results indicate that LMCNs have dynamic end-to-end latency which fluctuates periodically with the constellation movement.Compared to ground data center networks(GDCNs),LMCNs deteriorate the computing and storage job throughput,which can be alleviated by the utilization of erasure codes and data flow scheduling of worker nodes.
基金partially supported by the Guangdong Basic and Applied Basic Research Foundation(2023A1515011531)the National Natural Science Foundation of China under Grant 62173356+2 种基金the Science and Technology Development Fund(FDCT),Macao SAR,under Grant 0019/2021/AZhuhai Industry-University-Research Project with Hongkong and Macao under Grant ZH22017002210014PWCthe Key Technologies for Scheduling and Optimization of Complex Distributed Manufacturing Systems(22JR10KA007).
文摘The flow shop scheduling problem is important for the manufacturing industry.Effective flow shop scheduling can bring great benefits to the industry.However,there are few types of research on Distributed Hybrid Flow Shop Problems(DHFSP)by learning assisted meta-heuristics.This work addresses a DHFSP with minimizing the maximum completion time(Makespan).First,a mathematical model is developed for the concerned DHFSP.Second,four Q-learning-assisted meta-heuristics,e.g.,genetic algorithm(GA),artificial bee colony algorithm(ABC),particle swarm optimization(PSO),and differential evolution(DE),are proposed.According to the nature of DHFSP,six local search operations are designed for finding high-quality solutions in local space.Instead of randomselection,Q-learning assists meta-heuristics in choosing the appropriate local search operations during iterations.Finally,based on 60 cases,comprehensive numerical experiments are conducted to assess the effectiveness of the proposed algorithms.The experimental results and discussions prove that using Q-learning to select appropriate local search operations is more effective than the random strategy.To verify the competitiveness of the Q-learning assistedmeta-heuristics,they are compared with the improved iterated greedy algorithm(IIG),which is also for solving DHFSP.The Friedman test is executed on the results by five algorithms.It is concluded that the performance of four Q-learning-assisted meta-heuristics are better than IIG,and the Q-learning-assisted PSO shows the best competitiveness.
基金supported in part by the National Natural Science Foundation of China(62073189,62173207)the Taishan Scholar Project of Shandong Province(tsqn202211129)。
文摘This paper is aimed at the distributed fault estimation issue associated with the potential loss of actuator efficiency for a type of discrete-time nonlinear systems with sensor saturation.For the distributed estimation structure under consideration,an estimation center is not necessary,and the estimator derives its information from itself and neighboring nodes,which fuses the state vector and the measurement vector.In an effort to cut down data conflicts in communication networks,the stochastic communication protocol(SCP)is employed so that the output signals from sensors can be selected.Additionally,a recursive security estimator scheme is created since attackers randomly inject malicious signals into the selected data.On this basis,sufficient conditions for a fault estimator with less conservatism are presented which ensure an upper bound of the estimation error covariance and the mean-square exponential boundedness of the estimating error.Finally,a numerical example is used to show the reliability and effectiveness of the considered distributed estimation algorithm.
文摘This paper proposes a novel approach for identifying distributed dynamic loads in the time domain.Using polynomial andmodal analysis,the load is transformed intomodal space for coefficient identification.This allows the distributed dynamic load with a two-dimensional form in terms of time and space to be simultaneously identified in the form of modal force,thereby achieving dimensionality reduction.The Impulse-based Force Estimation Algorithm is proposed to identify dynamic loads in the time domain.Firstly,the algorithm establishes a recursion scheme based on convolution integral,enabling it to identify loads with a long history and rapidly changing forms over time.Secondly,the algorithm introduces moving mean and polynomial fitting to detrend,enhancing its applicability in load estimation.The aforementioned methodology successfully accomplishes the reconstruction of distributed,instead of centralized,dynamic loads on the continuum in the time domain by utilizing acceleration response.To validate the effectiveness of the method,computational and experimental verification were conducted.
基金Project supported by Natural Science Foundation of Shandong Province,China (Grant Nos.ZR2021MF049,ZR2022LLZ012,and ZR2021LLZ001)。
文摘Quantum error correction is a crucial technology for realizing quantum computers.These computers achieve faulttolerant quantum computing by detecting and correcting errors using decoding algorithms.Quantum error correction using neural network-based machine learning methods is a promising approach that is adapted to physical systems without the need to build noise models.In this paper,we use a distributed decoding strategy,which effectively alleviates the problem of exponential growth of the training set required for neural networks as the code distance of quantum error-correcting codes increases.Our decoding algorithm is based on renormalization group decoding and recurrent neural network decoder.The recurrent neural network is trained through the ResNet architecture to improve its decoding accuracy.Then we test the decoding performance of our distributed strategy decoder,recurrent neural network decoder,and the classic minimum weight perfect matching(MWPM)decoder for rotated surface codes with different code distances under the circuit noise model,the thresholds of these three decoders are about 0.0052,0.0051,and 0.0049,respectively.Our results demonstrate that the distributed strategy decoder outperforms the other two decoders,achieving approximately a 5%improvement in decoding efficiency compared to the MWPM decoder and approximately a 2%improvement compared to the recurrent neural network decoder.
基金supported in part by the“Pioneer”and“Leading Goose”R&D Program of Zhejiang(Grant No.2022C03174)the National Natural Science Foundation of China(No.92067103)+4 种基金the Key Research and Development Program of Shaanxi,China(No.2021ZDLGY06-02)the Natural Science Foundation of Shaanxi Province(No.2019ZDLGY12-02)the Shaanxi Innovation Team Project(No.2018TD-007)the Xi'an Science and technology Innovation Plan(No.201809168CX9JC10)the Fundamental Research Funds for the Central Universities(No.YJS2212)and National 111 Program of China B16037.
文摘The security of Federated Learning(FL)/Distributed Machine Learning(DML)is gravely threatened by data poisoning attacks,which destroy the usability of the model by contaminating training samples,so such attacks are called causative availability indiscriminate attacks.Facing the problem that existing data sanitization methods are hard to apply to real-time applications due to their tedious process and heavy computations,we propose a new supervised batch detection method for poison,which can fleetly sanitize the training dataset before the local model training.We design a training dataset generation method that helps to enhance accuracy and uses data complexity features to train a detection model,which will be used in an efficient batch hierarchical detection process.Our model stockpiles knowledge about poison,which can be expanded by retraining to adapt to new attacks.Being neither attack-specific nor scenario-specific,our method is applicable to FL/DML or other online or offline scenarios.
基金supported in part by the the Natural Science Foundation of Shanghai(20ZR1421600)Research Fund of Guangxi Key Lab of Multi-Source Information Mining&Security(MIMS21-M-02).
文摘False data injection attack(FDIA)is an attack that affects the stability of grid cyber-physical system(GCPS)by evading the detecting mechanism of bad data.Existing FDIA detection methods usually employ complex neural networkmodels to detect FDIA attacks.However,they overlook the fact that FDIA attack samples at public-private network edges are extremely sparse,making it difficult for neural network models to obtain sufficient samples to construct a robust detection model.To address this problem,this paper designs an efficient sample generative adversarial model of FDIA attack in public-private network edge,which can effectively bypass the detectionmodel to threaten the power grid system.A generative adversarial network(GAN)framework is first constructed by combining residual networks(ResNet)with fully connected networks(FCN).Then,a sparse adversarial learning model is built by integrating the time-aligned data and normal data,which is used to learn the distribution characteristics between normal data and attack data through iterative confrontation.Furthermore,we introduce a Gaussian hybrid distributionmatrix by aggregating the network structure of attack data characteristics and normal data characteristics,which can connect and calculate FDIA data with normal characteristics.Finally,efficient FDIA attack samples can be sequentially generated through interactive adversarial learning.Extensive simulation experiments are conducted with IEEE 14-bus and IEEE 118-bus system data,and the results demonstrate that the generated attack samples of the proposed model can present superior performance compared to state-of-the-art models in terms of attack strength,robustness,and covert capability.
基金supported by the National Natural Science Foundation of China(Basic Science Center Program)(61988101)the Joint Fund of Ministry of Education for Equipment Pre-research (8091B022234)+3 种基金Shanghai International Science and Technology Cooperation Program (21550712400)Shanghai Pilot Program for Basic Research (22TQ1400100-3)the Fundamental Research Funds for the Central UniversitiesShanghai Artifcial Intelligence Laboratory。
文摘In this paper, the optimal variational generalized Nash equilibrium(v-GNE) seeking problem in merely monotone games with linearly coupled cost functions is investigated, in which the feasible strategy domain of each agent is coupled through an affine constraint. A distributed algorithm based on the hybrid steepest descent method is first proposed to seek the optimal v-GNE. Then, an accelerated algorithm with relaxation is proposed and analyzed, which has the potential to further improve the convergence speed to the optimal v-GNE. Some sufficient conditions in both algorithms are obtained to ensure the global convergence towards the optimal v-GNE. To illustrate the performance of the algorithms, numerical simulation is conducted based on a networked Nash-Cournot game with bounded market capacities.
基金funding support from the National Natural Science Foundation of China(Grant No.52204030)Youth Innovation and Technology Support Program for Higher Education Institutions of Shandong Province,China(Grant No.2022KJ070)the National Natural Science Foundation of China Enterprise Innovation and Development Joint Fund Project(Grant No.U19B6003).
文摘Fiber-optic distributed strain sensing(FO-DSS)has been successful in monitoring strain changes along horizontal wellbores in hydraulically fractured reservoirs.However,the mechanism driving the various FO-DSS responses associated with near-wellbore hydraulic fracture properties is still unclear.To address this knowledge gap,we use coupled wellbore-reservoir-geomechanics simulations to study measured strain-change behavior and infer hydraulic fracture characteristics.The crossflow among fractures is captured through explicit modeling of the transient wellbore flow.In addition,local grid refinement is applied to accurately capture strain changes along the fiber.A Base Case model was designed with four fractures of varying properties,simulating strain change signals when the production well is shut-in for 10 d after 240 d of production and reopened for 2 d.Strain-pressure plots for different fracture clusters were used to gain insights into inferring fracture properties using DSS data.When comparing the model with and without the wellbore,distinct strain change signals were observed,emphasizing the importance of incorporating the wellbore in FO-DSS modeling.The effects of fracture spacing and matrix permeability on strain change signals were thoroughly investigated.The results of our numerical study can improve the understanding of the relation between DSS signals and fracture hydraulic properties,thus maximizing the value of the dataset for fracture diagnostics and characterization.
文摘The Internet of Things(IoT)is a growing technology that allows the sharing of data with other devices across wireless networks.Specifically,IoT systems are vulnerable to cyberattacks due to its opennes The proposed work intends to implement a new security framework for detecting the most specific and harmful intrusions in IoT networks.In this framework,a Covariance Linear Learning Embedding Selection(CL2ES)methodology is used at first to extract the features highly associated with the IoT intrusions.Then,the Kernel Distributed Bayes Classifier(KDBC)is created to forecast attacks based on the probability distribution value precisely.In addition,a unique Mongolian Gazellas Optimization(MGO)algorithm is used to optimize the weight value for the learning of the classifier.The effectiveness of the proposed CL2ES-KDBC framework has been assessed using several IoT cyber-attack datasets,The obtained results are then compared with current classification methods regarding accuracy(97%),precision(96.5%),and other factors.Computational analysis of the CL2ES-KDBC system on IoT intrusion datasets is performed,which provides valuable insight into its performance,efficiency,and suitability for securing IoT networks.