Mg-Al alloys have excellent strength and ductility but relatively low thermal conductivity due to Al addition.The accurate prediction of thermal conductivity is a prerequisite for designing Mg-Al alloys with high ther...Mg-Al alloys have excellent strength and ductility but relatively low thermal conductivity due to Al addition.The accurate prediction of thermal conductivity is a prerequisite for designing Mg-Al alloys with high thermal conductivity.Thus,databases for predicting temperature-and composition-dependent thermal conductivities must be established.In this study,Mg-Al-La alloys with different contents of Al2La,Al3La,and Al11La3phases and solid solubility of Al in the α-Mg phase were designed.The influence of the second phase(s) and Al solid solubility on thermal conductivity was investigated.Experimental results revealed a second phase transformation from Al_(2)La to Al_(3)La and further to Al_(11)La_(3)with the increasing Al content at a constant La amount.The degree of the negative effect of the second phase(s) on thermal diffusivity followed the sequence of Al2La>Al3La>Al_(11)La_(3).Compared with the second phase,an increase in the solid solubility of Al in α-Mg remarkably reduced the thermal conductivity.On the basis of the experimental data,a database of the reciprocal thermal diffusivity of the Mg-Al-La system was established by calculation of the phase diagram (CALPHAD)method.With a standard error of±1.2 W/(m·K),the predicted results were in good agreement with the experimental data.The established database can be used to design Mg-Al alloys with high thermal conductivity and provide valuable guidance for expanding their application prospects.展开更多
Most of the neural network architectures are based on human experience,which requires a long and tedious trial-and-error process.Neural architecture search(NAS)attempts to detect effective architectures without human ...Most of the neural network architectures are based on human experience,which requires a long and tedious trial-and-error process.Neural architecture search(NAS)attempts to detect effective architectures without human intervention.Evolutionary algorithms(EAs)for NAS can find better solutions than human-designed architectures by exploring a large search space for possible architectures.Using multiobjective EAs for NAS,optimal neural architectures that meet various performance criteria can be explored and discovered efficiently.Furthermore,hardware-accelerated NAS methods can improve the efficiency of the NAS.While existing reviews have mainly focused on different strategies to complete NAS,a few studies have explored the use of EAs for NAS.In this paper,we summarize and explore the use of EAs for NAS,as well as large-scale multiobjective optimization strategies and hardware-accelerated NAS methods.NAS performs well in healthcare applications,such as medical image analysis,classification of disease diagnosis,and health monitoring.EAs for NAS can automate the search process and optimize multiple objectives simultaneously in a given healthcare task.Deep neural network has been successfully used in healthcare,but it lacks interpretability.Medical data is highly sensitive,and privacy leaks are frequently reported in the healthcare industry.To solve these problems,in healthcare,we propose an interpretable neuroevolution framework based on federated learning to address search efficiency and privacy protection.Moreover,we also point out future research directions for evolutionary NAS.Overall,for researchers who want to use EAs to optimize NNs in healthcare,we analyze the advantages and disadvantages of doing so to provide detailed guidance,and propose an interpretable privacy-preserving framework for healthcare applications.展开更多
The interconnection bottleneck caused by limitations of cable number, inner space and cooling power of dilution refrigerators has been an outstanding challenge for building scalable superconducting quantum computers w...The interconnection bottleneck caused by limitations of cable number, inner space and cooling power of dilution refrigerators has been an outstanding challenge for building scalable superconducting quantum computers with the increasing number of qubits in quantum processors. To surmount such an obstacle, it is desirable to integrate qubits with quantum–classical interface(QCI) circuits based on rapid single flux quantum(RSFQ) circuits. In this work, a digital flux tuner for qubits(DFTQ) is proposed for manipulating flux of qubits as a crucial part of the interface circuit. A schematic diagram of the DFTQ is presented, consisting of a coarse tuning unit and a fine-tuning unit for providing magnetic flux with different precision to qubits. The method of using DFTQ to provide flux for gate operations is discussed from the optimization of circuit design and input signal. To verify the effectiveness of the method, simulations of a single DFTQ and quantum gates including a Z gate and an iSWAP gate with DFTQs are performed for flux-tunable transmons. The quantum process tomography corresponding to the two gates is also carried out to analyze the sources of gate error. The results of tomography show that the gate fidelities independent of the initial states of the Z gate and the iSWAP gate are 99.935% and 99.676%,respectively. With DFTQs inside, the QCI would be a powerful tool for building large-scale quantum computers.展开更多
The single-shot readout data process is essential for the realization of high-fidelity qubits and fault-tolerant quantum algorithms in semiconductor quantum dots. However, the fidelity and visibility of the readout pr...The single-shot readout data process is essential for the realization of high-fidelity qubits and fault-tolerant quantum algorithms in semiconductor quantum dots. However, the fidelity and visibility of the readout process are sensitive to the choice of the thresholds and limited by the experimental hardware. By demonstrating the linear dependence between the measured spin state probabilities and readout visibilities along with dark counts, we describe an alternative threshold-independent method for the single-shot readout of spin qubits in semiconductor quantum dots. We can obtain the extrapolated spin state probabilities of the prepared probabilities of the excited spin state through the threshold-independent method. We then analyze the corresponding errors of the method, finding that errors of the extrapolated probabilities cannot be neglected with no constraints on the readout time and threshold voltage. Therefore, by limiting the readout time and threshold voltage, we ensure the accuracy of the extrapolated probability. We then prove that the efficiency and robustness of this method are 60 times larger than those of the most commonly used method. Moreover, we discuss the influence of the electron temperature on the effective area with a fixed external magnetic field and provide a preliminary demonstration for a single-shot readout of up to 0.7K/1.5T in the future.展开更多
Electron spins confined in semiconductor quantum dots(QDs)are one of potential candidates for physical implementation of scalable quantum information processing technologies.Tunnel coupling based inter exchange intera...Electron spins confined in semiconductor quantum dots(QDs)are one of potential candidates for physical implementation of scalable quantum information processing technologies.Tunnel coupling based inter exchange interaction between QDs is crucial in achieving single-qubit manipulation,two-qubit gate,quantum communication and quantum simulation.This review first provides a theoretical perspective that surveys a general framework,including the Helter−London approach,the Hund−Mulliken approach,and the Hubbard model,to describe the inter exchange interactions between semiconductor quantum dots.An electrical method to control the inter exchange interaction in a realistic device is proposed as well.Then the significant achievements of inter exchange interaction in manipulating single qubits,achieving two-qubit gates,performing quantum communication and quantum simulation are reviewed.The last part is a summary of this review.展开更多
The growing development of the Internet of Things(IoT)is accelerating the emergence and growth of new IoT services and applications,which will result in massive amounts of data being generated,transmitted and pro-cess...The growing development of the Internet of Things(IoT)is accelerating the emergence and growth of new IoT services and applications,which will result in massive amounts of data being generated,transmitted and pro-cessed in wireless communication networks.Mobile Edge Computing(MEC)is a desired paradigm to timely process the data from IoT for value maximization.In MEC,a number of computing-capable devices are deployed at the network edge near data sources to support edge computing,such that the long network transmission delay in cloud computing paradigm could be avoided.Since an edge device might not always have sufficient resources to process the massive amount of data,computation offloading is significantly important considering the coop-eration among edge devices.However,the dynamic traffic characteristics and heterogeneous computing capa-bilities of edge devices challenge the offloading.In addition,different scheduling schemes might provide different computation delays to the offloaded tasks.Thus,offloading in mobile nodes and scheduling in the MEC server are coupled to determine service delay.This paper seeks to guarantee low delay for computation intensive applica-tions by jointly optimizing the offloading and scheduling in such an MEC system.We propose a Delay-Greedy Computation Offloading(DGCO)algorithm to make offloading decisions for new tasks in distributed computing-enabled mobile devices.A Reinforcement Learning-based Parallel Scheduling(RLPS)algorithm is further designed to schedule offloaded tasks in the multi-core MEC server.With an offloading delay broadcast mechanism,the DGCO and RLPS cooperate to achieve the goal of delay-guarantee-ratio maximization.Finally,the simulation results show that our proposal can bound the end-to-end delay of various tasks.Even under slightly heavy task load,the delay-guarantee-ratio given by DGCO-RLPS can still approximate 95%,while that given by benchmarked algorithms is reduced to intolerable value.The simulation results are demonstrated the effective-ness of DGCO-RLPS for delay guarantee in MEC.展开更多
On the basis of computational fluid dynamics,the flow field characteristics of multi-trophic artificial reefs,including the flow field distribution features of a single reef under three different velocities and the ef...On the basis of computational fluid dynamics,the flow field characteristics of multi-trophic artificial reefs,including the flow field distribution features of a single reef under three different velocities and the effect of spacing between reefs on flow scale and the flow state,were analyzed.Results indicate upwelling,slow flow,and eddy around a single reef.Maximum velocity,height,and volume of upwelling in front of a single reef were positively correlated with inflow velocity.The length and volume of slow flow increased with the increase in inflow velocity.Eddies were present both inside and backward,and vorticity was positively correlated with inflow velocity.Space between reefs had a minor influence on the maximum velocity and height of upwelling.With the increase in space from 0.5 L to 1.5 L(L is the reef lehgth),the length of slow flow in the front and back of the combined reefs increased slightly.When the space was 2.0 L,the length of the slow flow decreased.In four different spaces,eddies were present inside and at the back of each reef.The maximum vorticity was negatively correlated with space from 0.5 L to 1.5 L,but under 2.0 L space,the maximum vorticity was close to the vorticity of a single reef under the same inflow velocity.展开更多
This paper presents a novel approach to proxy blind signatures in the realm of quantum circuits,aiming to enhance security while safeguarding sensitive information.The main objective of this research is to introduce a...This paper presents a novel approach to proxy blind signatures in the realm of quantum circuits,aiming to enhance security while safeguarding sensitive information.The main objective of this research is to introduce a quantum proxy blind signature(QPBS)protocol that utilizes quantum logical gates and quantum measurement techniques.The QPBS protocol is constructed by the initial phase,proximal blinding message phase,remote authorization and signature phase,remote validation,and de-blinding phase.This innovative design ensures a secure mechanism for signing documents without revealing the content to the proxy signer,providing practical security authentication in a quantum environment under the assumption that the CNOT gates are securely implemented.Unlike existing approaches,our proposed QPBS protocol eliminates the need for quantum entanglement preparation,thus simplifying the implementation process.To assess the effectiveness and robustness of the QPBS protocol,we conduct comprehensive simulation studies in both ideal and noisy quantum environments on the IBM quantum cloud platform.The results demonstrate the superior performance of the QPBS algorithm,highlighting its resilience against repudiation and forgeability,which are key security concerns in the realm of proxy blind signatures.Furthermore,we have established authentic security thresholds(82.102%)in the presence of real noise,thereby emphasizing the practicality of our proposed solution.展开更多
Practical real-world scenarios such as the Internet,social networks,and biological networks present the challenges of data scarcity and complex correlations,which limit the applications of artificial intelligence.The ...Practical real-world scenarios such as the Internet,social networks,and biological networks present the challenges of data scarcity and complex correlations,which limit the applications of artificial intelligence.The graph structure is a typical tool used to formulate such correlations,it is incapable of modeling highorder correlations among different objects in systems;thus,the graph structure cannot fully convey the intricate correlations among objects.Confronted with the aforementioned two challenges,hypergraph computation models high-order correlations among data,knowledge,and rules through hyperedges and leverages these high-order correlations to enhance the data.Additionally,hypergraph computation achieves collaborative computation using data and high-order correlations,thereby offering greater modeling flexibility.In particular,we introduce three types of hypergraph computation methods:①hypergraph structure modeling,②hypergraph semantic computing,and③efficient hypergraph computing.We then specify how to adopt hypergraph computation in practice by focusing on specific tasks such as three-dimensional(3D)object recognition,revealing that hypergraph computation can reduce the data requirement by 80%while achieving comparable performance or improve the performance by 52%given the same data,compared with a traditional data-based method.A comprehensive overview of the applications of hypergraph computation in diverse domains,such as intelligent medicine and computer vision,is also provided.Finally,we introduce an open-source deep learning library,DeepHypergraph(DHG),which can serve as a tool for the practical usage of hypergraph computation.展开更多
An efficient approach is proposed for the equivalent linearization of frame structures with plastic hinges under nonstationary seismic excitations.The concentrated plastic hinges,described by the Bouc-Wen model,are as...An efficient approach is proposed for the equivalent linearization of frame structures with plastic hinges under nonstationary seismic excitations.The concentrated plastic hinges,described by the Bouc-Wen model,are assumed to occur at the two ends of a linear-elastic beam element.The auxiliary differential equations governing the plastic rotational displacements and their corresponding hysteretic displacements are replaced with linearized differential equations.Then,the two sets of equations of motion for the original nonlinear system can be reduced to an expanded-order equivalent linearized equation of motion for equivalent linear systems.To solve the equation of motion for equivalent linear systems,the nonstationary random vibration analysis is carried out based on the explicit time-domain method with high efficiency.Finally,the proposed treatment method for initial values of equivalent parameters is investigated in conjunction with parallel computing technology,which provides a new way of obtaining the equivalent linear systems at different time instants.Based on the explicit time-domain method,the key responses of interest of the converged equivalent linear system can be calculated through dimension reduction analysis with high efficiency.Numerical examples indicate that the proposed approach has high computational efficiency,and shows good applicability to weak nonlinear and medium-intensity nonlinear systems.展开更多
Although predictor-corrector methods have been extensively applied,they might not meet the requirements of practical applications and engineering tasks,particularly when high accuracy and efficiency are necessary.A no...Although predictor-corrector methods have been extensively applied,they might not meet the requirements of practical applications and engineering tasks,particularly when high accuracy and efficiency are necessary.A novel class of correctors based on feedback-accelerated Picard iteration(FAPI)is proposed to further enhance computational performance.With optimal feedback terms that do not require inversion of matrices,significantly faster convergence speed and higher numerical accuracy are achieved by these correctors compared with their counterparts;however,the computational complexities are comparably low.These advantages enable nonlinear engineering problems to be solved quickly and accurately,even with rough initial guesses from elementary predictors.The proposed method offers flexibility,enabling the use of the generated correctors for either bulk processing of collocation nodes in a domain or successive corrections of a single node in a finite difference approach.In our method,the functional formulas of FAPI are discretized into numerical forms using the collocation approach.These collocated iteration formulas can directly solve nonlinear problems,but they may require significant computational resources because of the manipulation of high-dimensionalmatrices.To address this,the collocated iteration formulas are further converted into finite difference forms,enabling the design of lightweight predictor-corrector algorithms for real-time computation.The generality of the proposed method is illustrated by deriving new correctors for three commonly employed finite-difference approaches:the modified Euler approach,the Adams-Bashforth-Moulton approach,and the implicit Runge-Kutta approach.Subsequently,the updated approaches are tested in solving strongly nonlinear problems,including the Matthieu equation,the Duffing equation,and the low-earth-orbit tracking problem.The numerical findings confirm the computational accuracy and efficiency of the derived predictor-corrector algorithms.展开更多
Owing to the complex lithology of unconventional reservoirs,field interpreters usually need to provide a basis for interpretation using logging simulation models.Among the various detection tools that use nuclear sour...Owing to the complex lithology of unconventional reservoirs,field interpreters usually need to provide a basis for interpretation using logging simulation models.Among the various detection tools that use nuclear sources,the detector response can reflect various types of information of the medium.The Monte Carlo method is one of the primary methods used to obtain nuclear detection responses in complex environments.However,this requires a computational process with extensive random sampling,consumes considerable resources,and does not provide real-time response results.Therefore,a novel fast forward computational method(FFCM)for nuclear measurement that uses volumetric detection constraints to rapidly calculate the detector response in various complex environments is proposed.First,the data library required for the FFCM is built by collecting the detection volume,detector counts,and flux sensitivity functions through a Monte Carlo simulation.Then,based on perturbation theory and the Rytov approximation,a model for the detector response is derived using the flux sensitivity function method and a one-group diffusion model.The environmental perturbation is constrained to optimize the model according to the tool structure and the impact of the formation and borehole within the effective detection volume.Finally,the method is applied to a neutron porosity tool for verification.In various complex simulation environments,the maximum relative error between the calculated porosity results of Monte Carlo and FFCM was 6.80%,with a rootmean-square error of 0.62 p.u.In field well applications,the formation porosity model obtained using FFCM was in good agreement with the model obtained by interpreters,which demonstrates the validity and accuracy of the proposed method.展开更多
Rechargeable Zn-air batteries(ZAB)have drawn extensive attention due to their eco-friendliness and safety.However,the lack of high-performance and low-cost oxygen redox reactions(OER and ORR)catalysts has become one o...Rechargeable Zn-air batteries(ZAB)have drawn extensive attention due to their eco-friendliness and safety.However,the lack of high-performance and low-cost oxygen redox reactions(OER and ORR)catalysts has become one of the main stumbling blocks in their development.Herein,we successfully fabricate a CoFe nanobubble encapsulated in nitrogen-doped carbon nanocage on wood carbon support(CoFe@NC/WC)via pyrolysis of a novel Prussian blue analog(PBA)/spruce precursor.The hierarchical CoFe@NC/WC catalyst exhibits an excellent potential difference of 0.74 V between the OER potential at 10 mA cm^(-2)and half-wave potential of ORR in 0.1 M KOH,comparable to recently reported preeminent electrocatalysts.Further,CoFe@NC/WC shows outstanding electrochemical performance in liquid ZAB,with a peak power density of 138.9 mW cm^(-2)and a specific capacity of 763.5 mAh g^(-1).More importantly,a bacterial cellulose nanofiber reinforced polyacrylic acid(BC-PAA)hydrogel electrolyte shows ultrahigh tensile-breaking stress of 1.58 MPa.In conjunction with the as-prepared CoFe@NC/WC catalyst,BC-PAA-based wearable ZAB displays impressive rechargeability and foldability,and can power portable electronics,such as electronic timer and mobile phone,in bent states.This work provides a new approach toward high-activity and low-cost catalysts for ZAB.展开更多
Deep learning has emerged in many practical applications,such as image classification,fault diagnosis,and object detection.More recently,convolutional neural networks(CNNs),representative models of deep learning,have ...Deep learning has emerged in many practical applications,such as image classification,fault diagnosis,and object detection.More recently,convolutional neural networks(CNNs),representative models of deep learning,have been used to solve fault detection.However,the current design of CNNs for fault detection of wind turbine blades is highly dependent on domain knowledge and requires a large amount of trial and error.For this reason,an evolutionary YOLOv8 network has been developed to automatically find the network architecture for wind turbine blade-based fault detection.YOLOv8 is a CNN-backed object detection model.Specifically,to reduce the parameter count,we first design an improved FasterNet module based on the Partial Convolution(PConv)operator.Then,to enhance convergence performance,we improve the loss function based on the efficient complete intersection over the union.Based on this,a flexible variable-length encoding is proposed,and the corresponding reproduction operators are designed.Related experimental results confirmthat the proposed approach can achieve better fault detection results and improve by 2.6%in mean precision at 50(mAP50)compared to the existing methods.Additionally,compared to training with the YOLOv8n model,the YOLOBFE model reduces the training parameters by 933,937 and decreases the GFLOPS(Giga Floating Point Operations Per Second)by 1.1.展开更多
Majorana quantum computation offers a potential approach to securely manipulating and storing quantum data in a topological manner that may effectively resist the decoherence induced by local noise. However, actual Ma...Majorana quantum computation offers a potential approach to securely manipulating and storing quantum data in a topological manner that may effectively resist the decoherence induced by local noise. However, actual Majorana qubit setups are susceptible to noise. In this study, from a quantum dynamics perspective, we develop a noise model for Majorana qubits that accounts for quasi-particle poisoning and Majorana overlapping with fluctuation. Furthermore, we focus on Majorana parity readout methodologies, specifically those leveraging an ancillary quantum dot, and carry out an indepth exploration of continuous measurement techniques founded on the quantum jump model of a quantum point contact.Utilizing these methodologies, we proceed to analyze the influence of noise on the afore-mentioned noise model, employing numerical computation to evaluate the power spectrum and frequency curve. In the culmination of our study, we put forward a strategy to benchmark the presence and detailed properties of noise in Majorana qubits.展开更多
Neuromorphic computing,inspired by the human brain,uses memristor devices for complex tasks.Recent studies show that self-organizing random nanowires can implement neuromorphic information processing,enabling data ana...Neuromorphic computing,inspired by the human brain,uses memristor devices for complex tasks.Recent studies show that self-organizing random nanowires can implement neuromorphic information processing,enabling data analysis.This paper presents a model based on these nanowire networks,with an improved conductance variation profile.We suggest using these networks for temporal information processing via a reservoir computing scheme and propose an efficient data encoding method using voltage pulses.The nanowire network layer generates dynamic behaviors for pulse voltages,allowing time series prediction analysis.Our experiment uses a double stochastic nanowire network architecture for processing multiple input signals,outperforming traditional reservoir computing in terms of fewer nodes,enriched dynamics and improved prediction accuracy.Experimental results confirm the high accuracy of this architecture on multiple real-time series datasets,making neuromorphic nanowire networks promising for physical implementation of reservoir computing.展开更多
This paper presents a comprehensive exploration into the integration of Internet of Things(IoT),big data analysis,cloud computing,and Artificial Intelligence(AI),which has led to an unprecedented era of connectivity.W...This paper presents a comprehensive exploration into the integration of Internet of Things(IoT),big data analysis,cloud computing,and Artificial Intelligence(AI),which has led to an unprecedented era of connectivity.We delve into the emerging trend of machine learning on embedded devices,enabling tasks in resource-limited environ-ments.However,the widespread adoption of machine learning raises significant privacy concerns,necessitating the development of privacy-preserving techniques.One such technique,secure multi-party computation(MPC),allows collaborative computations without exposing private inputs.Despite its potential,complex protocols and communication interactions hinder performance,especially on resource-constrained devices.Efforts to enhance efficiency have been made,but scalability remains a challenge.Given the success of GPUs in deep learning,lever-aging embedded GPUs,such as those offered by NVIDIA,emerges as a promising solution.Therefore,we propose an Embedded GPU-based Secure Two-party Computation(EG-STC)framework for Artificial Intelligence(AI)systems.To the best of our knowledge,this work represents the first endeavor to fully implement machine learning model training based on secure two-party computing on the Embedded GPU platform.Our experimental results demonstrate the effectiveness of EG-STC.On an embedded GPU with a power draw of 5 W,our implementation achieved a secure two-party matrix multiplication throughput of 5881.5 kilo-operations per millisecond(kops/ms),with an energy efficiency ratio of 1176.3 kops/ms/W.Furthermore,leveraging our EG-STC framework,we achieved an overall time acceleration ratio of 5–6 times compared to solutions running on server-grade CPUs.Our solution also exhibited a reduced runtime,requiring only 60%to 70%of the runtime of previously best-known methods on the same platform.In summary,our research contributes to the advancement of secure and efficient machine learning implementations on resource-constrained embedded devices,paving the way for broader adoption of AI technologies in various applications.展开更多
It is important to calculate the reachable domain(RD)of the manned lunar mission to evaluate whether a lunar landing site could be reached by the spacecraft. In this paper, the RD of free return orbits is quickly eval...It is important to calculate the reachable domain(RD)of the manned lunar mission to evaluate whether a lunar landing site could be reached by the spacecraft. In this paper, the RD of free return orbits is quickly evaluated and calculated via the classification and regression neural networks. An efficient databasegeneration method is developed for obtaining eight types of free return orbits and then the RD is defined by the orbit’s inclination and right ascension of ascending node(RAAN) at the perilune. A classify neural network and a regression network are trained respectively. The former is built for classifying the type of the RD, and the latter is built for calculating the inclination and RAAN of the RD. The simulation results show that two neural networks are well trained. The classification model has an accuracy of more than 99% and the mean square error of the regression model is less than 0.01°on the test set. Moreover, a serial strategy is proposed to combine the two surrogate models and a recognition tool is built to evaluate whether a lunar site could be reached. The proposed deep learning method shows the superiority in computation efficiency compared with the traditional double two-body model.展开更多
Model checking is an automated formal verification method to verify whether epistemic multi-agent systems adhere to property specifications.Although there is an extensive literature on qualitative properties such as s...Model checking is an automated formal verification method to verify whether epistemic multi-agent systems adhere to property specifications.Although there is an extensive literature on qualitative properties such as safety and liveness,there is still a lack of quantitative and uncertain property verifications for these systems.In uncertain environments,agents must make judicious decisions based on subjective epistemic.To verify epistemic and measurable properties in multi-agent systems,this paper extends fuzzy computation tree logic by introducing epistemic modalities and proposing a new Fuzzy Computation Tree Logic of Knowledge(FCTLK).We represent fuzzy multi-agent systems as distributed knowledge bases with fuzzy epistemic interpreted systems.In addition,we provide a transformation algorithm from fuzzy epistemic interpreted systems to fuzzy Kripke structures,as well as transformation rules from FCTLK formulas to Fuzzy Computation Tree Logic(FCTL)formulas.Accordingly,we transform the FCTLK model checking problem into the FCTL model checking.This enables the verification of FCTLK formulas by using the fuzzy model checking algorithm of FCTL without additional computational overheads.Finally,we present correctness proofs and complexity analyses of the proposed algorithms.Additionally,we further illustrate the practical application of our approach through an example of a train control system.展开更多
The use of the Internet of Things(IoT)is expanding at an unprecedented scale in many critical applications due to the ability to interconnect and utilize a plethora of wide range of devices.In critical infrastructure ...The use of the Internet of Things(IoT)is expanding at an unprecedented scale in many critical applications due to the ability to interconnect and utilize a plethora of wide range of devices.In critical infrastructure domains like oil and gas supply,intelligent transportation,power grids,and autonomous agriculture,it is essential to guarantee the confidentiality,integrity,and authenticity of data collected and exchanged.However,the limited resources coupled with the heterogeneity of IoT devices make it inefficient or sometimes infeasible to achieve secure data transmission using traditional cryptographic techniques.Consequently,designing a lightweight secure data transmission scheme is becoming essential.In this article,we propose lightweight secure data transmission(LSDT)scheme for IoT environments.LSDT consists of three phases and utilizes an effective combination of symmetric keys and the Elliptic Curve Menezes-Qu-Vanstone asymmetric key agreement protocol.We design the simulation environment and experiments to evaluate the performance of the LSDT scheme in terms of communication and computation costs.Security and performance analysis indicates that the LSDT scheme is secure,suitable for IoT applications,and performs better in comparison to other related security schemes.展开更多
基金financially supported by the National Key Research and Development Program of China (No.2021YFB3701001)the National Natural Science Foundation of China (No.U2102212)+1 种基金the Shanghai Rising-Star Program (No.21QA1403200)the Shanghai Engineering Research Center for Metal Parts Green Remanufacture (No.19DZ2252900) from Shanghai Engineering Research Center Construction Project。
文摘Mg-Al alloys have excellent strength and ductility but relatively low thermal conductivity due to Al addition.The accurate prediction of thermal conductivity is a prerequisite for designing Mg-Al alloys with high thermal conductivity.Thus,databases for predicting temperature-and composition-dependent thermal conductivities must be established.In this study,Mg-Al-La alloys with different contents of Al2La,Al3La,and Al11La3phases and solid solubility of Al in the α-Mg phase were designed.The influence of the second phase(s) and Al solid solubility on thermal conductivity was investigated.Experimental results revealed a second phase transformation from Al_(2)La to Al_(3)La and further to Al_(11)La_(3)with the increasing Al content at a constant La amount.The degree of the negative effect of the second phase(s) on thermal diffusivity followed the sequence of Al2La>Al3La>Al_(11)La_(3).Compared with the second phase,an increase in the solid solubility of Al in α-Mg remarkably reduced the thermal conductivity.On the basis of the experimental data,a database of the reciprocal thermal diffusivity of the Mg-Al-La system was established by calculation of the phase diagram (CALPHAD)method.With a standard error of±1.2 W/(m·K),the predicted results were in good agreement with the experimental data.The established database can be used to design Mg-Al alloys with high thermal conductivity and provide valuable guidance for expanding their application prospects.
基金supported in part by the National Natural Science Foundation of China (NSFC) under Grant No.61976242in part by the Natural Science Fund of Hebei Province for Distinguished Young Scholars under Grant No.F2021202010+2 种基金in part by the Fundamental Scientific Research Funds for Interdisciplinary Team of Hebei University of Technology under Grant No.JBKYTD2002funded by Science and Technology Project of Hebei Education Department under Grant No.JZX2023007supported by 2022 Interdisciplinary Postgraduate Training Program of Hebei University of Technology under Grant No.HEBUT-YXKJC-2022122.
文摘Most of the neural network architectures are based on human experience,which requires a long and tedious trial-and-error process.Neural architecture search(NAS)attempts to detect effective architectures without human intervention.Evolutionary algorithms(EAs)for NAS can find better solutions than human-designed architectures by exploring a large search space for possible architectures.Using multiobjective EAs for NAS,optimal neural architectures that meet various performance criteria can be explored and discovered efficiently.Furthermore,hardware-accelerated NAS methods can improve the efficiency of the NAS.While existing reviews have mainly focused on different strategies to complete NAS,a few studies have explored the use of EAs for NAS.In this paper,we summarize and explore the use of EAs for NAS,as well as large-scale multiobjective optimization strategies and hardware-accelerated NAS methods.NAS performs well in healthcare applications,such as medical image analysis,classification of disease diagnosis,and health monitoring.EAs for NAS can automate the search process and optimize multiple objectives simultaneously in a given healthcare task.Deep neural network has been successfully used in healthcare,but it lacks interpretability.Medical data is highly sensitive,and privacy leaks are frequently reported in the healthcare industry.To solve these problems,in healthcare,we propose an interpretable neuroevolution framework based on federated learning to address search efficiency and privacy protection.Moreover,we also point out future research directions for evolutionary NAS.Overall,for researchers who want to use EAs to optimize NNs in healthcare,we analyze the advantages and disadvantages of doing so to provide detailed guidance,and propose an interpretable privacy-preserving framework for healthcare applications.
文摘The interconnection bottleneck caused by limitations of cable number, inner space and cooling power of dilution refrigerators has been an outstanding challenge for building scalable superconducting quantum computers with the increasing number of qubits in quantum processors. To surmount such an obstacle, it is desirable to integrate qubits with quantum–classical interface(QCI) circuits based on rapid single flux quantum(RSFQ) circuits. In this work, a digital flux tuner for qubits(DFTQ) is proposed for manipulating flux of qubits as a crucial part of the interface circuit. A schematic diagram of the DFTQ is presented, consisting of a coarse tuning unit and a fine-tuning unit for providing magnetic flux with different precision to qubits. The method of using DFTQ to provide flux for gate operations is discussed from the optimization of circuit design and input signal. To verify the effectiveness of the method, simulations of a single DFTQ and quantum gates including a Z gate and an iSWAP gate with DFTQs are performed for flux-tunable transmons. The quantum process tomography corresponding to the two gates is also carried out to analyze the sources of gate error. The results of tomography show that the gate fidelities independent of the initial states of the Z gate and the iSWAP gate are 99.935% and 99.676%,respectively. With DFTQs inside, the QCI would be a powerful tool for building large-scale quantum computers.
基金Project supported by the National Natural Science Foundation of China (Grant Nos.12074368,92165207,12034018,and 62004185)the Anhui Province Natural Science Foundation (Grant No.2108085J03)the USTC Tang Scholarship。
文摘The single-shot readout data process is essential for the realization of high-fidelity qubits and fault-tolerant quantum algorithms in semiconductor quantum dots. However, the fidelity and visibility of the readout process are sensitive to the choice of the thresholds and limited by the experimental hardware. By demonstrating the linear dependence between the measured spin state probabilities and readout visibilities along with dark counts, we describe an alternative threshold-independent method for the single-shot readout of spin qubits in semiconductor quantum dots. We can obtain the extrapolated spin state probabilities of the prepared probabilities of the excited spin state through the threshold-independent method. We then analyze the corresponding errors of the method, finding that errors of the extrapolated probabilities cannot be neglected with no constraints on the readout time and threshold voltage. Therefore, by limiting the readout time and threshold voltage, we ensure the accuracy of the extrapolated probability. We then prove that the efficiency and robustness of this method are 60 times larger than those of the most commonly used method. Moreover, we discuss the influence of the electron temperature on the effective area with a fixed external magnetic field and provide a preliminary demonstration for a single-shot readout of up to 0.7K/1.5T in the future.
基金funded by National Natural Science Foundation of China,(Grant Nos.11974030 and 92165208)。
文摘Electron spins confined in semiconductor quantum dots(QDs)are one of potential candidates for physical implementation of scalable quantum information processing technologies.Tunnel coupling based inter exchange interaction between QDs is crucial in achieving single-qubit manipulation,two-qubit gate,quantum communication and quantum simulation.This review first provides a theoretical perspective that surveys a general framework,including the Helter−London approach,the Hund−Mulliken approach,and the Hubbard model,to describe the inter exchange interactions between semiconductor quantum dots.An electrical method to control the inter exchange interaction in a realistic device is proposed as well.Then the significant achievements of inter exchange interaction in manipulating single qubits,achieving two-qubit gates,performing quantum communication and quantum simulation are reviewed.The last part is a summary of this review.
基金supported in part by the National Natural Science Foundation of China under Grant 61901128,62273109the Natural Science Foundation of the Jiangsu Higher Education Institutions of China(21KJB510032).
文摘The growing development of the Internet of Things(IoT)is accelerating the emergence and growth of new IoT services and applications,which will result in massive amounts of data being generated,transmitted and pro-cessed in wireless communication networks.Mobile Edge Computing(MEC)is a desired paradigm to timely process the data from IoT for value maximization.In MEC,a number of computing-capable devices are deployed at the network edge near data sources to support edge computing,such that the long network transmission delay in cloud computing paradigm could be avoided.Since an edge device might not always have sufficient resources to process the massive amount of data,computation offloading is significantly important considering the coop-eration among edge devices.However,the dynamic traffic characteristics and heterogeneous computing capa-bilities of edge devices challenge the offloading.In addition,different scheduling schemes might provide different computation delays to the offloaded tasks.Thus,offloading in mobile nodes and scheduling in the MEC server are coupled to determine service delay.This paper seeks to guarantee low delay for computation intensive applica-tions by jointly optimizing the offloading and scheduling in such an MEC system.We propose a Delay-Greedy Computation Offloading(DGCO)algorithm to make offloading decisions for new tasks in distributed computing-enabled mobile devices.A Reinforcement Learning-based Parallel Scheduling(RLPS)algorithm is further designed to schedule offloaded tasks in the multi-core MEC server.With an offloading delay broadcast mechanism,the DGCO and RLPS cooperate to achieve the goal of delay-guarantee-ratio maximization.Finally,the simulation results show that our proposal can bound the end-to-end delay of various tasks.Even under slightly heavy task load,the delay-guarantee-ratio given by DGCO-RLPS can still approximate 95%,while that given by benchmarked algorithms is reduced to intolerable value.The simulation results are demonstrated the effective-ness of DGCO-RLPS for delay guarantee in MEC.
基金supported by the National Natural Science Foundation of China(No.32002442)the National Key R&D Program(No.2019YFD0902101).
文摘On the basis of computational fluid dynamics,the flow field characteristics of multi-trophic artificial reefs,including the flow field distribution features of a single reef under three different velocities and the effect of spacing between reefs on flow scale and the flow state,were analyzed.Results indicate upwelling,slow flow,and eddy around a single reef.Maximum velocity,height,and volume of upwelling in front of a single reef were positively correlated with inflow velocity.The length and volume of slow flow increased with the increase in inflow velocity.Eddies were present both inside and backward,and vorticity was positively correlated with inflow velocity.Space between reefs had a minor influence on the maximum velocity and height of upwelling.With the increase in space from 0.5 L to 1.5 L(L is the reef lehgth),the length of slow flow in the front and back of the combined reefs increased slightly.When the space was 2.0 L,the length of the slow flow decreased.In four different spaces,eddies were present inside and at the back of each reef.The maximum vorticity was negatively correlated with space from 0.5 L to 1.5 L,but under 2.0 L space,the maximum vorticity was close to the vorticity of a single reef under the same inflow velocity.
基金Project supported by the General Project of Natural Science Foundation of Hunan Province(Grant Nos.2024JJ5273 and 2023JJ50328)the Scientific Research Project of Education Department of Hunan Province(Grant Nos.22A0049 and 22B0699)。
文摘This paper presents a novel approach to proxy blind signatures in the realm of quantum circuits,aiming to enhance security while safeguarding sensitive information.The main objective of this research is to introduce a quantum proxy blind signature(QPBS)protocol that utilizes quantum logical gates and quantum measurement techniques.The QPBS protocol is constructed by the initial phase,proximal blinding message phase,remote authorization and signature phase,remote validation,and de-blinding phase.This innovative design ensures a secure mechanism for signing documents without revealing the content to the proxy signer,providing practical security authentication in a quantum environment under the assumption that the CNOT gates are securely implemented.Unlike existing approaches,our proposed QPBS protocol eliminates the need for quantum entanglement preparation,thus simplifying the implementation process.To assess the effectiveness and robustness of the QPBS protocol,we conduct comprehensive simulation studies in both ideal and noisy quantum environments on the IBM quantum cloud platform.The results demonstrate the superior performance of the QPBS algorithm,highlighting its resilience against repudiation and forgeability,which are key security concerns in the realm of proxy blind signatures.Furthermore,we have established authentic security thresholds(82.102%)in the presence of real noise,thereby emphasizing the practicality of our proposed solution.
文摘Practical real-world scenarios such as the Internet,social networks,and biological networks present the challenges of data scarcity and complex correlations,which limit the applications of artificial intelligence.The graph structure is a typical tool used to formulate such correlations,it is incapable of modeling highorder correlations among different objects in systems;thus,the graph structure cannot fully convey the intricate correlations among objects.Confronted with the aforementioned two challenges,hypergraph computation models high-order correlations among data,knowledge,and rules through hyperedges and leverages these high-order correlations to enhance the data.Additionally,hypergraph computation achieves collaborative computation using data and high-order correlations,thereby offering greater modeling flexibility.In particular,we introduce three types of hypergraph computation methods:①hypergraph structure modeling,②hypergraph semantic computing,and③efficient hypergraph computing.We then specify how to adopt hypergraph computation in practice by focusing on specific tasks such as three-dimensional(3D)object recognition,revealing that hypergraph computation can reduce the data requirement by 80%while achieving comparable performance or improve the performance by 52%given the same data,compared with a traditional data-based method.A comprehensive overview of the applications of hypergraph computation in diverse domains,such as intelligent medicine and computer vision,is also provided.Finally,we introduce an open-source deep learning library,DeepHypergraph(DHG),which can serve as a tool for the practical usage of hypergraph computation.
基金Fundamental Research Funds for the Central Universities under Grant No.2682022CX072the Research and Development Plan in Key Areas of Guangdong Province under Grant No.2020B0202010008。
文摘An efficient approach is proposed for the equivalent linearization of frame structures with plastic hinges under nonstationary seismic excitations.The concentrated plastic hinges,described by the Bouc-Wen model,are assumed to occur at the two ends of a linear-elastic beam element.The auxiliary differential equations governing the plastic rotational displacements and their corresponding hysteretic displacements are replaced with linearized differential equations.Then,the two sets of equations of motion for the original nonlinear system can be reduced to an expanded-order equivalent linearized equation of motion for equivalent linear systems.To solve the equation of motion for equivalent linear systems,the nonstationary random vibration analysis is carried out based on the explicit time-domain method with high efficiency.Finally,the proposed treatment method for initial values of equivalent parameters is investigated in conjunction with parallel computing technology,which provides a new way of obtaining the equivalent linear systems at different time instants.Based on the explicit time-domain method,the key responses of interest of the converged equivalent linear system can be calculated through dimension reduction analysis with high efficiency.Numerical examples indicate that the proposed approach has high computational efficiency,and shows good applicability to weak nonlinear and medium-intensity nonlinear systems.
基金work is supported by the Fundamental Research Funds for the Central Universities(No.3102019HTQD014)of Northwestern Polytechnical UniversityFunding of National Key Laboratory of Astronautical Flight DynamicsYoung Talent Support Project of Shaanxi State.
文摘Although predictor-corrector methods have been extensively applied,they might not meet the requirements of practical applications and engineering tasks,particularly when high accuracy and efficiency are necessary.A novel class of correctors based on feedback-accelerated Picard iteration(FAPI)is proposed to further enhance computational performance.With optimal feedback terms that do not require inversion of matrices,significantly faster convergence speed and higher numerical accuracy are achieved by these correctors compared with their counterparts;however,the computational complexities are comparably low.These advantages enable nonlinear engineering problems to be solved quickly and accurately,even with rough initial guesses from elementary predictors.The proposed method offers flexibility,enabling the use of the generated correctors for either bulk processing of collocation nodes in a domain or successive corrections of a single node in a finite difference approach.In our method,the functional formulas of FAPI are discretized into numerical forms using the collocation approach.These collocated iteration formulas can directly solve nonlinear problems,but they may require significant computational resources because of the manipulation of high-dimensionalmatrices.To address this,the collocated iteration formulas are further converted into finite difference forms,enabling the design of lightweight predictor-corrector algorithms for real-time computation.The generality of the proposed method is illustrated by deriving new correctors for three commonly employed finite-difference approaches:the modified Euler approach,the Adams-Bashforth-Moulton approach,and the implicit Runge-Kutta approach.Subsequently,the updated approaches are tested in solving strongly nonlinear problems,including the Matthieu equation,the Duffing equation,and the low-earth-orbit tracking problem.The numerical findings confirm the computational accuracy and efficiency of the derived predictor-corrector algorithms.
基金This work is supported by National Natural Science Foundation of China(Nos.U23B20151 and 52171253).
文摘Owing to the complex lithology of unconventional reservoirs,field interpreters usually need to provide a basis for interpretation using logging simulation models.Among the various detection tools that use nuclear sources,the detector response can reflect various types of information of the medium.The Monte Carlo method is one of the primary methods used to obtain nuclear detection responses in complex environments.However,this requires a computational process with extensive random sampling,consumes considerable resources,and does not provide real-time response results.Therefore,a novel fast forward computational method(FFCM)for nuclear measurement that uses volumetric detection constraints to rapidly calculate the detector response in various complex environments is proposed.First,the data library required for the FFCM is built by collecting the detection volume,detector counts,and flux sensitivity functions through a Monte Carlo simulation.Then,based on perturbation theory and the Rytov approximation,a model for the detector response is derived using the flux sensitivity function method and a one-group diffusion model.The environmental perturbation is constrained to optimize the model according to the tool structure and the impact of the formation and borehole within the effective detection volume.Finally,the method is applied to a neutron porosity tool for verification.In various complex simulation environments,the maximum relative error between the calculated porosity results of Monte Carlo and FFCM was 6.80%,with a rootmean-square error of 0.62 p.u.In field well applications,the formation porosity model obtained using FFCM was in good agreement with the model obtained by interpreters,which demonstrates the validity and accuracy of the proposed method.
基金supported by the Innovation and Technology Commission(Grant no.PRP/032/20FX)the RFBR(Grant no.20-03-00772)
文摘Rechargeable Zn-air batteries(ZAB)have drawn extensive attention due to their eco-friendliness and safety.However,the lack of high-performance and low-cost oxygen redox reactions(OER and ORR)catalysts has become one of the main stumbling blocks in their development.Herein,we successfully fabricate a CoFe nanobubble encapsulated in nitrogen-doped carbon nanocage on wood carbon support(CoFe@NC/WC)via pyrolysis of a novel Prussian blue analog(PBA)/spruce precursor.The hierarchical CoFe@NC/WC catalyst exhibits an excellent potential difference of 0.74 V between the OER potential at 10 mA cm^(-2)and half-wave potential of ORR in 0.1 M KOH,comparable to recently reported preeminent electrocatalysts.Further,CoFe@NC/WC shows outstanding electrochemical performance in liquid ZAB,with a peak power density of 138.9 mW cm^(-2)and a specific capacity of 763.5 mAh g^(-1).More importantly,a bacterial cellulose nanofiber reinforced polyacrylic acid(BC-PAA)hydrogel electrolyte shows ultrahigh tensile-breaking stress of 1.58 MPa.In conjunction with the as-prepared CoFe@NC/WC catalyst,BC-PAA-based wearable ZAB displays impressive rechargeability and foldability,and can power portable electronics,such as electronic timer and mobile phone,in bent states.This work provides a new approach toward high-activity and low-cost catalysts for ZAB.
基金supported by the Liaoning Province Applied Basic Research Program Project of China(Grant:2023JH2/101300065)the Liaoning Province Science and Technology Plan Joint Fund(2023-MSLH-221).
文摘Deep learning has emerged in many practical applications,such as image classification,fault diagnosis,and object detection.More recently,convolutional neural networks(CNNs),representative models of deep learning,have been used to solve fault detection.However,the current design of CNNs for fault detection of wind turbine blades is highly dependent on domain knowledge and requires a large amount of trial and error.For this reason,an evolutionary YOLOv8 network has been developed to automatically find the network architecture for wind turbine blade-based fault detection.YOLOv8 is a CNN-backed object detection model.Specifically,to reduce the parameter count,we first design an improved FasterNet module based on the Partial Convolution(PConv)operator.Then,to enhance convergence performance,we improve the loss function based on the efficient complete intersection over the union.Based on this,a flexible variable-length encoding is proposed,and the corresponding reproduction operators are designed.Related experimental results confirmthat the proposed approach can achieve better fault detection results and improve by 2.6%in mean precision at 50(mAP50)compared to the existing methods.Additionally,compared to training with the YOLOv8n model,the YOLOBFE model reduces the training parameters by 933,937 and decreases the GFLOPS(Giga Floating Point Operations Per Second)by 1.1.
基金supported by the Innovation Program for Quantum Science and Technology (Grant No.2021ZD0302400)the National Natural Science Foundation of China (Grants No.11974198)the Natural Science Foundation of Shandong Province of China (Grant No.ZR2021MA091)。
文摘Majorana quantum computation offers a potential approach to securely manipulating and storing quantum data in a topological manner that may effectively resist the decoherence induced by local noise. However, actual Majorana qubit setups are susceptible to noise. In this study, from a quantum dynamics perspective, we develop a noise model for Majorana qubits that accounts for quasi-particle poisoning and Majorana overlapping with fluctuation. Furthermore, we focus on Majorana parity readout methodologies, specifically those leveraging an ancillary quantum dot, and carry out an indepth exploration of continuous measurement techniques founded on the quantum jump model of a quantum point contact.Utilizing these methodologies, we proceed to analyze the influence of noise on the afore-mentioned noise model, employing numerical computation to evaluate the power spectrum and frequency curve. In the culmination of our study, we put forward a strategy to benchmark the presence and detailed properties of noise in Majorana qubits.
基金Project supported by the National Natural Science Foundation of China (Grant Nos. U20A20227,62076208, and 62076207)Chongqing Talent Plan “Contract System” Project (Grant No. CQYC20210302257)+3 种基金National Key Laboratory of Smart Vehicle Safety Technology Open Fund Project (Grant No. IVSTSKL-202309)the Chongqing Technology Innovation and Application Development Special Major Project (Grant No. CSTB2023TIAD-STX0020)College of Artificial Intelligence, Southwest UniversityState Key Laboratory of Intelligent Vehicle Safety Technology
文摘Neuromorphic computing,inspired by the human brain,uses memristor devices for complex tasks.Recent studies show that self-organizing random nanowires can implement neuromorphic information processing,enabling data analysis.This paper presents a model based on these nanowire networks,with an improved conductance variation profile.We suggest using these networks for temporal information processing via a reservoir computing scheme and propose an efficient data encoding method using voltage pulses.The nanowire network layer generates dynamic behaviors for pulse voltages,allowing time series prediction analysis.Our experiment uses a double stochastic nanowire network architecture for processing multiple input signals,outperforming traditional reservoir computing in terms of fewer nodes,enriched dynamics and improved prediction accuracy.Experimental results confirm the high accuracy of this architecture on multiple real-time series datasets,making neuromorphic nanowire networks promising for physical implementation of reservoir computing.
基金supported in part by Major Science and Technology Demonstration Project of Jiangsu Provincial Key R&D Program under Grant No.BE2023025in part by the National Natural Science Foundation of China under Grant No.62302238+2 种基金in part by the Natural Science Foundation of Jiangsu Province under Grant No.BK20220388in part by the Natural Science Research Project of Colleges and Universities in Jiangsu Province under Grant No.22KJB520004in part by the China Postdoctoral Science Foundation under Grant No.2022M711689.
文摘This paper presents a comprehensive exploration into the integration of Internet of Things(IoT),big data analysis,cloud computing,and Artificial Intelligence(AI),which has led to an unprecedented era of connectivity.We delve into the emerging trend of machine learning on embedded devices,enabling tasks in resource-limited environ-ments.However,the widespread adoption of machine learning raises significant privacy concerns,necessitating the development of privacy-preserving techniques.One such technique,secure multi-party computation(MPC),allows collaborative computations without exposing private inputs.Despite its potential,complex protocols and communication interactions hinder performance,especially on resource-constrained devices.Efforts to enhance efficiency have been made,but scalability remains a challenge.Given the success of GPUs in deep learning,lever-aging embedded GPUs,such as those offered by NVIDIA,emerges as a promising solution.Therefore,we propose an Embedded GPU-based Secure Two-party Computation(EG-STC)framework for Artificial Intelligence(AI)systems.To the best of our knowledge,this work represents the first endeavor to fully implement machine learning model training based on secure two-party computing on the Embedded GPU platform.Our experimental results demonstrate the effectiveness of EG-STC.On an embedded GPU with a power draw of 5 W,our implementation achieved a secure two-party matrix multiplication throughput of 5881.5 kilo-operations per millisecond(kops/ms),with an energy efficiency ratio of 1176.3 kops/ms/W.Furthermore,leveraging our EG-STC framework,we achieved an overall time acceleration ratio of 5–6 times compared to solutions running on server-grade CPUs.Our solution also exhibited a reduced runtime,requiring only 60%to 70%of the runtime of previously best-known methods on the same platform.In summary,our research contributes to the advancement of secure and efficient machine learning implementations on resource-constrained embedded devices,paving the way for broader adoption of AI technologies in various applications.
基金supported by the National Natural Science Foundation of China (12072365)the Natural Science Foundation of Hunan Province of China (2020JJ4657)。
文摘It is important to calculate the reachable domain(RD)of the manned lunar mission to evaluate whether a lunar landing site could be reached by the spacecraft. In this paper, the RD of free return orbits is quickly evaluated and calculated via the classification and regression neural networks. An efficient databasegeneration method is developed for obtaining eight types of free return orbits and then the RD is defined by the orbit’s inclination and right ascension of ascending node(RAAN) at the perilune. A classify neural network and a regression network are trained respectively. The former is built for classifying the type of the RD, and the latter is built for calculating the inclination and RAAN of the RD. The simulation results show that two neural networks are well trained. The classification model has an accuracy of more than 99% and the mean square error of the regression model is less than 0.01°on the test set. Moreover, a serial strategy is proposed to combine the two surrogate models and a recognition tool is built to evaluate whether a lunar site could be reached. The proposed deep learning method shows the superiority in computation efficiency compared with the traditional double two-body model.
基金The work is partially supported by Natural Science Foundation of Ningxia(Grant No.AAC03300)National Natural Science Foundation of China(Grant No.61962001)Graduate Innovation Project of North Minzu University(Grant No.YCX23152).
文摘Model checking is an automated formal verification method to verify whether epistemic multi-agent systems adhere to property specifications.Although there is an extensive literature on qualitative properties such as safety and liveness,there is still a lack of quantitative and uncertain property verifications for these systems.In uncertain environments,agents must make judicious decisions based on subjective epistemic.To verify epistemic and measurable properties in multi-agent systems,this paper extends fuzzy computation tree logic by introducing epistemic modalities and proposing a new Fuzzy Computation Tree Logic of Knowledge(FCTLK).We represent fuzzy multi-agent systems as distributed knowledge bases with fuzzy epistemic interpreted systems.In addition,we provide a transformation algorithm from fuzzy epistemic interpreted systems to fuzzy Kripke structures,as well as transformation rules from FCTLK formulas to Fuzzy Computation Tree Logic(FCTL)formulas.Accordingly,we transform the FCTLK model checking problem into the FCTL model checking.This enables the verification of FCTLK formulas by using the fuzzy model checking algorithm of FCTL without additional computational overheads.Finally,we present correctness proofs and complexity analyses of the proposed algorithms.Additionally,we further illustrate the practical application of our approach through an example of a train control system.
基金support of the Interdisciplinary Research Center for Intelligent Secure Systems(IRC-ISS)Internal Fund Grant#INSS2202.
文摘The use of the Internet of Things(IoT)is expanding at an unprecedented scale in many critical applications due to the ability to interconnect and utilize a plethora of wide range of devices.In critical infrastructure domains like oil and gas supply,intelligent transportation,power grids,and autonomous agriculture,it is essential to guarantee the confidentiality,integrity,and authenticity of data collected and exchanged.However,the limited resources coupled with the heterogeneity of IoT devices make it inefficient or sometimes infeasible to achieve secure data transmission using traditional cryptographic techniques.Consequently,designing a lightweight secure data transmission scheme is becoming essential.In this article,we propose lightweight secure data transmission(LSDT)scheme for IoT environments.LSDT consists of three phases and utilizes an effective combination of symmetric keys and the Elliptic Curve Menezes-Qu-Vanstone asymmetric key agreement protocol.We design the simulation environment and experiments to evaluate the performance of the LSDT scheme in terms of communication and computation costs.Security and performance analysis indicates that the LSDT scheme is secure,suitable for IoT applications,and performs better in comparison to other related security schemes.