The evolution of the probability density function of a stochastic dynamical system over time can be described by a Fokker–Planck–Kolmogorov(FPK) equation, the solution of which determines the distribution of macrosc...The evolution of the probability density function of a stochastic dynamical system over time can be described by a Fokker–Planck–Kolmogorov(FPK) equation, the solution of which determines the distribution of macroscopic variables in the stochastic dynamic system. Traditional methods for solving these equations often struggle with computational efficiency and scalability, particularly in high-dimensional contexts. To address these challenges, this paper proposes a novel deep learning method based on prior knowledge with dual training to solve the stationary FPK equations. Initially, the neural network is pre-trained through the prior knowledge obtained by Monte Carlo simulation(MCS). Subsequently, the second training phase incorporates the FPK differential operator into the loss function, while a supervisory term consisting of local maximum points is specifically included to mitigate the generation of zero solutions. This dual-training strategy not only expedites convergence but also enhances computational efficiency, making the method well-suited for high-dimensional systems. Numerical examples, including two different two-dimensional(2D), six-dimensional(6D), and eight-dimensional(8D) systems, are conducted to assess the efficacy of the proposed method. The results demonstrate robust performance in terms of both computational speed and accuracy for solving FPK equations in the first three systems. While the method is also applicable to high-dimensional systems, such as 8D, it should be noted that computational efficiency may be marginally compromised due to data volume constraints.展开更多
Facial beauty analysis is an important topic in human society.It may be used as a guidance for face beautification applications such as cosmetic surgery.Deep neural networks(DNNs)have recently been adopted for facial ...Facial beauty analysis is an important topic in human society.It may be used as a guidance for face beautification applications such as cosmetic surgery.Deep neural networks(DNNs)have recently been adopted for facial beauty analysis and have achieved remarkable performance.However,most existing DNN-based models regard facial beauty analysis as a normal classification task.They ignore important prior knowledge in traditional machine learning models which illustrate the significant contribution of the geometric features in facial beauty analysis.To be specific,landmarks of the whole face and facial organs are introduced to extract geometric features to make the decision.Inspired by this,we introduce a novel dual-branch network for facial beauty analysis:one branch takes the Swin Transformer as the backbone to model the full face and global patterns,and another branch focuses on the masked facial organs with the residual network to model the local patterns of certain facial parts.Additionally,the designed multi-scale feature fusion module can further facilitate our network to learn complementary semantic information between the two branches.In model optimisation,we propose a hybrid loss function,where especially geometric regulation is introduced by regressing the facial landmarks and it can force the extracted features to convey facial geometric features.Experiments performed on the SCUT-FBP5500 dataset and the SCUT-FBP dataset demonstrate that our model outperforms the state-of-the-art convolutional neural networks models,which proves the effectiveness of the proposed geometric regularisation and dual-branch structure with the hybrid network.To the best of our knowledge,this is the first study to introduce a Vision Transformer into the facial beauty analysis task.展开更多
Plug-in Hybrid Electric Vehicles(PHEVs)represent an innovative breed of transportation,harnessing diverse power sources for enhanced performance.Energy management strategies(EMSs)that coordinate and control different ...Plug-in Hybrid Electric Vehicles(PHEVs)represent an innovative breed of transportation,harnessing diverse power sources for enhanced performance.Energy management strategies(EMSs)that coordinate and control different energy sources is a critical component of PHEV control technology,directly impacting overall vehicle performance.This study proposes an improved deep reinforcement learning(DRL)-based EMSthat optimizes realtime energy allocation and coordinates the operation of multiple power sources.Conventional DRL algorithms struggle to effectively explore all possible state-action combinations within high-dimensional state and action spaces.They often fail to strike an optimal balance between exploration and exploitation,and their assumption of a static environment limits their ability to adapt to changing conditions.Moreover,these algorithms suffer from low sample efficiency.Collectively,these factors contribute to convergence difficulties,low learning efficiency,and instability.To address these challenges,the Deep Deterministic Policy Gradient(DDPG)algorithm is enhanced using entropy regularization and a summation tree-based Prioritized Experience Replay(PER)method,aiming to improve exploration performance and learning efficiency from experience samples.Additionally,the correspondingMarkovDecision Process(MDP)is established.Finally,an EMSbased on the improvedDRLmodel is presented.Comparative simulation experiments are conducted against rule-based,optimization-based,andDRL-based EMSs.The proposed strategy exhibitsminimal deviation fromthe optimal solution obtained by the dynamic programming(DP)strategy that requires global information.In the typical driving scenarios based onWorld Light Vehicle Test Cycle(WLTC)and New European Driving Cycle(NEDC),the proposed method achieved a fuel consumption of 2698.65 g and an Equivalent Fuel Consumption(EFC)of 2696.77 g.Compared to the DP strategy baseline,the proposed method improved the fuel efficiency variances(FEV)by 18.13%,15.1%,and 8.37%over the Deep QNetwork(DQN),Double DRL(DDRL),and original DDPG methods,respectively.The observational outcomes demonstrate that the proposed EMS based on improved DRL framework possesses good real-time performance,stability,and reliability,effectively optimizing vehicle economy and fuel consumption.展开更多
For some important object recognition applications such as intelligent robots and unmanned driving, images are collected on a consecutive basis and associated among themselves, besides, the scenes have steady prior fe...For some important object recognition applications such as intelligent robots and unmanned driving, images are collected on a consecutive basis and associated among themselves, besides, the scenes have steady prior features. Yet existing technologies do not take full advantage of this information. In order to take object recognition further than existing algorithms in the above application, an object recognition method that fuses temporal sequence with scene priori information is proposed. This method first employs YOLOv3 as the basic algorithm to recognize objects in single-frame images, then the DeepSort algorithm to establish association among potential objects recognized in images of different moments, and finally the confidence fusion method and temporal boundary processing method designed herein to fuse, at the decision level, temporal sequence information with scene priori information. Experiments using public datasets and self-built industrial scene datasets show that due to the expansion of information sources, the quality of single-frame images has less impact on the recognition results, whereby the object recognition is greatly improved. It is presented herein as a widely applicable framework for the fusion of information under multiple classes. All the object recognition algorithms that output object class, location information and recognition confidence at the same time can be integrated into this information fusion framework to improve performance.展开更多
240 nm AlGaN-based micro-LEDs with different sizes are designed and fabricated.Then,the external quantum efficiency(EQE)and light extraction efficiency(LEE)are systematically investigated by comparing size and edge ef...240 nm AlGaN-based micro-LEDs with different sizes are designed and fabricated.Then,the external quantum efficiency(EQE)and light extraction efficiency(LEE)are systematically investigated by comparing size and edge effects.Here,it is revealed that the peak optical output power increases by 81.83%with the size shrinking from 50.0 to 25.0μm.Thereinto,the LEE increases by 26.21%and the LEE enhancement mainly comes from the sidewall light extraction.Most notably,transversemagnetic(TM)mode light intensifies faster as the size shrinks due to the tilted mesa side-wall and Al reflector design.However,when it turns to 12.5μm sized micro-LEDs,the output power is lower than 25.0μm sized ones.The underlying mechanism is that even though protected by SiO2 passivation,the edge effect which leads to current leakage and Shockley-Read-Hall(SRH)recombination deteriorates rapidly with the size further shrinking.Moreover,the ratio of the p-contact area to mesa area is much lower,which deteriorates the p-type current spreading at the mesa edge.These findings show a role of thumb for the design of high efficiency micro-LEDs with wavelength below 250 nm,which will pave the way for wide applications of deep ultraviolet(DUV)micro-LEDs.展开更多
Limited by the dynamic range of the detector,saturation artifacts usually occur in optical coherence tomography(OCT)imaging for high scattering media.The available methods are difficult to remove saturation artifacts ...Limited by the dynamic range of the detector,saturation artifacts usually occur in optical coherence tomography(OCT)imaging for high scattering media.The available methods are difficult to remove saturation artifacts and restore texture completely in OCT images.We proposed a deep learning-based inpainting method of saturation artifacts in this paper.The generation mechanism of saturation artifacts was analyzed,and experimental and simulated datasets were built based on the mechanism.Enhanced super-resolution generative adversarial networks were trained by the clear–saturated phantom image pairs.The perfect reconstructed results of experimental zebrafish and thyroid OCT images proved its feasibility,strong generalization,and robustness.展开更多
Deep and ultra-deep reservoirs have gradually become the primary focus of hydrocarbon exploration as a result of a series of significant discoveries in deep hydrocarbon exploration worldwide.These reservoirs present u...Deep and ultra-deep reservoirs have gradually become the primary focus of hydrocarbon exploration as a result of a series of significant discoveries in deep hydrocarbon exploration worldwide.These reservoirs present unique challenges due to their deep burial depth(4500-8882 m),low matrix permeability,complex crustal stress conditions,high temperature and pressure(HTHP,150-200℃,105-155 MPa),coupled with high salinity of formation water.Consequently,the costs associated with their exploitation and development are exceptionally high.In deep and ultra-deep reservoirs,hydraulic fracturing is commonly used to achieve high and stable production.During hydraulic fracturing,a substantial volume of fluid is injected into the reservoir.However,statistical analysis reveals that the flowback rate is typically less than 30%,leaving the majority of the fluid trapped within the reservoir.Therefore,hydraulic fracturing in deep reservoirs not only enhances the reservoir permeability by creating artificial fractures but also damages reservoirs due to the fracturing fluids involved.The challenging“three-high”environment of a deep reservoir,characterized by high temperature,high pressure,and high salinity,exacerbates conventional forms of damage,including water sensitivity,retention of fracturing fluids,rock creep,and proppant breakage.In addition,specific damage mechanisms come into play,such as fracturing fluid decomposition at elevated temperatures and proppant diagenetic reactions at HTHP conditions.Presently,the foremost concern in deep oil and gas development lies in effectively assessing the damage inflicted on these reservoirs by hydraulic fracturing,comprehending the underlying mechanisms,and selecting appropriate solutions.It's noteworthy that the majority of existing studies on reservoir damage primarily focus on conventional reservoirs,with limited attention given to deep reservoirs and a lack of systematic summaries.In light of this,our approach entails initially summarizing the current knowledge pertaining to the types of fracturing fluids employed in deep and ultra-deep reservoirs.Subsequently,we delve into a systematic examination of the damage processes and mechanisms caused by fracturing fluids within the context of hydraulic fracturing in deep reservoirs,taking into account the unique reservoir characteristics of high temperature,high pressure,and high in-situ stress.In addition,we provide an overview of research progress related to high-temperature deep reservoir fracturing fluid and the damage of aqueous fracturing fluids to rock matrix,both artificial and natural fractures,and sand-packed fractures.We conclude by offering a summary of current research advancements and future directions,which hold significant potential for facilitating the efficient development of deep oil and gas reservoirs while effectively mitigating reservoir damage.展开更多
The amount of oxygen blown into the converter is one of the key parameters for the control of the converter blowing process,which directly affects the tap-to-tap time of converter. In this study, a hybrid model based ...The amount of oxygen blown into the converter is one of the key parameters for the control of the converter blowing process,which directly affects the tap-to-tap time of converter. In this study, a hybrid model based on oxygen balance mechanism (OBM) and deep neural network (DNN) was established for predicting oxygen blowing time in converter. A three-step method was utilized in the hybrid model. First, the oxygen consumption volume was predicted by the OBM model and DNN model, respectively. Second, a more accurate oxygen consumption volume was obtained by integrating the OBM model and DNN model. Finally, the converter oxygen blowing time was calculated according to the oxygen consumption volume and the oxygen supply intensity of each heat. The proposed hybrid model was verified using the actual data collected from an integrated steel plant in China, and compared with multiple linear regression model, OBM model, and neural network model including extreme learning machine, back propagation neural network, and DNN. The test results indicate that the hybrid model with a network structure of 3 hidden layer layers, 32-16-8 neurons per hidden layer, and 0.1 learning rate has the best prediction accuracy and stronger generalization ability compared with other models. The predicted hit ratio of oxygen consumption volume within the error±300 m^(3)is 96.67%;determination coefficient (R^(2)) and root mean square error (RMSE) are0.6984 and 150.03 m^(3), respectively. The oxygen blow time prediction hit ratio within the error±0.6 min is 89.50%;R2and RMSE are0.9486 and 0.3592 min, respectively. As a result, the proposed model can effectively predict the oxygen consumption volume and oxygen blowing time in the converter.展开更多
Avatars, as promising digital representations and service assistants of users in Metaverses, can enable drivers and passengers to immerse themselves in 3D virtual services and spaces of UAV-assisted vehicular Metavers...Avatars, as promising digital representations and service assistants of users in Metaverses, can enable drivers and passengers to immerse themselves in 3D virtual services and spaces of UAV-assisted vehicular Metaverses. However, avatar tasks include a multitude of human-to-avatar and avatar-to-avatar interactive applications, e.g., augmented reality navigation,which consumes intensive computing resources. It is inefficient and impractical for vehicles to process avatar tasks locally. Fortunately, migrating avatar tasks to the nearest roadside units(RSU)or unmanned aerial vehicles(UAV) for execution is a promising solution to decrease computation overhead and reduce task processing latency, while the high mobility of vehicles brings challenges for vehicles to independently perform avatar migration decisions depending on current and future vehicle status. To address these challenges, in this paper, we propose a novel avatar task migration system based on multi-agent deep reinforcement learning(MADRL) to execute immersive vehicular avatar tasks dynamically. Specifically, we first formulate the problem of avatar task migration from vehicles to RSUs/UAVs as a partially observable Markov decision process that can be solved by MADRL algorithms. We then design the multi-agent proximal policy optimization(MAPPO) approach as the MADRL algorithm for the avatar task migration problem. To overcome slow convergence resulting from the curse of dimensionality and non-stationary issues caused by shared parameters in MAPPO, we further propose a transformer-based MAPPO approach via sequential decision-making models for the efficient representation of relationships among agents. Finally, to motivate terrestrial or non-terrestrial edge servers(e.g., RSUs or UAVs) to share computation resources and ensure traceability of the sharing records, we apply smart contracts and blockchain technologies to achieve secure sharing management. Numerical results demonstrate that the proposed approach outperforms the MAPPO approach by around 2% and effectively reduces approximately 20% of the latency of avatar task execution in UAV-assisted vehicular Metaverses.展开更多
Dear Editor,This letter presents a novel segmentation approach that leverages dendritic neurons to tackle the challenges of medical imaging segmentation.In this study,we enhance the segmentation accuracy based on a Se...Dear Editor,This letter presents a novel segmentation approach that leverages dendritic neurons to tackle the challenges of medical imaging segmentation.In this study,we enhance the segmentation accuracy based on a SegNet variant including an encoder-decoder structure,an upsampling index,and a deep supervision method.Furthermore,we introduce a dendritic neuron-based convolutional block to enable nonlinear feature mapping,thereby further improving the effectiveness of our approach.展开更多
Accurate soil moisture(SM)prediction is critical for understanding hydrological processes.Physics-based(PB)models exhibit large uncertainties in SM predictions arising from uncertain parameterizations and insufficient...Accurate soil moisture(SM)prediction is critical for understanding hydrological processes.Physics-based(PB)models exhibit large uncertainties in SM predictions arising from uncertain parameterizations and insufficient representation of land-surface processes.In addition to PB models,deep learning(DL)models have been widely used in SM predictions recently.However,few pure DL models have notably high success rates due to lacking physical information.Thus,we developed hybrid models to effectively integrate the outputs of PB models into DL models to improve SM predictions.To this end,we first developed a hybrid model based on the attention mechanism to take advantage of PB models at each forecast time scale(attention model).We further built an ensemble model that combined the advantages of different hybrid schemes(ensemble model).We utilized SM forecasts from the Global Forecast System to enhance the convolutional long short-term memory(ConvLSTM)model for 1–16 days of SM predictions.The performances of the proposed hybrid models were investigated and compared with two existing hybrid models.The results showed that the attention model could leverage benefits of PB models and achieved the best predictability of drought events among the different hybrid models.Moreover,the ensemble model performed best among all hybrid models at all forecast time scales and different soil conditions.It is highlighted that the ensemble model outperformed the pure DL model over 79.5%of in situ stations for 16-day predictions.These findings suggest that our proposed hybrid models can adequately exploit the benefits of PB model outputs to aid DL models in making SM predictions.展开更多
The scarcity of in-situ ocean observations poses a challenge for real-time information acquisition in the ocean.Among the crucial hydroacoustic environmental parameters,ocean sound velocity exhibits significant spatia...The scarcity of in-situ ocean observations poses a challenge for real-time information acquisition in the ocean.Among the crucial hydroacoustic environmental parameters,ocean sound velocity exhibits significant spatial and temporal variability and it is highly relevant to oceanic research.In this study,we propose a new data-driven approach,leveraging deep learning techniques,for the prediction of sound velocity fields(SVFs).Our novel spatiotemporal prediction model,STLSTM-SA,combines Spatiotemporal Long Short-Term Memory(ST-LSTM) with a self-attention mechanism to enable accurate and real-time prediction of SVFs.To circumvent the limited amount of observational data,we employ transfer learning by first training the model using reanalysis datasets,followed by fine-tuning it using in-situ analysis data to obtain the final prediction model.By utilizing the historical 12-month SVFs as input,our model predicts the SVFs for the subsequent three months.We compare the performance of five models:Artificial Neural Networks(ANN),Long ShortTerm Memory(LSTM),Convolutional LSTM(ConvLSTM),ST-LSTM,and our proposed ST-LSTM-SA model in a test experiment spanning 2019 to 2022.Our results demonstrate that the ST-LSTM-SA model significantly improves the prediction accuracy and stability of sound velocity in both temporal and spatial dimensions.The ST-LSTM-SA model not only accurately predicts the ocean sound velocity field(SVF),but also provides valuable insights for spatiotemporal prediction of other oceanic environmental variables.展开更多
Thunderstorm gusts are a common form of severe convective weather in the warm season in North China,and it is of great importance to correctly forecast them.At present,the forecasting of thunderstorm gusts is mainly b...Thunderstorm gusts are a common form of severe convective weather in the warm season in North China,and it is of great importance to correctly forecast them.At present,the forecasting of thunderstorm gusts is mainly based on traditional subjective methods,which fails to achieve high-resolution and high-frequency gridded forecasts based on multiple observation sources.In this paper,we propose a deep learning method called Thunderstorm Gusts TransU-net(TGTransUnet)to forecast thunderstorm gusts in North China based on multi-source gridded product data from the Institute of Urban Meteorology(IUM)with a lead time of 1 to 6 h.To determine the specific range of thunderstorm gusts,we combine three meteorological variables:radar reflectivity factor,lightning location,and 1-h maximum instantaneous wind speed from automatic weather stations(AWSs),and obtain a reasonable ground truth of thunderstorm gusts.Then,we transform the forecasting problem into an image-to-image problem in deep learning under the TG-TransUnet architecture,which is based on convolutional neural networks and a transformer.The analysis and forecast data of the enriched multi-source gridded comprehensive forecasting system for the period 2021–23 are then used as training,validation,and testing datasets.Finally,the performance of TG-TransUnet is compared with other methods.The results show that TG-TransUnet has the best prediction results at 1–6 h.The IUM is currently using this model to support the forecasting of thunderstorm gusts in North China.展开更多
The great potentials of massive Multiple-Input Multiple-Output(MIMO)in Frequency Division Duplex(FDD)mode can be fully exploited when the downlink Channel State Information(CSI)is available at base stations.However,th...The great potentials of massive Multiple-Input Multiple-Output(MIMO)in Frequency Division Duplex(FDD)mode can be fully exploited when the downlink Channel State Information(CSI)is available at base stations.However,the accurate CsI is difficult to obtain due to the large amount of feedback overhead caused by massive antennas.In this paper,we propose a deep learning based joint channel estimation and feedback framework,which comprehensively realizes the estimation,compression,and reconstruction of downlink channels in FDD massive MIMO systems.Two networks are constructed to perform estimation and feedback explicitly and implicitly.The explicit network adopts a multi-Signal-to-Noise-Ratios(SNRs)technique to obtain a single trained channel estimation subnet that works well with different SNRs and employs a deep residual network to reconstruct the channels,while the implicit network directly compresses pilots and sends them back to reduce network parameters.Quantization module is also designed to generate data-bearing bitstreams.Simulation results show that the two proposed networks exhibit excellent performance of reconstruction and are robust to different environments and quantization errors.展开更多
Battery management systems(BMSs) play a vital role in ensuring efficient and reliable operations of lithium-ion batteries.The main function of the BMSs is to estimate battery states and diagnose battery health using b...Battery management systems(BMSs) play a vital role in ensuring efficient and reliable operations of lithium-ion batteries.The main function of the BMSs is to estimate battery states and diagnose battery health using battery open-circuit voltage(OCV).However,acquiring the complete OCV data online can be a challenging endeavor due to the time-consuming measurement process or the need for specific operating conditions required by OCV estimation models.In addressing these concerns,this study introduces a deep neural network-combined framework for accurate and robust OCV estimation,utilizing partial daily charging data.We incorporate a generative deep learning model to extract aging-related features from data and generate high-fidelity OCV curves.Correlation analysis is employed to identify the optimal partial charging data,optimizing the OCV estimation precision while preserving exceptional flexibility.The validation results,using data from nickel-cobalt-magnesium(NCM) batteries,illustrate the accurate estimation of the complete OCV-capacity curve,with an average root mean square errors(RMSE) of less than 3 mAh.Achieving this level of precision for OCV estimation requires only around 50 s collection of partial charging data.Further validations on diverse battery types operating under various conditions confirm the effectiveness of our proposed method.Additional cases of precise health diagnosis based on OCV highlight the significance of conducting online OCV estimation.Our method provides a flexible approach to achieve complete OCV estimation and holds promise for generalization to other tasks in BMSs.展开更多
The advent of pandemics such as COVID-19 significantly impacts human behaviour and lives every day.Therefore,it is essential to make medical services connected to internet,available in every remote location during the...The advent of pandemics such as COVID-19 significantly impacts human behaviour and lives every day.Therefore,it is essential to make medical services connected to internet,available in every remote location during these situations.Also,the security issues in the Internet of Medical Things(IoMT)used in these service,make the situation even more critical because cyberattacks on the medical devices might cause treatment delays or clinical failures.Hence,services in the healthcare ecosystem need rapid,uninterrupted,and secure facilities.The solution provided in this research addresses security concerns and services availability for patients with critical health in remote areas.This research aims to develop an intelligent Software Defined Networks(SDNs)enabled secure framework for IoT healthcare ecosystem.We propose a hybrid of machine learning and deep learning techniques(DNN+SVM)to identify network intrusions in the sensor-based healthcare data.In addition,this system can efficiently monitor connected devices and suspicious behaviours.Finally,we evaluate the performance of our proposed framework using various performance metrics based on the healthcare application scenarios.the experimental results show that the proposed approach effectively detects and mitigates attacks in the SDN-enabled IoT networks and performs better that other state-of-art-approaches.展开更多
Although Federated Deep Learning(FDL)enables distributed machine learning in the Internet of Vehicles(IoV),it requires multiple clients to upload model parameters,thus still existing unavoidable communication overhead...Although Federated Deep Learning(FDL)enables distributed machine learning in the Internet of Vehicles(IoV),it requires multiple clients to upload model parameters,thus still existing unavoidable communication overhead and data privacy risks.The recently proposed Swarm Learning(SL)provides a decentralized machine learning approach for unit edge computing and blockchain-based coordination.A Swarm-Federated Deep Learning framework in the IoV system(IoV-SFDL)that integrates SL into the FDL framework is proposed in this paper.The IoV-SFDL organizes vehicles to generate local SL models with adjacent vehicles based on the blockchain empowered SL,then aggregates the global FDL model among different SL groups with a credibility weights prediction algorithm.Extensive experimental results show that compared with the baseline frameworks,the proposed IoV-SFDL framework reduces the overhead of client-to-server communication by 16.72%,while the model performance improves by about 5.02%for the same training iterations.展开更多
In recent years,deep learning methods have gradually been applied to prediction tasks related to Arctic sea ice concentration,but relatively little research has been conducted for larger spatial and temporal scales,ma...In recent years,deep learning methods have gradually been applied to prediction tasks related to Arctic sea ice concentration,but relatively little research has been conducted for larger spatial and temporal scales,mainly due to the limited time coverage of observations and reanalysis data.Meanwhile,deep learning predictions of sea ice thickness(SIT)have yet to receive ample attention.In this study,two data-driven deep learning(DL)models are built based on the ConvLSTM and fully convolutional U-net(FC-Unet)algorithms and trained using CMIP6 historical simulations for transfer learning and fine-tuned using reanalysis/observations.These models enable monthly predictions of Arctic SIT without considering the complex physical processes involved.Through comprehensive assessments of prediction skills by season and region,the results suggest that using a broader set of CMIP6 data for transfer learning,as well as incorporating multiple climate variables as predictors,contribute to better prediction results,although both DL models can effectively predict the spatiotemporal features of SIT anomalies.Regarding the predicted SIT anomalies of the FC-Unet model,the spatial correlations with reanalysis reach an average level of 89%over all months,while the temporal anomaly correlation coefficients are close to unity in most cases.The models also demonstrate robust performances in predicting SIT and SIE during extreme events.The effectiveness and reliability of the proposed deep transfer learning models in predicting Arctic SIT can facilitate more accurate pan-Arctic predictions,aiding climate change research and real-time business applications.展开更多
The Tibetan Plateau(TP)is a prevalent region for convection systems due to its unique thermodynamic forcing.This study investigated isolated deep convections(IDCs),which have a smaller spatial and temporal size than m...The Tibetan Plateau(TP)is a prevalent region for convection systems due to its unique thermodynamic forcing.This study investigated isolated deep convections(IDCs),which have a smaller spatial and temporal size than mesoscale convective systems(MCSs),over the TP in the rainy season(June-September)during 2001–2020.The authors used satellite precipitation and brightness temperature observations from the Global Precipitation Measurement mission.Results show that IDCs mainly concentrate over the southern TP.The IDC number per rainy season decreases from around 140 over the southern TP to around 10 over the northern TP,with an average 54.2.The initiation time of IDCs exhibits an obvious diurnal cycle,with the peak at 1400–1500 LST and the valley at 0900–1000 LST.Most IDCs last less than five hours and more than half appear for only one hour.IDCs generally have a cold cloud area of 7422.9 km^(2),containing a precipitation area of approximately 65%.The larger the IDC,the larger the fraction of intense precipitation it contains.IDCs contribute approximately 20%–30%to total precipitation and approximately 30%–40%to extreme precipitation over the TP,with a larger percentage in July and August than in June and September.In terms of spatial distribution,IDCs contribute more to both total precipitation and extreme precipitation over the TP compared to the surrounding plain regions.IDCs over the TP account for a larger fraction than MCSs,indicating the important role of IDCs over the region.展开更多
基金Project supported by the National Natural Science Foundation of China (Grant No.12172226)。
文摘The evolution of the probability density function of a stochastic dynamical system over time can be described by a Fokker–Planck–Kolmogorov(FPK) equation, the solution of which determines the distribution of macroscopic variables in the stochastic dynamic system. Traditional methods for solving these equations often struggle with computational efficiency and scalability, particularly in high-dimensional contexts. To address these challenges, this paper proposes a novel deep learning method based on prior knowledge with dual training to solve the stationary FPK equations. Initially, the neural network is pre-trained through the prior knowledge obtained by Monte Carlo simulation(MCS). Subsequently, the second training phase incorporates the FPK differential operator into the loss function, while a supervisory term consisting of local maximum points is specifically included to mitigate the generation of zero solutions. This dual-training strategy not only expedites convergence but also enhances computational efficiency, making the method well-suited for high-dimensional systems. Numerical examples, including two different two-dimensional(2D), six-dimensional(6D), and eight-dimensional(8D) systems, are conducted to assess the efficacy of the proposed method. The results demonstrate robust performance in terms of both computational speed and accuracy for solving FPK equations in the first three systems. While the method is also applicable to high-dimensional systems, such as 8D, it should be noted that computational efficiency may be marginally compromised due to data volume constraints.
基金Shenzhen Science and Technology Program,Grant/Award Number:ZDSYS20211021111415025Shenzhen Institute of Artificial Intelligence and Robotics for SocietyYouth Science and Technology Talents Development Project of Guizhou Education Department,Grant/Award Number:QianJiaoheKYZi[2018]459。
文摘Facial beauty analysis is an important topic in human society.It may be used as a guidance for face beautification applications such as cosmetic surgery.Deep neural networks(DNNs)have recently been adopted for facial beauty analysis and have achieved remarkable performance.However,most existing DNN-based models regard facial beauty analysis as a normal classification task.They ignore important prior knowledge in traditional machine learning models which illustrate the significant contribution of the geometric features in facial beauty analysis.To be specific,landmarks of the whole face and facial organs are introduced to extract geometric features to make the decision.Inspired by this,we introduce a novel dual-branch network for facial beauty analysis:one branch takes the Swin Transformer as the backbone to model the full face and global patterns,and another branch focuses on the masked facial organs with the residual network to model the local patterns of certain facial parts.Additionally,the designed multi-scale feature fusion module can further facilitate our network to learn complementary semantic information between the two branches.In model optimisation,we propose a hybrid loss function,where especially geometric regulation is introduced by regressing the facial landmarks and it can force the extracted features to convey facial geometric features.Experiments performed on the SCUT-FBP5500 dataset and the SCUT-FBP dataset demonstrate that our model outperforms the state-of-the-art convolutional neural networks models,which proves the effectiveness of the proposed geometric regularisation and dual-branch structure with the hybrid network.To the best of our knowledge,this is the first study to introduce a Vision Transformer into the facial beauty analysis task.
文摘Plug-in Hybrid Electric Vehicles(PHEVs)represent an innovative breed of transportation,harnessing diverse power sources for enhanced performance.Energy management strategies(EMSs)that coordinate and control different energy sources is a critical component of PHEV control technology,directly impacting overall vehicle performance.This study proposes an improved deep reinforcement learning(DRL)-based EMSthat optimizes realtime energy allocation and coordinates the operation of multiple power sources.Conventional DRL algorithms struggle to effectively explore all possible state-action combinations within high-dimensional state and action spaces.They often fail to strike an optimal balance between exploration and exploitation,and their assumption of a static environment limits their ability to adapt to changing conditions.Moreover,these algorithms suffer from low sample efficiency.Collectively,these factors contribute to convergence difficulties,low learning efficiency,and instability.To address these challenges,the Deep Deterministic Policy Gradient(DDPG)algorithm is enhanced using entropy regularization and a summation tree-based Prioritized Experience Replay(PER)method,aiming to improve exploration performance and learning efficiency from experience samples.Additionally,the correspondingMarkovDecision Process(MDP)is established.Finally,an EMSbased on the improvedDRLmodel is presented.Comparative simulation experiments are conducted against rule-based,optimization-based,andDRL-based EMSs.The proposed strategy exhibitsminimal deviation fromthe optimal solution obtained by the dynamic programming(DP)strategy that requires global information.In the typical driving scenarios based onWorld Light Vehicle Test Cycle(WLTC)and New European Driving Cycle(NEDC),the proposed method achieved a fuel consumption of 2698.65 g and an Equivalent Fuel Consumption(EFC)of 2696.77 g.Compared to the DP strategy baseline,the proposed method improved the fuel efficiency variances(FEV)by 18.13%,15.1%,and 8.37%over the Deep QNetwork(DQN),Double DRL(DDRL),and original DDPG methods,respectively.The observational outcomes demonstrate that the proposed EMS based on improved DRL framework possesses good real-time performance,stability,and reliability,effectively optimizing vehicle economy and fuel consumption.
文摘For some important object recognition applications such as intelligent robots and unmanned driving, images are collected on a consecutive basis and associated among themselves, besides, the scenes have steady prior features. Yet existing technologies do not take full advantage of this information. In order to take object recognition further than existing algorithms in the above application, an object recognition method that fuses temporal sequence with scene priori information is proposed. This method first employs YOLOv3 as the basic algorithm to recognize objects in single-frame images, then the DeepSort algorithm to establish association among potential objects recognized in images of different moments, and finally the confidence fusion method and temporal boundary processing method designed herein to fuse, at the decision level, temporal sequence information with scene priori information. Experiments using public datasets and self-built industrial scene datasets show that due to the expansion of information sources, the quality of single-frame images has less impact on the recognition results, whereby the object recognition is greatly improved. It is presented herein as a widely applicable framework for the fusion of information under multiple classes. All the object recognition algorithms that output object class, location information and recognition confidence at the same time can be integrated into this information fusion framework to improve performance.
基金This work was supported by National Key R&D Program of China(2022YFB3605103)the National Natural Science Foundation of China(62204241,U22A2084,62121005,and 61827813)+3 种基金the Natural Science Foundation of Jilin Province(20230101345JC,20230101360JC,and 20230101107JC)the Youth Innovation Promotion Association of CAS(2023223)the Young Elite Scientist Sponsorship Program By CAST(YESS20200182)the CAS Talents Program(E30122E4M0).
文摘240 nm AlGaN-based micro-LEDs with different sizes are designed and fabricated.Then,the external quantum efficiency(EQE)and light extraction efficiency(LEE)are systematically investigated by comparing size and edge effects.Here,it is revealed that the peak optical output power increases by 81.83%with the size shrinking from 50.0 to 25.0μm.Thereinto,the LEE increases by 26.21%and the LEE enhancement mainly comes from the sidewall light extraction.Most notably,transversemagnetic(TM)mode light intensifies faster as the size shrinks due to the tilted mesa side-wall and Al reflector design.However,when it turns to 12.5μm sized micro-LEDs,the output power is lower than 25.0μm sized ones.The underlying mechanism is that even though protected by SiO2 passivation,the edge effect which leads to current leakage and Shockley-Read-Hall(SRH)recombination deteriorates rapidly with the size further shrinking.Moreover,the ratio of the p-contact area to mesa area is much lower,which deteriorates the p-type current spreading at the mesa edge.These findings show a role of thumb for the design of high efficiency micro-LEDs with wavelength below 250 nm,which will pave the way for wide applications of deep ultraviolet(DUV)micro-LEDs.
基金supported by the National Natural Science Foundation of China(62375144 and 61875092)Tianjin Foundation of Natural Science(21JCYBJC00260)Beijing-Tianjin-Hebei Basic Research Cooperation Special Program(19JCZDJC65300).
文摘Limited by the dynamic range of the detector,saturation artifacts usually occur in optical coherence tomography(OCT)imaging for high scattering media.The available methods are difficult to remove saturation artifacts and restore texture completely in OCT images.We proposed a deep learning-based inpainting method of saturation artifacts in this paper.The generation mechanism of saturation artifacts was analyzed,and experimental and simulated datasets were built based on the mechanism.Enhanced super-resolution generative adversarial networks were trained by the clear–saturated phantom image pairs.The perfect reconstructed results of experimental zebrafish and thyroid OCT images proved its feasibility,strong generalization,and robustness.
基金Dao-Bing Wang was supported by the Beijing Natural Science Foundation Project(No.3222030)the National Natural Science Foundation of China(No.52274002)+1 种基金the PetroChina Science and Technology Innovation Foundation Project(No.2021DQ02-0201)Fu-Jian Zhou was supported by the National Natural Science Foundation of China(No.52174045).
文摘Deep and ultra-deep reservoirs have gradually become the primary focus of hydrocarbon exploration as a result of a series of significant discoveries in deep hydrocarbon exploration worldwide.These reservoirs present unique challenges due to their deep burial depth(4500-8882 m),low matrix permeability,complex crustal stress conditions,high temperature and pressure(HTHP,150-200℃,105-155 MPa),coupled with high salinity of formation water.Consequently,the costs associated with their exploitation and development are exceptionally high.In deep and ultra-deep reservoirs,hydraulic fracturing is commonly used to achieve high and stable production.During hydraulic fracturing,a substantial volume of fluid is injected into the reservoir.However,statistical analysis reveals that the flowback rate is typically less than 30%,leaving the majority of the fluid trapped within the reservoir.Therefore,hydraulic fracturing in deep reservoirs not only enhances the reservoir permeability by creating artificial fractures but also damages reservoirs due to the fracturing fluids involved.The challenging“three-high”environment of a deep reservoir,characterized by high temperature,high pressure,and high salinity,exacerbates conventional forms of damage,including water sensitivity,retention of fracturing fluids,rock creep,and proppant breakage.In addition,specific damage mechanisms come into play,such as fracturing fluid decomposition at elevated temperatures and proppant diagenetic reactions at HTHP conditions.Presently,the foremost concern in deep oil and gas development lies in effectively assessing the damage inflicted on these reservoirs by hydraulic fracturing,comprehending the underlying mechanisms,and selecting appropriate solutions.It's noteworthy that the majority of existing studies on reservoir damage primarily focus on conventional reservoirs,with limited attention given to deep reservoirs and a lack of systematic summaries.In light of this,our approach entails initially summarizing the current knowledge pertaining to the types of fracturing fluids employed in deep and ultra-deep reservoirs.Subsequently,we delve into a systematic examination of the damage processes and mechanisms caused by fracturing fluids within the context of hydraulic fracturing in deep reservoirs,taking into account the unique reservoir characteristics of high temperature,high pressure,and high in-situ stress.In addition,we provide an overview of research progress related to high-temperature deep reservoir fracturing fluid and the damage of aqueous fracturing fluids to rock matrix,both artificial and natural fractures,and sand-packed fractures.We conclude by offering a summary of current research advancements and future directions,which hold significant potential for facilitating the efficient development of deep oil and gas reservoirs while effectively mitigating reservoir damage.
基金financially supported by the National Natural Science Foundation of China (Nos.51974023 and52374321)the funding of State Key Laboratory of Advanced Metallurgy,University of Science and Technology Beijing,China (No.41620007)。
文摘The amount of oxygen blown into the converter is one of the key parameters for the control of the converter blowing process,which directly affects the tap-to-tap time of converter. In this study, a hybrid model based on oxygen balance mechanism (OBM) and deep neural network (DNN) was established for predicting oxygen blowing time in converter. A three-step method was utilized in the hybrid model. First, the oxygen consumption volume was predicted by the OBM model and DNN model, respectively. Second, a more accurate oxygen consumption volume was obtained by integrating the OBM model and DNN model. Finally, the converter oxygen blowing time was calculated according to the oxygen consumption volume and the oxygen supply intensity of each heat. The proposed hybrid model was verified using the actual data collected from an integrated steel plant in China, and compared with multiple linear regression model, OBM model, and neural network model including extreme learning machine, back propagation neural network, and DNN. The test results indicate that the hybrid model with a network structure of 3 hidden layer layers, 32-16-8 neurons per hidden layer, and 0.1 learning rate has the best prediction accuracy and stronger generalization ability compared with other models. The predicted hit ratio of oxygen consumption volume within the error±300 m^(3)is 96.67%;determination coefficient (R^(2)) and root mean square error (RMSE) are0.6984 and 150.03 m^(3), respectively. The oxygen blow time prediction hit ratio within the error±0.6 min is 89.50%;R2and RMSE are0.9486 and 0.3592 min, respectively. As a result, the proposed model can effectively predict the oxygen consumption volume and oxygen blowing time in the converter.
基金supported in part by NSFC (62102099, U22A2054, 62101594)in part by the Pearl River Talent Recruitment Program (2021QN02S643)+9 种基金Guangzhou Basic Research Program (2023A04J1699)in part by the National Research Foundation, SingaporeInfocomm Media Development Authority under its Future Communications Research Development ProgrammeDSO National Laboratories under the AI Singapore Programme under AISG Award No AISG2-RP-2020-019Energy Research Test-Bed and Industry Partnership Funding Initiative, Energy Grid (EG) 2.0 programmeDesCartes and the Campus for Research Excellence and Technological Enterprise (CREATE) programmeMOE Tier 1 under Grant RG87/22in part by the Singapore University of Technology and Design (SUTD) (SRG-ISTD-2021- 165)in part by the SUTD-ZJU IDEA Grant SUTD-ZJU (VP) 202102in part by the Ministry of Education, Singapore, through its SUTD Kickstarter Initiative (SKI 20210204)。
文摘Avatars, as promising digital representations and service assistants of users in Metaverses, can enable drivers and passengers to immerse themselves in 3D virtual services and spaces of UAV-assisted vehicular Metaverses. However, avatar tasks include a multitude of human-to-avatar and avatar-to-avatar interactive applications, e.g., augmented reality navigation,which consumes intensive computing resources. It is inefficient and impractical for vehicles to process avatar tasks locally. Fortunately, migrating avatar tasks to the nearest roadside units(RSU)or unmanned aerial vehicles(UAV) for execution is a promising solution to decrease computation overhead and reduce task processing latency, while the high mobility of vehicles brings challenges for vehicles to independently perform avatar migration decisions depending on current and future vehicle status. To address these challenges, in this paper, we propose a novel avatar task migration system based on multi-agent deep reinforcement learning(MADRL) to execute immersive vehicular avatar tasks dynamically. Specifically, we first formulate the problem of avatar task migration from vehicles to RSUs/UAVs as a partially observable Markov decision process that can be solved by MADRL algorithms. We then design the multi-agent proximal policy optimization(MAPPO) approach as the MADRL algorithm for the avatar task migration problem. To overcome slow convergence resulting from the curse of dimensionality and non-stationary issues caused by shared parameters in MAPPO, we further propose a transformer-based MAPPO approach via sequential decision-making models for the efficient representation of relationships among agents. Finally, to motivate terrestrial or non-terrestrial edge servers(e.g., RSUs or UAVs) to share computation resources and ensure traceability of the sharing records, we apply smart contracts and blockchain technologies to achieve secure sharing management. Numerical results demonstrate that the proposed approach outperforms the MAPPO approach by around 2% and effectively reduces approximately 20% of the latency of avatar task execution in UAV-assisted vehicular Metaverses.
基金partially supported by the Japan Society for the Promotion of Science(JSPS)KAKENHI(JP22H03643)Japan Science and Technology Agency(JST)Support for Pioneering Research Initiated by the Next Generation(SPRING)(JPMJSP2145)JST through the Establishment of University Fellowships Towards the Creation of Science Technology Innovation(JPMJFS2115)。
文摘Dear Editor,This letter presents a novel segmentation approach that leverages dendritic neurons to tackle the challenges of medical imaging segmentation.In this study,we enhance the segmentation accuracy based on a SegNet variant including an encoder-decoder structure,an upsampling index,and a deep supervision method.Furthermore,we introduce a dendritic neuron-based convolutional block to enable nonlinear feature mapping,thereby further improving the effectiveness of our approach.
基金supported by the Natural Science Foundation of China(Grant Nos.42088101 and 42205149)Zhongwang WEI was supported by the Natural Science Foundation of China(Grant No.42075158)+1 种基金Wei SHANGGUAN was supported by the Natural Science Foundation of China(Grant No.41975122)Yonggen ZHANG was supported by the National Natural Science Foundation of Tianjin(Grant No.20JCQNJC01660).
文摘Accurate soil moisture(SM)prediction is critical for understanding hydrological processes.Physics-based(PB)models exhibit large uncertainties in SM predictions arising from uncertain parameterizations and insufficient representation of land-surface processes.In addition to PB models,deep learning(DL)models have been widely used in SM predictions recently.However,few pure DL models have notably high success rates due to lacking physical information.Thus,we developed hybrid models to effectively integrate the outputs of PB models into DL models to improve SM predictions.To this end,we first developed a hybrid model based on the attention mechanism to take advantage of PB models at each forecast time scale(attention model).We further built an ensemble model that combined the advantages of different hybrid schemes(ensemble model).We utilized SM forecasts from the Global Forecast System to enhance the convolutional long short-term memory(ConvLSTM)model for 1–16 days of SM predictions.The performances of the proposed hybrid models were investigated and compared with two existing hybrid models.The results showed that the attention model could leverage benefits of PB models and achieved the best predictability of drought events among the different hybrid models.Moreover,the ensemble model performed best among all hybrid models at all forecast time scales and different soil conditions.It is highlighted that the ensemble model outperformed the pure DL model over 79.5%of in situ stations for 16-day predictions.These findings suggest that our proposed hybrid models can adequately exploit the benefits of PB model outputs to aid DL models in making SM predictions.
基金supported by the National Natural Science Foundation of China(Grant No.42004030)Basic Scientific Fund for National Public Research Institutes of China(Grant No.2022S03)+1 种基金Science and Technology Innovation Project(LSKJ202205102)funded by Laoshan Laboratory,and the National Key Research and Development Program of China(2020YFB0505805).
文摘The scarcity of in-situ ocean observations poses a challenge for real-time information acquisition in the ocean.Among the crucial hydroacoustic environmental parameters,ocean sound velocity exhibits significant spatial and temporal variability and it is highly relevant to oceanic research.In this study,we propose a new data-driven approach,leveraging deep learning techniques,for the prediction of sound velocity fields(SVFs).Our novel spatiotemporal prediction model,STLSTM-SA,combines Spatiotemporal Long Short-Term Memory(ST-LSTM) with a self-attention mechanism to enable accurate and real-time prediction of SVFs.To circumvent the limited amount of observational data,we employ transfer learning by first training the model using reanalysis datasets,followed by fine-tuning it using in-situ analysis data to obtain the final prediction model.By utilizing the historical 12-month SVFs as input,our model predicts the SVFs for the subsequent three months.We compare the performance of five models:Artificial Neural Networks(ANN),Long ShortTerm Memory(LSTM),Convolutional LSTM(ConvLSTM),ST-LSTM,and our proposed ST-LSTM-SA model in a test experiment spanning 2019 to 2022.Our results demonstrate that the ST-LSTM-SA model significantly improves the prediction accuracy and stability of sound velocity in both temporal and spatial dimensions.The ST-LSTM-SA model not only accurately predicts the ocean sound velocity field(SVF),but also provides valuable insights for spatiotemporal prediction of other oceanic environmental variables.
基金supported in part by the Beijing Natural Science Foundation(Grant No.8222051)the National Key R&D Program of China(Grant No.2022YFC3004103)+2 种基金the National Natural Foundation of China(Grant Nos.42275003 and 42275012)the China Meteorological Administration Key Innovation Team(Grant Nos.CMA2022ZD04 and CMA2022ZD07)the Beijing Science and Technology Program(Grant No.Z221100005222012).
文摘Thunderstorm gusts are a common form of severe convective weather in the warm season in North China,and it is of great importance to correctly forecast them.At present,the forecasting of thunderstorm gusts is mainly based on traditional subjective methods,which fails to achieve high-resolution and high-frequency gridded forecasts based on multiple observation sources.In this paper,we propose a deep learning method called Thunderstorm Gusts TransU-net(TGTransUnet)to forecast thunderstorm gusts in North China based on multi-source gridded product data from the Institute of Urban Meteorology(IUM)with a lead time of 1 to 6 h.To determine the specific range of thunderstorm gusts,we combine three meteorological variables:radar reflectivity factor,lightning location,and 1-h maximum instantaneous wind speed from automatic weather stations(AWSs),and obtain a reasonable ground truth of thunderstorm gusts.Then,we transform the forecasting problem into an image-to-image problem in deep learning under the TG-TransUnet architecture,which is based on convolutional neural networks and a transformer.The analysis and forecast data of the enriched multi-source gridded comprehensive forecasting system for the period 2021–23 are then used as training,validation,and testing datasets.Finally,the performance of TG-TransUnet is compared with other methods.The results show that TG-TransUnet has the best prediction results at 1–6 h.The IUM is currently using this model to support the forecasting of thunderstorm gusts in North China.
基金supported in part by the National Natural Science Foundation of China(NSFC)under Grants 61941104,61921004the Key Research and Development Program of Shandong Province under Grant 2020CXGC010108+1 种基金the Southeast University-China Mobile Research Institute Joint Innovation Centersupported in part by the Scientific Research Foundation of Graduate School of Southeast University under Grant YBPY2118.
文摘The great potentials of massive Multiple-Input Multiple-Output(MIMO)in Frequency Division Duplex(FDD)mode can be fully exploited when the downlink Channel State Information(CSI)is available at base stations.However,the accurate CsI is difficult to obtain due to the large amount of feedback overhead caused by massive antennas.In this paper,we propose a deep learning based joint channel estimation and feedback framework,which comprehensively realizes the estimation,compression,and reconstruction of downlink channels in FDD massive MIMO systems.Two networks are constructed to perform estimation and feedback explicitly and implicitly.The explicit network adopts a multi-Signal-to-Noise-Ratios(SNRs)technique to obtain a single trained channel estimation subnet that works well with different SNRs and employs a deep residual network to reconstruct the channels,while the implicit network directly compresses pilots and sends them back to reduce network parameters.Quantization module is also designed to generate data-bearing bitstreams.Simulation results show that the two proposed networks exhibit excellent performance of reconstruction and are robust to different environments and quantization errors.
基金This work was supported by the National Key R&D Program of China(2021YFB2402002)the Beijing Natural Science Foundation(L223013)the Chongqing Automobile Collaborative Innovation Centre(No.2022CDJDX-004).
文摘Battery management systems(BMSs) play a vital role in ensuring efficient and reliable operations of lithium-ion batteries.The main function of the BMSs is to estimate battery states and diagnose battery health using battery open-circuit voltage(OCV).However,acquiring the complete OCV data online can be a challenging endeavor due to the time-consuming measurement process or the need for specific operating conditions required by OCV estimation models.In addressing these concerns,this study introduces a deep neural network-combined framework for accurate and robust OCV estimation,utilizing partial daily charging data.We incorporate a generative deep learning model to extract aging-related features from data and generate high-fidelity OCV curves.Correlation analysis is employed to identify the optimal partial charging data,optimizing the OCV estimation precision while preserving exceptional flexibility.The validation results,using data from nickel-cobalt-magnesium(NCM) batteries,illustrate the accurate estimation of the complete OCV-capacity curve,with an average root mean square errors(RMSE) of less than 3 mAh.Achieving this level of precision for OCV estimation requires only around 50 s collection of partial charging data.Further validations on diverse battery types operating under various conditions confirm the effectiveness of our proposed method.Additional cases of precise health diagnosis based on OCV highlight the significance of conducting online OCV estimation.Our method provides a flexible approach to achieve complete OCV estimation and holds promise for generalization to other tasks in BMSs.
文摘The advent of pandemics such as COVID-19 significantly impacts human behaviour and lives every day.Therefore,it is essential to make medical services connected to internet,available in every remote location during these situations.Also,the security issues in the Internet of Medical Things(IoMT)used in these service,make the situation even more critical because cyberattacks on the medical devices might cause treatment delays or clinical failures.Hence,services in the healthcare ecosystem need rapid,uninterrupted,and secure facilities.The solution provided in this research addresses security concerns and services availability for patients with critical health in remote areas.This research aims to develop an intelligent Software Defined Networks(SDNs)enabled secure framework for IoT healthcare ecosystem.We propose a hybrid of machine learning and deep learning techniques(DNN+SVM)to identify network intrusions in the sensor-based healthcare data.In addition,this system can efficiently monitor connected devices and suspicious behaviours.Finally,we evaluate the performance of our proposed framework using various performance metrics based on the healthcare application scenarios.the experimental results show that the proposed approach effectively detects and mitigates attacks in the SDN-enabled IoT networks and performs better that other state-of-art-approaches.
基金supported by the National Natural Science Foundation of China(NSFC)under Grant 62071179.
文摘Although Federated Deep Learning(FDL)enables distributed machine learning in the Internet of Vehicles(IoV),it requires multiple clients to upload model parameters,thus still existing unavoidable communication overhead and data privacy risks.The recently proposed Swarm Learning(SL)provides a decentralized machine learning approach for unit edge computing and blockchain-based coordination.A Swarm-Federated Deep Learning framework in the IoV system(IoV-SFDL)that integrates SL into the FDL framework is proposed in this paper.The IoV-SFDL organizes vehicles to generate local SL models with adjacent vehicles based on the blockchain empowered SL,then aggregates the global FDL model among different SL groups with a credibility weights prediction algorithm.Extensive experimental results show that compared with the baseline frameworks,the proposed IoV-SFDL framework reduces the overhead of client-to-server communication by 16.72%,while the model performance improves by about 5.02%for the same training iterations.
基金supported by the National Natural Science Foundation of China(Grant Nos.41976193 and 42176243).
文摘In recent years,deep learning methods have gradually been applied to prediction tasks related to Arctic sea ice concentration,but relatively little research has been conducted for larger spatial and temporal scales,mainly due to the limited time coverage of observations and reanalysis data.Meanwhile,deep learning predictions of sea ice thickness(SIT)have yet to receive ample attention.In this study,two data-driven deep learning(DL)models are built based on the ConvLSTM and fully convolutional U-net(FC-Unet)algorithms and trained using CMIP6 historical simulations for transfer learning and fine-tuned using reanalysis/observations.These models enable monthly predictions of Arctic SIT without considering the complex physical processes involved.Through comprehensive assessments of prediction skills by season and region,the results suggest that using a broader set of CMIP6 data for transfer learning,as well as incorporating multiple climate variables as predictors,contribute to better prediction results,although both DL models can effectively predict the spatiotemporal features of SIT anomalies.Regarding the predicted SIT anomalies of the FC-Unet model,the spatial correlations with reanalysis reach an average level of 89%over all months,while the temporal anomaly correlation coefficients are close to unity in most cases.The models also demonstrate robust performances in predicting SIT and SIE during extreme events.The effectiveness and reliability of the proposed deep transfer learning models in predicting Arctic SIT can facilitate more accurate pan-Arctic predictions,aiding climate change research and real-time business applications.
基金supported by the National Natural Science Foundation of China[grant number 42105064]the Second Tibetan Plateau Scientific Expedition and Research(STEP)program[grant number 2019QZKK0102]the special fund of the Yunnan University“double first-class”construction.
文摘The Tibetan Plateau(TP)is a prevalent region for convection systems due to its unique thermodynamic forcing.This study investigated isolated deep convections(IDCs),which have a smaller spatial and temporal size than mesoscale convective systems(MCSs),over the TP in the rainy season(June-September)during 2001–2020.The authors used satellite precipitation and brightness temperature observations from the Global Precipitation Measurement mission.Results show that IDCs mainly concentrate over the southern TP.The IDC number per rainy season decreases from around 140 over the southern TP to around 10 over the northern TP,with an average 54.2.The initiation time of IDCs exhibits an obvious diurnal cycle,with the peak at 1400–1500 LST and the valley at 0900–1000 LST.Most IDCs last less than five hours and more than half appear for only one hour.IDCs generally have a cold cloud area of 7422.9 km^(2),containing a precipitation area of approximately 65%.The larger the IDC,the larger the fraction of intense precipitation it contains.IDCs contribute approximately 20%–30%to total precipitation and approximately 30%–40%to extreme precipitation over the TP,with a larger percentage in July and August than in June and September.In terms of spatial distribution,IDCs contribute more to both total precipitation and extreme precipitation over the TP compared to the surrounding plain regions.IDCs over the TP account for a larger fraction than MCSs,indicating the important role of IDCs over the region.