Since the effectiveness of extracting fault features is not high under traditional bearing fault diagnosis method, a bearing fault diagnosis method based on Deep Auto-encoder Network (DAEN) optimized by Cloud Adaptive...Since the effectiveness of extracting fault features is not high under traditional bearing fault diagnosis method, a bearing fault diagnosis method based on Deep Auto-encoder Network (DAEN) optimized by Cloud Adaptive Particle Swarm Optimization (CAPSO) was proposed. On the basis of analyzing CAPSO and DAEN, the CAPSO-DAEN fault diagnosis model is built. The model uses the randomness and stability of CAPSO algorithm to optimize the connection weight of DAEN, to reduce the constraints on the weights and extract fault features adaptively. Finally, efficient and accurate fault diagnosis can be implemented with the Softmax classifier. The results of test show that the proposed method has higher diagnostic accuracy and more stable diagnosis results than those based on the DAEN, Support Vector Machine (SVM) and the Back Propagation algorithm (BP) under appropriate parameters.展开更多
To address the difficulties in fusing multi-mode sensor data for complex industrial machinery, an adaptive deep coupling convolutional auto-encoder (ADCCAE) fusion method was proposed. First, the multi-mode features e...To address the difficulties in fusing multi-mode sensor data for complex industrial machinery, an adaptive deep coupling convolutional auto-encoder (ADCCAE) fusion method was proposed. First, the multi-mode features extracted synchronously by the CCAE were stacked and fed to the multi-channel convolution layers for fusion. Then, the fused data was passed to all connection layers for compression and fed to the Softmax module for classification. Finally, the coupling loss function coefficients and the network parameters were optimized through an adaptive approach using the gray wolf optimization (GWO) algorithm. Experimental comparisons showed that the proposed ADCCAE fusion model was superior to existing models for multi-mode data fusion.展开更多
Intelligent diagnosis approaches with shallow architectural models play an essential role in healthcare.Deep Learning(DL)models with unsupervised learning concepts have been proposed because high-quality feature extra...Intelligent diagnosis approaches with shallow architectural models play an essential role in healthcare.Deep Learning(DL)models with unsupervised learning concepts have been proposed because high-quality feature extraction and adequate labelled details significantly influence shallow models.On the other hand,skin lesionbased segregation and disintegration procedures play an essential role in earlier skin cancer detection.However,artefacts,an unclear boundary,poor contrast,and different lesion sizes make detection difficult.To address the issues in skin lesion diagnosis,this study creates the UDLS-DDOA model,an intelligent Unsupervised Deep Learning-based Stacked Auto-encoder(UDLS)optimized by Dynamic Differential Annealed Optimization(DDOA).Pre-processing,segregation,feature removal or separation,and disintegration are part of the proposed skin lesion diagnosis model.Pre-processing of skin lesion images occurs at the initial level for noise removal in the image using the Top hat filter and painting methodology.Following that,a Fuzzy C-Means(FCM)segregation procedure is performed using a Quasi-Oppositional Elephant Herd Optimization(QOEHO)algorithm.Besides,a novel feature extraction technique using the UDLS technique is applied where the parameter tuning takes place using DDOA.In the end,the disintegration procedure would be accomplished using a SoftMax(SM)classifier.The UDLS-DDOA model is tested against the International Skin Imaging Collaboration(ISIC)dataset,and the experimental results are examined using various computational attributes.The simulation results demonstrated that the UDLS-DDOA model outperformed the compared methods significantly.展开更多
With the rapid development of deep learning methods, the data-driven approach has shown powerful advantages over the model-driven one. In this paper, we propose an end-to-end autoencoder communication system based on ...With the rapid development of deep learning methods, the data-driven approach has shown powerful advantages over the model-driven one. In this paper, we propose an end-to-end autoencoder communication system based on Deep Residual Shrinkage Networks (DRSNs), where neural networks (DNNs) are used to implement the coding, decoding, modulation and demodulation functions of the communication system. Our proposed autoencoder communication system can better reduce the signal noise by adding an “attention mechanism” and “soft thresholding” modules and has better performance at various signal-to-noise ratios (SNR). Also, we have shown through comparative experiments that the system can operate at moderate block lengths and support different throughputs. It has been shown to work efficiently in the AWGN channel. Simulation results show that our model has a higher Bit-Error-Rate (BER) gain and greatly improved decoding performance compared to conventional modulation and classical autoencoder systems at various signal-to-noise ratios.展开更多
Healthcare systems nowadays depend on IoT sensors for sending data over the internet as a common practice.Encryption ofmedical images is very important to secure patient information.Encrypting these images consumes a ...Healthcare systems nowadays depend on IoT sensors for sending data over the internet as a common practice.Encryption ofmedical images is very important to secure patient information.Encrypting these images consumes a lot of time onedge computing;therefore,theuse of anauto-encoder for compressionbefore encodingwill solve such a problem.In this paper,we use an auto-encoder to compress amedical image before encryption,and an encryption output(vector)is sent out over the network.On the other hand,a decoder was used to reproduce the original image back after the vector was received and decrypted.Two convolutional neural networks were conducted to evaluate our proposed approach:The first one is the auto-encoder,which is utilized to compress and encrypt the images,and the other assesses the classification accuracy of the image after decryption and decoding.Different hyperparameters of the encoder were tested,followed by the classification of the image to verify that no critical information was lost,to test the encryption and encoding resolution.In this approach,sixteen hyperparameter permutations are utilized,but this research discusses three main cases in detail.The first case shows that the combination of Mean Square Logarithmic Error(MSLE),ADAgrad,two layers for the auto-encoder,and ReLU had the best auto-encoder results with a Mean Absolute Error(MAE)=0.221 after 50 epochs and 75%classification with the best result for the classification algorithm.The second case shows the reflection of auto-encoder results on the classification results which is a combination ofMean Square Error(MSE),RMSprop,three layers for the auto-encoder,and ReLU,which had the best classification accuracy of 65%,the auto-encoder gives MAE=0.31 after 50 epochs.The third case is the worst,which is the combination of the hinge,RMSprop,three layers for the auto-encoder,and ReLU,providing accuracy of 20%and MAE=0.485.展开更多
Dear Editor,This letter presents a novel segmentation approach that leverages dendritic neurons to tackle the challenges of medical imaging segmentation.In this study,we enhance the segmentation accuracy based on a Se...Dear Editor,This letter presents a novel segmentation approach that leverages dendritic neurons to tackle the challenges of medical imaging segmentation.In this study,we enhance the segmentation accuracy based on a SegNet variant including an encoder-decoder structure,an upsampling index,and a deep supervision method.Furthermore,we introduce a dendritic neuron-based convolutional block to enable nonlinear feature mapping,thereby further improving the effectiveness of our approach.展开更多
Limited by the dynamic range of the detector,saturation artifacts usually occur in optical coherence tomography(OCT)imaging for high scattering media.The available methods are difficult to remove saturation artifacts ...Limited by the dynamic range of the detector,saturation artifacts usually occur in optical coherence tomography(OCT)imaging for high scattering media.The available methods are difficult to remove saturation artifacts and restore texture completely in OCT images.We proposed a deep learning-based inpainting method of saturation artifacts in this paper.The generation mechanism of saturation artifacts was analyzed,and experimental and simulated datasets were built based on the mechanism.Enhanced super-resolution generative adversarial networks were trained by the clear–saturated phantom image pairs.The perfect reconstructed results of experimental zebrafish and thyroid OCT images proved its feasibility,strong generalization,and robustness.展开更多
The scarcity of in-situ ocean observations poses a challenge for real-time information acquisition in the ocean.Among the crucial hydroacoustic environmental parameters,ocean sound velocity exhibits significant spatia...The scarcity of in-situ ocean observations poses a challenge for real-time information acquisition in the ocean.Among the crucial hydroacoustic environmental parameters,ocean sound velocity exhibits significant spatial and temporal variability and it is highly relevant to oceanic research.In this study,we propose a new data-driven approach,leveraging deep learning techniques,for the prediction of sound velocity fields(SVFs).Our novel spatiotemporal prediction model,STLSTM-SA,combines Spatiotemporal Long Short-Term Memory(ST-LSTM) with a self-attention mechanism to enable accurate and real-time prediction of SVFs.To circumvent the limited amount of observational data,we employ transfer learning by first training the model using reanalysis datasets,followed by fine-tuning it using in-situ analysis data to obtain the final prediction model.By utilizing the historical 12-month SVFs as input,our model predicts the SVFs for the subsequent three months.We compare the performance of five models:Artificial Neural Networks(ANN),Long ShortTerm Memory(LSTM),Convolutional LSTM(ConvLSTM),ST-LSTM,and our proposed ST-LSTM-SA model in a test experiment spanning 2019 to 2022.Our results demonstrate that the ST-LSTM-SA model significantly improves the prediction accuracy and stability of sound velocity in both temporal and spatial dimensions.The ST-LSTM-SA model not only accurately predicts the ocean sound velocity field(SVF),but also provides valuable insights for spatiotemporal prediction of other oceanic environmental variables.展开更多
Avatars, as promising digital representations and service assistants of users in Metaverses, can enable drivers and passengers to immerse themselves in 3D virtual services and spaces of UAV-assisted vehicular Metavers...Avatars, as promising digital representations and service assistants of users in Metaverses, can enable drivers and passengers to immerse themselves in 3D virtual services and spaces of UAV-assisted vehicular Metaverses. However, avatar tasks include a multitude of human-to-avatar and avatar-to-avatar interactive applications, e.g., augmented reality navigation,which consumes intensive computing resources. It is inefficient and impractical for vehicles to process avatar tasks locally. Fortunately, migrating avatar tasks to the nearest roadside units(RSU)or unmanned aerial vehicles(UAV) for execution is a promising solution to decrease computation overhead and reduce task processing latency, while the high mobility of vehicles brings challenges for vehicles to independently perform avatar migration decisions depending on current and future vehicle status. To address these challenges, in this paper, we propose a novel avatar task migration system based on multi-agent deep reinforcement learning(MADRL) to execute immersive vehicular avatar tasks dynamically. Specifically, we first formulate the problem of avatar task migration from vehicles to RSUs/UAVs as a partially observable Markov decision process that can be solved by MADRL algorithms. We then design the multi-agent proximal policy optimization(MAPPO) approach as the MADRL algorithm for the avatar task migration problem. To overcome slow convergence resulting from the curse of dimensionality and non-stationary issues caused by shared parameters in MAPPO, we further propose a transformer-based MAPPO approach via sequential decision-making models for the efficient representation of relationships among agents. Finally, to motivate terrestrial or non-terrestrial edge servers(e.g., RSUs or UAVs) to share computation resources and ensure traceability of the sharing records, we apply smart contracts and blockchain technologies to achieve secure sharing management. Numerical results demonstrate that the proposed approach outperforms the MAPPO approach by around 2% and effectively reduces approximately 20% of the latency of avatar task execution in UAV-assisted vehicular Metaverses.展开更多
Deep and ultra-deep reservoirs have gradually become the primary focus of hydrocarbon exploration as a result of a series of significant discoveries in deep hydrocarbon exploration worldwide.These reservoirs present u...Deep and ultra-deep reservoirs have gradually become the primary focus of hydrocarbon exploration as a result of a series of significant discoveries in deep hydrocarbon exploration worldwide.These reservoirs present unique challenges due to their deep burial depth(4500-8882 m),low matrix permeability,complex crustal stress conditions,high temperature and pressure(HTHP,150-200℃,105-155 MPa),coupled with high salinity of formation water.Consequently,the costs associated with their exploitation and development are exceptionally high.In deep and ultra-deep reservoirs,hydraulic fracturing is commonly used to achieve high and stable production.During hydraulic fracturing,a substantial volume of fluid is injected into the reservoir.However,statistical analysis reveals that the flowback rate is typically less than 30%,leaving the majority of the fluid trapped within the reservoir.Therefore,hydraulic fracturing in deep reservoirs not only enhances the reservoir permeability by creating artificial fractures but also damages reservoirs due to the fracturing fluids involved.The challenging“three-high”environment of a deep reservoir,characterized by high temperature,high pressure,and high salinity,exacerbates conventional forms of damage,including water sensitivity,retention of fracturing fluids,rock creep,and proppant breakage.In addition,specific damage mechanisms come into play,such as fracturing fluid decomposition at elevated temperatures and proppant diagenetic reactions at HTHP conditions.Presently,the foremost concern in deep oil and gas development lies in effectively assessing the damage inflicted on these reservoirs by hydraulic fracturing,comprehending the underlying mechanisms,and selecting appropriate solutions.It's noteworthy that the majority of existing studies on reservoir damage primarily focus on conventional reservoirs,with limited attention given to deep reservoirs and a lack of systematic summaries.In light of this,our approach entails initially summarizing the current knowledge pertaining to the types of fracturing fluids employed in deep and ultra-deep reservoirs.Subsequently,we delve into a systematic examination of the damage processes and mechanisms caused by fracturing fluids within the context of hydraulic fracturing in deep reservoirs,taking into account the unique reservoir characteristics of high temperature,high pressure,and high in-situ stress.In addition,we provide an overview of research progress related to high-temperature deep reservoir fracturing fluid and the damage of aqueous fracturing fluids to rock matrix,both artificial and natural fractures,and sand-packed fractures.We conclude by offering a summary of current research advancements and future directions,which hold significant potential for facilitating the efficient development of deep oil and gas reservoirs while effectively mitigating reservoir damage.展开更多
The great potentials of massive Multiple-Input Multiple-Output(MIMO)in Frequency Division Duplex(FDD)mode can be fully exploited when the downlink Channel State Information(CSI)is available at base stations.However,th...The great potentials of massive Multiple-Input Multiple-Output(MIMO)in Frequency Division Duplex(FDD)mode can be fully exploited when the downlink Channel State Information(CSI)is available at base stations.However,the accurate CsI is difficult to obtain due to the large amount of feedback overhead caused by massive antennas.In this paper,we propose a deep learning based joint channel estimation and feedback framework,which comprehensively realizes the estimation,compression,and reconstruction of downlink channels in FDD massive MIMO systems.Two networks are constructed to perform estimation and feedback explicitly and implicitly.The explicit network adopts a multi-Signal-to-Noise-Ratios(SNRs)technique to obtain a single trained channel estimation subnet that works well with different SNRs and employs a deep residual network to reconstruct the channels,while the implicit network directly compresses pilots and sends them back to reduce network parameters.Quantization module is also designed to generate data-bearing bitstreams.Simulation results show that the two proposed networks exhibit excellent performance of reconstruction and are robust to different environments and quantization errors.展开更多
Accurate soil moisture(SM)prediction is critical for understanding hydrological processes.Physics-based(PB)models exhibit large uncertainties in SM predictions arising from uncertain parameterizations and insufficient...Accurate soil moisture(SM)prediction is critical for understanding hydrological processes.Physics-based(PB)models exhibit large uncertainties in SM predictions arising from uncertain parameterizations and insufficient representation of land-surface processes.In addition to PB models,deep learning(DL)models have been widely used in SM predictions recently.However,few pure DL models have notably high success rates due to lacking physical information.Thus,we developed hybrid models to effectively integrate the outputs of PB models into DL models to improve SM predictions.To this end,we first developed a hybrid model based on the attention mechanism to take advantage of PB models at each forecast time scale(attention model).We further built an ensemble model that combined the advantages of different hybrid schemes(ensemble model).We utilized SM forecasts from the Global Forecast System to enhance the convolutional long short-term memory(ConvLSTM)model for 1–16 days of SM predictions.The performances of the proposed hybrid models were investigated and compared with two existing hybrid models.The results showed that the attention model could leverage benefits of PB models and achieved the best predictability of drought events among the different hybrid models.Moreover,the ensemble model performed best among all hybrid models at all forecast time scales and different soil conditions.It is highlighted that the ensemble model outperformed the pure DL model over 79.5%of in situ stations for 16-day predictions.These findings suggest that our proposed hybrid models can adequately exploit the benefits of PB model outputs to aid DL models in making SM predictions.展开更多
Thunderstorm gusts are a common form of severe convective weather in the warm season in North China,and it is of great importance to correctly forecast them.At present,the forecasting of thunderstorm gusts is mainly b...Thunderstorm gusts are a common form of severe convective weather in the warm season in North China,and it is of great importance to correctly forecast them.At present,the forecasting of thunderstorm gusts is mainly based on traditional subjective methods,which fails to achieve high-resolution and high-frequency gridded forecasts based on multiple observation sources.In this paper,we propose a deep learning method called Thunderstorm Gusts TransU-net(TGTransUnet)to forecast thunderstorm gusts in North China based on multi-source gridded product data from the Institute of Urban Meteorology(IUM)with a lead time of 1 to 6 h.To determine the specific range of thunderstorm gusts,we combine three meteorological variables:radar reflectivity factor,lightning location,and 1-h maximum instantaneous wind speed from automatic weather stations(AWSs),and obtain a reasonable ground truth of thunderstorm gusts.Then,we transform the forecasting problem into an image-to-image problem in deep learning under the TG-TransUnet architecture,which is based on convolutional neural networks and a transformer.The analysis and forecast data of the enriched multi-source gridded comprehensive forecasting system for the period 2021–23 are then used as training,validation,and testing datasets.Finally,the performance of TG-TransUnet is compared with other methods.The results show that TG-TransUnet has the best prediction results at 1–6 h.The IUM is currently using this model to support the forecasting of thunderstorm gusts in North China.展开更多
文摘Since the effectiveness of extracting fault features is not high under traditional bearing fault diagnosis method, a bearing fault diagnosis method based on Deep Auto-encoder Network (DAEN) optimized by Cloud Adaptive Particle Swarm Optimization (CAPSO) was proposed. On the basis of analyzing CAPSO and DAEN, the CAPSO-DAEN fault diagnosis model is built. The model uses the randomness and stability of CAPSO algorithm to optimize the connection weight of DAEN, to reduce the constraints on the weights and extract fault features adaptively. Finally, efficient and accurate fault diagnosis can be implemented with the Softmax classifier. The results of test show that the proposed method has higher diagnostic accuracy and more stable diagnosis results than those based on the DAEN, Support Vector Machine (SVM) and the Back Propagation algorithm (BP) under appropriate parameters.
文摘To address the difficulties in fusing multi-mode sensor data for complex industrial machinery, an adaptive deep coupling convolutional auto-encoder (ADCCAE) fusion method was proposed. First, the multi-mode features extracted synchronously by the CCAE were stacked and fed to the multi-channel convolution layers for fusion. Then, the fused data was passed to all connection layers for compression and fed to the Softmax module for classification. Finally, the coupling loss function coefficients and the network parameters were optimized through an adaptive approach using the gray wolf optimization (GWO) algorithm. Experimental comparisons showed that the proposed ADCCAE fusion model was superior to existing models for multi-mode data fusion.
基金deputyship for Research&Innovation,Ministry of Education in Saudi Arabia,for funding this research work through Project Number (IFP-2020-133).
文摘Intelligent diagnosis approaches with shallow architectural models play an essential role in healthcare.Deep Learning(DL)models with unsupervised learning concepts have been proposed because high-quality feature extraction and adequate labelled details significantly influence shallow models.On the other hand,skin lesionbased segregation and disintegration procedures play an essential role in earlier skin cancer detection.However,artefacts,an unclear boundary,poor contrast,and different lesion sizes make detection difficult.To address the issues in skin lesion diagnosis,this study creates the UDLS-DDOA model,an intelligent Unsupervised Deep Learning-based Stacked Auto-encoder(UDLS)optimized by Dynamic Differential Annealed Optimization(DDOA).Pre-processing,segregation,feature removal or separation,and disintegration are part of the proposed skin lesion diagnosis model.Pre-processing of skin lesion images occurs at the initial level for noise removal in the image using the Top hat filter and painting methodology.Following that,a Fuzzy C-Means(FCM)segregation procedure is performed using a Quasi-Oppositional Elephant Herd Optimization(QOEHO)algorithm.Besides,a novel feature extraction technique using the UDLS technique is applied where the parameter tuning takes place using DDOA.In the end,the disintegration procedure would be accomplished using a SoftMax(SM)classifier.The UDLS-DDOA model is tested against the International Skin Imaging Collaboration(ISIC)dataset,and the experimental results are examined using various computational attributes.The simulation results demonstrated that the UDLS-DDOA model outperformed the compared methods significantly.
文摘With the rapid development of deep learning methods, the data-driven approach has shown powerful advantages over the model-driven one. In this paper, we propose an end-to-end autoencoder communication system based on Deep Residual Shrinkage Networks (DRSNs), where neural networks (DNNs) are used to implement the coding, decoding, modulation and demodulation functions of the communication system. Our proposed autoencoder communication system can better reduce the signal noise by adding an “attention mechanism” and “soft thresholding” modules and has better performance at various signal-to-noise ratios (SNR). Also, we have shown through comparative experiments that the system can operate at moderate block lengths and support different throughputs. It has been shown to work efficiently in the AWGN channel. Simulation results show that our model has a higher Bit-Error-Rate (BER) gain and greatly improved decoding performance compared to conventional modulation and classical autoencoder systems at various signal-to-noise ratios.
基金funding was provided by the Institute for Research and Consulting Studies at King Khalid University through Corona Research(Fast Track)[Grant No.3-103S-2020].
文摘Healthcare systems nowadays depend on IoT sensors for sending data over the internet as a common practice.Encryption ofmedical images is very important to secure patient information.Encrypting these images consumes a lot of time onedge computing;therefore,theuse of anauto-encoder for compressionbefore encodingwill solve such a problem.In this paper,we use an auto-encoder to compress amedical image before encryption,and an encryption output(vector)is sent out over the network.On the other hand,a decoder was used to reproduce the original image back after the vector was received and decrypted.Two convolutional neural networks were conducted to evaluate our proposed approach:The first one is the auto-encoder,which is utilized to compress and encrypt the images,and the other assesses the classification accuracy of the image after decryption and decoding.Different hyperparameters of the encoder were tested,followed by the classification of the image to verify that no critical information was lost,to test the encryption and encoding resolution.In this approach,sixteen hyperparameter permutations are utilized,but this research discusses three main cases in detail.The first case shows that the combination of Mean Square Logarithmic Error(MSLE),ADAgrad,two layers for the auto-encoder,and ReLU had the best auto-encoder results with a Mean Absolute Error(MAE)=0.221 after 50 epochs and 75%classification with the best result for the classification algorithm.The second case shows the reflection of auto-encoder results on the classification results which is a combination ofMean Square Error(MSE),RMSprop,three layers for the auto-encoder,and ReLU,which had the best classification accuracy of 65%,the auto-encoder gives MAE=0.31 after 50 epochs.The third case is the worst,which is the combination of the hinge,RMSprop,three layers for the auto-encoder,and ReLU,providing accuracy of 20%and MAE=0.485.
基金partially supported by the Japan Society for the Promotion of Science(JSPS)KAKENHI(JP22H03643)Japan Science and Technology Agency(JST)Support for Pioneering Research Initiated by the Next Generation(SPRING)(JPMJSP2145)JST through the Establishment of University Fellowships Towards the Creation of Science Technology Innovation(JPMJFS2115)。
文摘Dear Editor,This letter presents a novel segmentation approach that leverages dendritic neurons to tackle the challenges of medical imaging segmentation.In this study,we enhance the segmentation accuracy based on a SegNet variant including an encoder-decoder structure,an upsampling index,and a deep supervision method.Furthermore,we introduce a dendritic neuron-based convolutional block to enable nonlinear feature mapping,thereby further improving the effectiveness of our approach.
基金supported by the National Natural Science Foundation of China(62375144 and 61875092)Tianjin Foundation of Natural Science(21JCYBJC00260)Beijing-Tianjin-Hebei Basic Research Cooperation Special Program(19JCZDJC65300).
文摘Limited by the dynamic range of the detector,saturation artifacts usually occur in optical coherence tomography(OCT)imaging for high scattering media.The available methods are difficult to remove saturation artifacts and restore texture completely in OCT images.We proposed a deep learning-based inpainting method of saturation artifacts in this paper.The generation mechanism of saturation artifacts was analyzed,and experimental and simulated datasets were built based on the mechanism.Enhanced super-resolution generative adversarial networks were trained by the clear–saturated phantom image pairs.The perfect reconstructed results of experimental zebrafish and thyroid OCT images proved its feasibility,strong generalization,and robustness.
基金supported by the National Natural Science Foundation of China(Grant No.42004030)Basic Scientific Fund for National Public Research Institutes of China(Grant No.2022S03)+1 种基金Science and Technology Innovation Project(LSKJ202205102)funded by Laoshan Laboratory,and the National Key Research and Development Program of China(2020YFB0505805).
文摘The scarcity of in-situ ocean observations poses a challenge for real-time information acquisition in the ocean.Among the crucial hydroacoustic environmental parameters,ocean sound velocity exhibits significant spatial and temporal variability and it is highly relevant to oceanic research.In this study,we propose a new data-driven approach,leveraging deep learning techniques,for the prediction of sound velocity fields(SVFs).Our novel spatiotemporal prediction model,STLSTM-SA,combines Spatiotemporal Long Short-Term Memory(ST-LSTM) with a self-attention mechanism to enable accurate and real-time prediction of SVFs.To circumvent the limited amount of observational data,we employ transfer learning by first training the model using reanalysis datasets,followed by fine-tuning it using in-situ analysis data to obtain the final prediction model.By utilizing the historical 12-month SVFs as input,our model predicts the SVFs for the subsequent three months.We compare the performance of five models:Artificial Neural Networks(ANN),Long ShortTerm Memory(LSTM),Convolutional LSTM(ConvLSTM),ST-LSTM,and our proposed ST-LSTM-SA model in a test experiment spanning 2019 to 2022.Our results demonstrate that the ST-LSTM-SA model significantly improves the prediction accuracy and stability of sound velocity in both temporal and spatial dimensions.The ST-LSTM-SA model not only accurately predicts the ocean sound velocity field(SVF),but also provides valuable insights for spatiotemporal prediction of other oceanic environmental variables.
基金supported in part by NSFC (62102099, U22A2054, 62101594)in part by the Pearl River Talent Recruitment Program (2021QN02S643)+9 种基金Guangzhou Basic Research Program (2023A04J1699)in part by the National Research Foundation, SingaporeInfocomm Media Development Authority under its Future Communications Research Development ProgrammeDSO National Laboratories under the AI Singapore Programme under AISG Award No AISG2-RP-2020-019Energy Research Test-Bed and Industry Partnership Funding Initiative, Energy Grid (EG) 2.0 programmeDesCartes and the Campus for Research Excellence and Technological Enterprise (CREATE) programmeMOE Tier 1 under Grant RG87/22in part by the Singapore University of Technology and Design (SUTD) (SRG-ISTD-2021- 165)in part by the SUTD-ZJU IDEA Grant SUTD-ZJU (VP) 202102in part by the Ministry of Education, Singapore, through its SUTD Kickstarter Initiative (SKI 20210204)。
文摘Avatars, as promising digital representations and service assistants of users in Metaverses, can enable drivers and passengers to immerse themselves in 3D virtual services and spaces of UAV-assisted vehicular Metaverses. However, avatar tasks include a multitude of human-to-avatar and avatar-to-avatar interactive applications, e.g., augmented reality navigation,which consumes intensive computing resources. It is inefficient and impractical for vehicles to process avatar tasks locally. Fortunately, migrating avatar tasks to the nearest roadside units(RSU)or unmanned aerial vehicles(UAV) for execution is a promising solution to decrease computation overhead and reduce task processing latency, while the high mobility of vehicles brings challenges for vehicles to independently perform avatar migration decisions depending on current and future vehicle status. To address these challenges, in this paper, we propose a novel avatar task migration system based on multi-agent deep reinforcement learning(MADRL) to execute immersive vehicular avatar tasks dynamically. Specifically, we first formulate the problem of avatar task migration from vehicles to RSUs/UAVs as a partially observable Markov decision process that can be solved by MADRL algorithms. We then design the multi-agent proximal policy optimization(MAPPO) approach as the MADRL algorithm for the avatar task migration problem. To overcome slow convergence resulting from the curse of dimensionality and non-stationary issues caused by shared parameters in MAPPO, we further propose a transformer-based MAPPO approach via sequential decision-making models for the efficient representation of relationships among agents. Finally, to motivate terrestrial or non-terrestrial edge servers(e.g., RSUs or UAVs) to share computation resources and ensure traceability of the sharing records, we apply smart contracts and blockchain technologies to achieve secure sharing management. Numerical results demonstrate that the proposed approach outperforms the MAPPO approach by around 2% and effectively reduces approximately 20% of the latency of avatar task execution in UAV-assisted vehicular Metaverses.
基金Dao-Bing Wang was supported by the Beijing Natural Science Foundation Project(No.3222030)the National Natural Science Foundation of China(No.52274002)+1 种基金the PetroChina Science and Technology Innovation Foundation Project(No.2021DQ02-0201)Fu-Jian Zhou was supported by the National Natural Science Foundation of China(No.52174045).
文摘Deep and ultra-deep reservoirs have gradually become the primary focus of hydrocarbon exploration as a result of a series of significant discoveries in deep hydrocarbon exploration worldwide.These reservoirs present unique challenges due to their deep burial depth(4500-8882 m),low matrix permeability,complex crustal stress conditions,high temperature and pressure(HTHP,150-200℃,105-155 MPa),coupled with high salinity of formation water.Consequently,the costs associated with their exploitation and development are exceptionally high.In deep and ultra-deep reservoirs,hydraulic fracturing is commonly used to achieve high and stable production.During hydraulic fracturing,a substantial volume of fluid is injected into the reservoir.However,statistical analysis reveals that the flowback rate is typically less than 30%,leaving the majority of the fluid trapped within the reservoir.Therefore,hydraulic fracturing in deep reservoirs not only enhances the reservoir permeability by creating artificial fractures but also damages reservoirs due to the fracturing fluids involved.The challenging“three-high”environment of a deep reservoir,characterized by high temperature,high pressure,and high salinity,exacerbates conventional forms of damage,including water sensitivity,retention of fracturing fluids,rock creep,and proppant breakage.In addition,specific damage mechanisms come into play,such as fracturing fluid decomposition at elevated temperatures and proppant diagenetic reactions at HTHP conditions.Presently,the foremost concern in deep oil and gas development lies in effectively assessing the damage inflicted on these reservoirs by hydraulic fracturing,comprehending the underlying mechanisms,and selecting appropriate solutions.It's noteworthy that the majority of existing studies on reservoir damage primarily focus on conventional reservoirs,with limited attention given to deep reservoirs and a lack of systematic summaries.In light of this,our approach entails initially summarizing the current knowledge pertaining to the types of fracturing fluids employed in deep and ultra-deep reservoirs.Subsequently,we delve into a systematic examination of the damage processes and mechanisms caused by fracturing fluids within the context of hydraulic fracturing in deep reservoirs,taking into account the unique reservoir characteristics of high temperature,high pressure,and high in-situ stress.In addition,we provide an overview of research progress related to high-temperature deep reservoir fracturing fluid and the damage of aqueous fracturing fluids to rock matrix,both artificial and natural fractures,and sand-packed fractures.We conclude by offering a summary of current research advancements and future directions,which hold significant potential for facilitating the efficient development of deep oil and gas reservoirs while effectively mitigating reservoir damage.
基金supported in part by the National Natural Science Foundation of China(NSFC)under Grants 61941104,61921004the Key Research and Development Program of Shandong Province under Grant 2020CXGC010108+1 种基金the Southeast University-China Mobile Research Institute Joint Innovation Centersupported in part by the Scientific Research Foundation of Graduate School of Southeast University under Grant YBPY2118.
文摘The great potentials of massive Multiple-Input Multiple-Output(MIMO)in Frequency Division Duplex(FDD)mode can be fully exploited when the downlink Channel State Information(CSI)is available at base stations.However,the accurate CsI is difficult to obtain due to the large amount of feedback overhead caused by massive antennas.In this paper,we propose a deep learning based joint channel estimation and feedback framework,which comprehensively realizes the estimation,compression,and reconstruction of downlink channels in FDD massive MIMO systems.Two networks are constructed to perform estimation and feedback explicitly and implicitly.The explicit network adopts a multi-Signal-to-Noise-Ratios(SNRs)technique to obtain a single trained channel estimation subnet that works well with different SNRs and employs a deep residual network to reconstruct the channels,while the implicit network directly compresses pilots and sends them back to reduce network parameters.Quantization module is also designed to generate data-bearing bitstreams.Simulation results show that the two proposed networks exhibit excellent performance of reconstruction and are robust to different environments and quantization errors.
基金supported by the Natural Science Foundation of China(Grant Nos.42088101 and 42205149)Zhongwang WEI was supported by the Natural Science Foundation of China(Grant No.42075158)+1 种基金Wei SHANGGUAN was supported by the Natural Science Foundation of China(Grant No.41975122)Yonggen ZHANG was supported by the National Natural Science Foundation of Tianjin(Grant No.20JCQNJC01660).
文摘Accurate soil moisture(SM)prediction is critical for understanding hydrological processes.Physics-based(PB)models exhibit large uncertainties in SM predictions arising from uncertain parameterizations and insufficient representation of land-surface processes.In addition to PB models,deep learning(DL)models have been widely used in SM predictions recently.However,few pure DL models have notably high success rates due to lacking physical information.Thus,we developed hybrid models to effectively integrate the outputs of PB models into DL models to improve SM predictions.To this end,we first developed a hybrid model based on the attention mechanism to take advantage of PB models at each forecast time scale(attention model).We further built an ensemble model that combined the advantages of different hybrid schemes(ensemble model).We utilized SM forecasts from the Global Forecast System to enhance the convolutional long short-term memory(ConvLSTM)model for 1–16 days of SM predictions.The performances of the proposed hybrid models were investigated and compared with two existing hybrid models.The results showed that the attention model could leverage benefits of PB models and achieved the best predictability of drought events among the different hybrid models.Moreover,the ensemble model performed best among all hybrid models at all forecast time scales and different soil conditions.It is highlighted that the ensemble model outperformed the pure DL model over 79.5%of in situ stations for 16-day predictions.These findings suggest that our proposed hybrid models can adequately exploit the benefits of PB model outputs to aid DL models in making SM predictions.
基金supported in part by the Beijing Natural Science Foundation(Grant No.8222051)the National Key R&D Program of China(Grant No.2022YFC3004103)+2 种基金the National Natural Foundation of China(Grant Nos.42275003 and 42275012)the China Meteorological Administration Key Innovation Team(Grant Nos.CMA2022ZD04 and CMA2022ZD07)the Beijing Science and Technology Program(Grant No.Z221100005222012).
文摘Thunderstorm gusts are a common form of severe convective weather in the warm season in North China,and it is of great importance to correctly forecast them.At present,the forecasting of thunderstorm gusts is mainly based on traditional subjective methods,which fails to achieve high-resolution and high-frequency gridded forecasts based on multiple observation sources.In this paper,we propose a deep learning method called Thunderstorm Gusts TransU-net(TGTransUnet)to forecast thunderstorm gusts in North China based on multi-source gridded product data from the Institute of Urban Meteorology(IUM)with a lead time of 1 to 6 h.To determine the specific range of thunderstorm gusts,we combine three meteorological variables:radar reflectivity factor,lightning location,and 1-h maximum instantaneous wind speed from automatic weather stations(AWSs),and obtain a reasonable ground truth of thunderstorm gusts.Then,we transform the forecasting problem into an image-to-image problem in deep learning under the TG-TransUnet architecture,which is based on convolutional neural networks and a transformer.The analysis and forecast data of the enriched multi-source gridded comprehensive forecasting system for the period 2021–23 are then used as training,validation,and testing datasets.Finally,the performance of TG-TransUnet is compared with other methods.The results show that TG-TransUnet has the best prediction results at 1–6 h.The IUM is currently using this model to support the forecasting of thunderstorm gusts in North China.