Recently, semantic segmentation has been widely applied toimage processing, scene understanding, and many others. Especially, indeep learning-based semantic segmentation, the U-Net with convolutionalencoder-decoder ar...Recently, semantic segmentation has been widely applied toimage processing, scene understanding, and many others. Especially, indeep learning-based semantic segmentation, the U-Net with convolutionalencoder-decoder architecture is a representative model which is proposed forimage segmentation in the biomedical field. It used max pooling operationfor reducing the size of image and making noise robust. However, instead ofreducing the complexity of the model, max pooling has the disadvantageof omitting some information about the image in reducing it. So, thispaper used two diagonal elements of down-sampling operation instead ofit. We think that the down-sampling feature maps have more informationintrinsically than max pooling feature maps because of keeping the Nyquisttheorem and extracting the latent information from them. In addition,this paper used two other diagonal elements for the skip connection. Indecoding, we used Subpixel Convolution rather than transposed convolutionto efficiently decode the encoded feature maps. Including all the ideas, thispaper proposed the new encoder-decoder model called Down-Sampling andSubpixel Convolution U-Net (DSSC-UNet). To prove the better performanceof the proposed model, this paper measured the performance of the UNetand DSSC-UNet on the Cityscapes. As a result, DSSC-UNet achieved89.6% Mean Intersection OverUnion (Mean-IoU) andU-Net achieved 85.6%Mean-IoU, confirming that DSSC-UNet achieved better performance.展开更多
Laminated composites are widely used in many engineering industries such as aircraft, spacecraft, boat hulls, racing car bodies, and storage tanks. We analyze the 3D deformations of a multilayered, linear elastic, ani...Laminated composites are widely used in many engineering industries such as aircraft, spacecraft, boat hulls, racing car bodies, and storage tanks. We analyze the 3D deformations of a multilayered, linear elastic, anisotropic rectangular plate subjected to arbitrary boundary conditions on one edge and simply supported on other edge. The rectangular laminate consists of anisotropic and homogeneous laminae of arbitrary thicknesses. This study presents the elastic analysis of laminated composite plates subjected to sinusoidal mechanical loading under arbitrary boundary conditions. Least square finite element solutions for displacements and stresses are investigated using a mathematical model, called a state-space model, which allows us to simultaneously solve for these field variables in the composite structure’s domain and ensure that continuity conditions are satisfied at layer interfaces. The governing equations are derived from this model using a numerical technique called the least-squares finite element method (LSFEM). These LSFEMs seek to minimize the squares of the governing equations and the associated side conditions residuals over the computational domain. The model is comprised of layerwise variables such as displacements, out-of-plane stresses, and in- plane strains, treated as independent variables. Numerical results are presented to demonstrate the response of the laminated composite plates under various arbitrary boundary conditions using LSFEM and compared with the 3D elasticity solution available in the literature.展开更多
Imagine numerous clients,each with personal data;individual inputs are severely corrupt,and a server only concerns the collective,statistically essential facets of this data.In several data mining methods,privacy has ...Imagine numerous clients,each with personal data;individual inputs are severely corrupt,and a server only concerns the collective,statistically essential facets of this data.In several data mining methods,privacy has become highly critical.As a result,various privacy-preserving data analysis technologies have emerged.Hence,we use the randomization process to reconstruct composite data attributes accurately.Also,we use privacy measures to estimate how much deception is required to guarantee privacy.There are several viable privacy protections;however,determining which one is the best is still a work in progress.This paper discusses the difficulty of measuring privacy while also offering numerous random sampling procedures and statistical and categorized data results.Further-more,this paper investigates the use of arbitrary nature with perturbations in privacy preservation.According to the research,arbitrary objects(most notably random matrices)have"predicted"frequency patterns.It shows how to recover crucial information from a sample damaged by a random number using an arbi-trary lattice spectral selection strategy.Thisfiltration system's conceptual frame-work posits,and extensive practicalfindings indicate that sparse data distortions preserve relatively modest privacy protection in various situations.As a result,the research framework is efficient and effective in maintaining data privacy and security.展开更多
In this paper,we propose an arbitrary decode-forward single-relay scheme for finite blocklength polar codes,which can be applied to the general symmetric discrete memoryless relay channel with orthogonal receiver comp...In this paper,we propose an arbitrary decode-forward single-relay scheme for finite blocklength polar codes,which can be applied to the general symmetric discrete memoryless relay channel with orthogonal receiver components.The relay node decodes the received message.The relay node selectively re-encodes the message and transmits it to the destination node.Furthermore,in order to minimize the upper-bound of the block error probability,we propose a selection strategy to decide the proper re-encoded bit set by the relay.Simulation results are presented to illustrate the improvement in decoding performance of the proposed scheme compared to conventional relay schemes in both additive white Gaussian noise(AWGN)channel and Rayleigh fading channel(RFC).展开更多
We are investigating the distributed optimization problem,where a network of nodes works together to minimize a global objective that is a finite sum of their stored local functions.Since nodes exchange optimization p...We are investigating the distributed optimization problem,where a network of nodes works together to minimize a global objective that is a finite sum of their stored local functions.Since nodes exchange optimization parameters through the wireless network,large-scale training models can create communication bottlenecks,resulting in slower training times.To address this issue,CHOCO-SGD was proposed,which allows compressing information with arbitrary precision without reducing the convergence rate for strongly convex objective functions.Nevertheless,most convex functions are not strongly convex(such as logistic regression or Lasso),which raises the question of whether this algorithm can be applied to non-strongly convex functions.In this paper,we provide the first theoretical analysis of the convergence rate of CHOCO-SGD on non-strongly convex objectives.We derive a sufficient condition,which limits the fidelity of compression,to guarantee convergence.Moreover,our analysis demonstrates that within the fidelity threshold,this algorithm can significantly reduce transmission burden while maintaining the same convergence rate order as its no-compression equivalent.Numerical experiments further validate the theoretical findings by demonstrating that CHOCO-SGD improves communication efficiency and keeps the same convergence rate order simultaneously.And experiments also show that the algorithm fails to converge with low compression fidelity and in time-varying topologies.Overall,our study offers valuable insights into the potential applicability of CHOCO-SGD for non-strongly convex objectives.Additionally,we provide practical guidelines for researchers seeking to utilize this algorithm in real-world scenarios.展开更多
文摘Recently, semantic segmentation has been widely applied toimage processing, scene understanding, and many others. Especially, indeep learning-based semantic segmentation, the U-Net with convolutionalencoder-decoder architecture is a representative model which is proposed forimage segmentation in the biomedical field. It used max pooling operationfor reducing the size of image and making noise robust. However, instead ofreducing the complexity of the model, max pooling has the disadvantageof omitting some information about the image in reducing it. So, thispaper used two diagonal elements of down-sampling operation instead ofit. We think that the down-sampling feature maps have more informationintrinsically than max pooling feature maps because of keeping the Nyquisttheorem and extracting the latent information from them. In addition,this paper used two other diagonal elements for the skip connection. Indecoding, we used Subpixel Convolution rather than transposed convolutionto efficiently decode the encoded feature maps. Including all the ideas, thispaper proposed the new encoder-decoder model called Down-Sampling andSubpixel Convolution U-Net (DSSC-UNet). To prove the better performanceof the proposed model, this paper measured the performance of the UNetand DSSC-UNet on the Cityscapes. As a result, DSSC-UNet achieved89.6% Mean Intersection OverUnion (Mean-IoU) andU-Net achieved 85.6%Mean-IoU, confirming that DSSC-UNet achieved better performance.
文摘Laminated composites are widely used in many engineering industries such as aircraft, spacecraft, boat hulls, racing car bodies, and storage tanks. We analyze the 3D deformations of a multilayered, linear elastic, anisotropic rectangular plate subjected to arbitrary boundary conditions on one edge and simply supported on other edge. The rectangular laminate consists of anisotropic and homogeneous laminae of arbitrary thicknesses. This study presents the elastic analysis of laminated composite plates subjected to sinusoidal mechanical loading under arbitrary boundary conditions. Least square finite element solutions for displacements and stresses are investigated using a mathematical model, called a state-space model, which allows us to simultaneously solve for these field variables in the composite structure’s domain and ensure that continuity conditions are satisfied at layer interfaces. The governing equations are derived from this model using a numerical technique called the least-squares finite element method (LSFEM). These LSFEMs seek to minimize the squares of the governing equations and the associated side conditions residuals over the computational domain. The model is comprised of layerwise variables such as displacements, out-of-plane stresses, and in- plane strains, treated as independent variables. Numerical results are presented to demonstrate the response of the laminated composite plates under various arbitrary boundary conditions using LSFEM and compared with the 3D elasticity solution available in the literature.
文摘Imagine numerous clients,each with personal data;individual inputs are severely corrupt,and a server only concerns the collective,statistically essential facets of this data.In several data mining methods,privacy has become highly critical.As a result,various privacy-preserving data analysis technologies have emerged.Hence,we use the randomization process to reconstruct composite data attributes accurately.Also,we use privacy measures to estimate how much deception is required to guarantee privacy.There are several viable privacy protections;however,determining which one is the best is still a work in progress.This paper discusses the difficulty of measuring privacy while also offering numerous random sampling procedures and statistical and categorized data results.Further-more,this paper investigates the use of arbitrary nature with perturbations in privacy preservation.According to the research,arbitrary objects(most notably random matrices)have"predicted"frequency patterns.It shows how to recover crucial information from a sample damaged by a random number using an arbi-trary lattice spectral selection strategy.Thisfiltration system's conceptual frame-work posits,and extensive practicalfindings indicate that sparse data distortions preserve relatively modest privacy protection in various situations.As a result,the research framework is efficient and effective in maintaining data privacy and security.
基金supported in part by the National Natural Science Foundation of China under Grant 92067202,Grant 62071058.
文摘In this paper,we propose an arbitrary decode-forward single-relay scheme for finite blocklength polar codes,which can be applied to the general symmetric discrete memoryless relay channel with orthogonal receiver components.The relay node decodes the received message.The relay node selectively re-encodes the message and transmits it to the destination node.Furthermore,in order to minimize the upper-bound of the block error probability,we propose a selection strategy to decide the proper re-encoded bit set by the relay.Simulation results are presented to illustrate the improvement in decoding performance of the proposed scheme compared to conventional relay schemes in both additive white Gaussian noise(AWGN)channel and Rayleigh fading channel(RFC).
基金supported in part by the Shanghai Natural Science Foundation under the Grant 22ZR1407000.
文摘We are investigating the distributed optimization problem,where a network of nodes works together to minimize a global objective that is a finite sum of their stored local functions.Since nodes exchange optimization parameters through the wireless network,large-scale training models can create communication bottlenecks,resulting in slower training times.To address this issue,CHOCO-SGD was proposed,which allows compressing information with arbitrary precision without reducing the convergence rate for strongly convex objective functions.Nevertheless,most convex functions are not strongly convex(such as logistic regression or Lasso),which raises the question of whether this algorithm can be applied to non-strongly convex functions.In this paper,we provide the first theoretical analysis of the convergence rate of CHOCO-SGD on non-strongly convex objectives.We derive a sufficient condition,which limits the fidelity of compression,to guarantee convergence.Moreover,our analysis demonstrates that within the fidelity threshold,this algorithm can significantly reduce transmission burden while maintaining the same convergence rate order as its no-compression equivalent.Numerical experiments further validate the theoretical findings by demonstrating that CHOCO-SGD improves communication efficiency and keeps the same convergence rate order simultaneously.And experiments also show that the algorithm fails to converge with low compression fidelity and in time-varying topologies.Overall,our study offers valuable insights into the potential applicability of CHOCO-SGD for non-strongly convex objectives.Additionally,we provide practical guidelines for researchers seeking to utilize this algorithm in real-world scenarios.