期刊文献+

U-shaped Vision Transformer and Its Application in Gear Pitting Measurement

下载PDF
导出
摘要 Although convolutional neural networks have become the mainstream segmentation model,the locality of convolution makes them cannot well learn global and long-range semantic information.To further improve the performance of segmentation models,we propose U-shaped vision Transformer(UsViT),a model based on Transformer and convolution.Specifically,residual Transformer blocks are designed in the encoder of UsViT,which take advantages of residual network and Transformer backbone at the same time.What is more,transpositions in each Transformer layer achieve the information interaction between spatial locations and feature channels,enhancing the capability of feature learning.In the decoder,for enhancing receptive field,different dilation rates are introduced to each convolutional layer.In addition,residual connections are applied to make the information propagation smoother when training the model.We first verify the superiority of UsViT on automatic portrait matting public dataset,which achieves 90.43%accuracy(Acc),95.56%Dice similarity coefficient,and 94.66%Intersection over Union with relatively fewer parameters.Finally,UsViT is applied to gear pitting measurement in gear contact fatigue test,and the comparative results indicate that UsViT can improve the Acc of pitting detection.
出处 《Journal of Dynamics, Monitoring and Diagnostics》 2022年第4期223-228,共6页 动力学、监测与诊断学报(英文)
基金 supported in part by National Natural Science Foundation of China under Grants 62033001 and 52175075.
  • 相关文献

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部