期刊文献+

双线性内卷神经网络用于眼底疾病图像分类 被引量:6

Bilinear involution neural network for image classification of fundus diseases
下载PDF
导出
摘要 由于眼底图像具有复杂程度高、个体差异弱、类间距离短等特点,纯卷积神经网络(CNN)和基于注意力的网络并不能在眼底疾病图像分类任务上达到令人满意的精度。因此,采用involution算子实现了注意力双线性内卷神经网络(ABINN)模型用于眼底疾病图像分类。ABINN模型的参数量仅是传统双线性卷积神经网络(BCNN)模型的11%,并提取了眼底图像的底层语义信息和空间结构信息进行二阶特征融合,是CNN和注意力方法的有效并联。此外,提出了两种基于involution算子实现注意力计算的实例化方法:基于图块的注意力子网络(AST)和基于像素的注意力子网络(ASX),这两种方法可以在CNN的基础结构内完成注意力的计算,从而使双线性子网络能在同一个架构下训练并进行特征融合。在公开眼底图像数据集OIA-ODIR上进行实验,结果显示ABINN模型的精度为85%,比通用BCNN模型提高了15.8个百分点,比TransEye模型提高了0.9个百分点。 Due to the high complexity, weak individual differences, and short inter-class distances of fundus image features, pure Convolutional Neural Networks(CNNs) and attention based networks cannot achieve satisfactory accuracy in fundus disease image classification tasks. To this end, Attention Bilinear Involution Neural Network(ABINN) model was implemented for fundus disease image classification by using the involution operator. The parameter amount of ABINN model was only 11% of that of the traditional Bilinear Convolutional Neural Network(BCNN) model. In ABINN model, the underlying semantic information and spatial structure information of the fundus image were extracted and the second-order features of them were fused. It is an effective parallel connection between CNN and attention method. In addition, two instantiation methods for attention calculation based on involution operator, Attention Subnetwork based on PaTch(AST) and Attention Subnetwork based on PiXel(ASX), were proposed. These two methods were able to calculate attention within the CNN basic structure, thereby enabling bilinear sub-networks to be trained and fused in the same architecture.Experimental results on public fundus image dataset OIA-ODIR show that ABINN model has the accuracy of 85%, which is 15. 8 percentage points higher than that of the common BCNN model and 0. 9 percentage points higher than that of TransEye(Transformer Eye) model.
作者 杨洪刚 陈洁洁 徐梦飞 YANG Honggang;CHEN Jiejie;XU Mengfei(College of Computer and Information Engineering,Hubei Normal University,Huangshi Hubei 435002,China)
出处 《计算机应用》 CSCD 北大核心 2023年第1期259-264,共6页 journal of Computer Applications
基金 国家自然科学基金资助项目(61976085)。
关键词 眼底图像 注意力机制 involution算子 二阶特征融合 OIA-ODIR数据集 fundus image attention mechanism involution operator second-order feature fusion OIA-ODIR dataset
  • 相关文献

参考文献3

二级参考文献30

  • 1陈粟,倪林.一种特征脸分析和小波变换相结合的人脸识别方法[J].计算机应用,2004,24(10):75-77. 被引量:11
  • 2周国民,陈勇,李国军.人脸识别中应用小波变换的两个关键问题[J].浙江大学学报(理学版),2005,32(1):34-38. 被引量:27
  • 3RUSSO F. An Image Enhancement Technique Combining Sharpening and Noise Reduction[J]. IEEE Transactions on Instrumentation and Measurement, 2002, 51(4): 824-828.
  • 4LAND EH. The Retinex theory of Color Vision[J]. Scientific American, 1977, 237(6) : 108 - 129.
  • 5FOI A. Another Wavelet Approach to Higher Dimensionality :Ridgelet transforms[A]. Tampere International Center for Signal Processing[C]. 2004.
  • 6DONOHO DL, DUNCAN MR, TRANSFORM DC. Strategy, Implementation and Experiments[EB/OL]. http://www-stat, stanford.edu/~ donoho/Reports/1999/DCvT, pdf.
  • 7STARCK JL, MURTAGH F, CANDeS EJ, et al. Gray and Color Image Contrast Enhancement by the Curvelet Transform[J]. IEEE Transactions on Image Processing, 2003, 12(6) : 706 -717.
  • 8KIM TK, PAIK JK, KANG BS. Contrast enhancement system using spatialIy adaptive histogram equalization with temporal filtering[J].IEEE Transactions on Consumer Electronics, 1998, 44( 1 ): 82 -86.
  • 9STARK JA, FITZGERALID WJ. An Alternative Algorithm for Adaptive Histogram Equalization[J]. Graphical Models and Image Processing, 1996, 58(2) : 180 - 185.
  • 10KIM JY, KIM LS, HWANG SH. An Advanced Contrast Enhancement Using Partially Overlapped Sub-Block Histogram Equalization[J]. IEEE Transactions on Circuits and Systems for Video Technology, 2001, 11(4): 475-484.

共引文献72

同被引文献51

引证文献6

二级引证文献2

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部