期刊文献+

迁移学习下高分快视数据道路快速提取 被引量:4

Rapid road extraction from quick view imagery of high-resolution satellites with transfer learning
原文传递
导出
摘要 目的传统的道路提取方法自动化程度不高,无法满足快速获取道路信息的需求。使用深度学习的道路提取方法多关注精度的提升,网络冗余度较高。而迁移学习通过将知识从源领域迁移到目标领域,可以快速完成目标学习任务。因此,本文利用高分辨率卫星快视数据快速获取的特性,构建了一种基于迁移学习的道路快速提取深度神经网络。方法采用基于预训练网络的迁移学习方法,可以将本文整个道路提取过程分为两个阶段:首先在开源大型数据库Image Net上训练源网络,保存此阶段最优模型;第2阶段迁移预训练保存的模型至目标网络,利用预训练保存的权重参数指导目标网络继续训练,此时快视数据作为输入,只做目标任务的定向微调,从而加速网络训练。总体来说,前期预训练是一个抽取通用特征参数的过程,目标训练是针对道路提取任务特化的过程。结果本文构建的基于迁移学习的快速道路提取网络,迁移预训练模型与不迁移相比验证精度提升6.0%,单幅尺寸为256×256像素的数据测试时间减少49.4%。快视数据测试集平均精度可达88.3%。截取一轨中7304×6980像素位于天津滨海新区的快视数据,可在54 s内完成道路提取。与其他迁移模型对比,本文方法在快速预测道路的同时且能达到较高的准确率。结论实验结果表明,本文针对高分卫星快视数据,提出的利用预训练模型初始化网络能有效利用权重参数,使模型趋于轻量化,使得精度提升的同时也加快了提取速度,能够实现道路信息快速精准获取。 Objective Quick view data generated by high-resolution satellites provide real-time reception and full resolution for quick view imaging.Such imaging offers a timely source of data for practical applications,such as fire detection,moving window display,disaster observation,and military information acquisition.Road extraction from remote sensing images has been a popular research topic in the field of remote sensing image analysis.Traditional object-oriented methods are not highly automated,and road features require prior knowledge for manual selection and design.These conditions lead to problems in real-time road information acquisition.The popular deep learning road extraction method mainly focuses on the improvement of precision and lacks research on the timeliness of road information extraction.Transfer learning can rapidly complete the task in the target area through weight sharing among different fields and make the model algorithm highly personalized.A transfer learning deep network for rapidly extracting roads is constructed to utilize quick view data from highresolution satellites.Method First,we propose a least-square fitting method of devignetting to solve the most serious radiation problem of TDICCD(time delay and integration charge coupled devices)vignetting phenomenon appearing in raw quick view data.The results of the preprocessing of the quick view data serve as our training dataset.Then,we choose Link Net as the target network after comparing the performance among different real-time semantic segmentation networks,such as ENet,U-Net,Link Net,and D-LinkNet.Link Net is efficient in computation memory,can learn from a relatively small training set,and allows residual unit ease training of deep networks.The rich bypass links each encoder with decoder.Thus,the networks can be designed with few parameters.The encoder starts with a kernel of size 7×7.In the next encoder block,its contracting path to capture context uses 3×3 full convolution.We use batch normalization in each convolutional layer,followed by Re LU nonlinearity.Reflection padding is used to extrapolate the missing context in the training data for predicting the pixels in the border region of the input image.The input of each encoder layer of Link Net is bypassed to the output of its corresponding decoder.Lost spatial information about the max pooling can then be recovered by the decoder and its upsampling operations.Finally,we modify Link Net to keep it consistent with Res Net34 network layer features,the so-called fine tuning,for accelerating Link Net network training process.Fine tuning is a useful efficient method of transfer learning.The use of Res Net34 weight parameter pretrained on Image Net initializing Link Net34 can accelerate the network convergence and lead to improved performance with almost no additional cost.Result In the process of devignetting quick view data,the least-square linear fitting method proposed in this study can efficiently remove the vignetting strip of the original image,which meets practical applications.In our road extraction experiment,Link Net34 using the pretrained Res Net34 as encoder has a 6%improvement in Dice accuracy compared with that when using Res Net34 not pretrained on the valid dataset.The time consumption of a single test feature map is reduced by 39 ms,and the test Dice accuracy can reach 88.3%.Pretrained networks substantially reduce training time that also helps prevent overfitting.Consequently,we achieve over 88%test accuracy and 40 ms test time on the quick view dataset.With an input feature map size of 3×256×256 pixels,the data of Tianjin Binhai with a size of 7304×6980 pixels take 54s.The original Link Net using Res Net18 as its encoder only has a Dice coefficient of 85.7%.We evaluate Res Net50 and Res Net101 as pretrained encoders.The Dice accuracy of the former is not improved,whereas the latter takes too much test time.We compare the performance of Link Net34 with those of three other popular deep transfer models for classification,namely,U-Net;two modifications of Ternaus Net and Alub Net using VGG11(visual geometry group)and Res Net34 as encoders separately;and a modification of D-LinkNet.The two U-Net modifications are likely to incorrectly recognize roads as background or recognize something nonroad,such as tree,as road.D-LinkNet has higher Dice than Link Net34 on the validation set,but the testing time takes59 ms more than that of Link Net34.Link Net34 avoids the weaknesses of Ternuas Net and Alub Net and makes better predictions than them.The small nonroad gap between two roads can also be avoided.Many methods mix the two roads into one.The method proposed in this study generally achieves good connectivity,accurate edge,and clear outline in the case of complete extraction of the entire road and fine location.It is especially suitable for rural linear roads and the extraction of area roads in towns.However,the extraction effect for complex road networks in urban areas is incomplete.Conclusion In this study,we build a deep transfer learning neural network,Link Net34,which uses a pretrained network,Res Net34,as an encoder.Res Net34 allows Link Net34 to learn without any significant increase in the number of parameters,solves the problem that the bottom layer features randomly initialized with weights of neural networks are inadequately rich,and accelerates network convergence.Our approach demonstrates the improvement in Link Net34 by the use of the pretrained encoder and the better performance of Link Net34 than other real-time segmentation architecture.The experimental results show that Link Net34 can handle road properties,such as narrowness,connectivity,complexity,and long span,to some extent.This architecture proves useful for binary classification with limited data and realizes fast and accurate acquisition of road information.Future research should consider increasing the quick view database.The pretrained network Link Net34 trains on the expanded quick view database and then transfers.The"semantic gap"between the source and target networks is reduced,and the data distribute similarly.These features are conducive to model initialization.
作者 张军军 万广通 张洪群 李山山 冯旭祥 Zhang Junjun;Wan Guangtong;Zhang Hongqun;Li Shanshan;Feng Xuxiang(Institute of Remote Sensing and Digital Earth,Chinese Academy of Sciences,Beijing 100094,China;University of Chinese Academy of Sciences,Beijing 100049,China)
出处 《中国图象图形学报》 CSCD 北大核心 2020年第7期1501-1512,共12页 Journal of Image and Graphics
基金 中国科学院战略性先导科技专项(A类)地球大数据科学工程子课题项目:CASEarth小卫星产品服务研究(XDA19010401) 中国科学院遥感与数字地球研究所集成课题项目(Y6JD260057)。
关键词 高分辨率卫星 快视数据 道路快速提取 迁移学习 微调 high-resolution satellite quick view data fast road extraction transfer learning fine-tuning
  • 相关文献

参考文献3

二级参考文献17

  • 1侯明辉.基于仿真数据的HJ-1C卫星SAR成像处理方法[J].中国科学:信息科学,2011,41(S1):42-54. 被引量:1
  • 2郭琨毅,盛新庆.高性能Beowulf集群的SAR并行成像的处理[J].现代雷达,2005,27(1):38-40. 被引量:4
  • 3熊君君,王贞松,姚建平,石长振.星载SAR实时成像处理器的FPGA实现[J].电子学报,2005,33(6):1070-1072. 被引量:19
  • 4Yassin M Y,IEEE Trans Pattern Analysis Machine Intelligence,2000年,22卷,3期,227页
  • 5Hu Xiangyun,Int Achives of Photogrammetry and Remote Sensing XXXIII(Part B3),2000年,994页
  • 6Gruen A,Photogrammetric Engineering and Remote Sensing,1997年,63卷,8期,985页
  • 7Steger C,Automatic Extraction of Manmade Objects from Aerial and Space Images(2),1997年,245页
  • 8TRINDER J C,Automatic Extraction of Manmade Objects from Aerial and Space Images(2),1997年,257页
  • 9Gruen A,ISPRS J Photogrammetry Remote Sensing,1995年,50卷,4期,11页
  • 10TRINDER J C,Automatic Extraction of Manmade Objects from Aerial and Space Images,1995年,95页

共引文献157

同被引文献27

引证文献4

二级引证文献6

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部