摘要
Deep learning with convolutional neural networks(CNNs)has achieved great success in the classification of various plant diseases.However,a limited number of studies have elucidated the process of inference,leaving it as an untouchable black box.Revealing the CNN to extract the learned feature as an interpretable form not only ensures its reliability but also enables the validation of the model authenticity and the training dataset by human intervention.In this study,a variety of neuron-wise and layer-wise visualization methods were applied using a CNN,trained with a publicly available plant disease image dataset.We showed that neural networks can capture the colors and textures of lesions specific to respective diseases upon diagnosis,which resembles human decision-making.While several visualizationmethods were used as they are,others had to be optimized to target a specific layer that fully captures the features to generate consequential outputs.Moreover,by interpreting the generated attention maps,we identified several layers that were not contributing to inference and removed such layers inside the network,decreasing the number of parameters by 75%without affecting the classification accuracy.The results provide an impetus for the CNN black box users in the field of plant science to better understand the diagnosis process and lead to further efficient use of deep learning for plant disease diagnosis.
基金
This research was supported by Japan Science and Tech-nology Agency (JST) PRESTO[Grants nos.JPMJPR17O5(Yosuke'Toda) and JPMJPR17O3(Fumio Okura)].