In the process of aquaculture,monitoring the number of fish bait particles is of great significance to improve the growth and welfare of fish.Although the counting method based on onvolutional neural network(CNN)achie...In the process of aquaculture,monitoring the number of fish bait particles is of great significance to improve the growth and welfare of fish.Although the counting method based on onvolutional neural network(CNN)achieve good accuracy and applicability,it has a high amount of parameters and computation,which limit the deployment on resource-constrained hardware devices.In order to solve the above problems,this paper proposes a lightweight bait particle counting method based on shift quantization and model pruning strategies.Firstly,we take corresponding lightweight strategies for different layers to flexibly balance the counting accuracy and performance of the model.In order to deeply lighten the counting model,the redundant and less informative weights of the model are removed through the combination of model quantization and pruning.The experimental results show that the compression rate is nearly 9 times.Finally,the quantization candidate value is refined by introducing a power-of-two addition term,which improves the matches of the weight distribution.By analyzing the experimental results,the counting loss at 3 bit is reduced by 35.31%.In summary,the lightweight bait particle counting model proposed in this paper achieves lossless counting accuracy and reduces the storage and computational overhead required for running convolutional neural networks.展开更多
基金supported by the National Key Research and Development Program of China(No.2019YFD0901000)。
文摘In the process of aquaculture,monitoring the number of fish bait particles is of great significance to improve the growth and welfare of fish.Although the counting method based on onvolutional neural network(CNN)achieve good accuracy and applicability,it has a high amount of parameters and computation,which limit the deployment on resource-constrained hardware devices.In order to solve the above problems,this paper proposes a lightweight bait particle counting method based on shift quantization and model pruning strategies.Firstly,we take corresponding lightweight strategies for different layers to flexibly balance the counting accuracy and performance of the model.In order to deeply lighten the counting model,the redundant and less informative weights of the model are removed through the combination of model quantization and pruning.The experimental results show that the compression rate is nearly 9 times.Finally,the quantization candidate value is refined by introducing a power-of-two addition term,which improves the matches of the weight distribution.By analyzing the experimental results,the counting loss at 3 bit is reduced by 35.31%.In summary,the lightweight bait particle counting model proposed in this paper achieves lossless counting accuracy and reduces the storage and computational overhead required for running convolutional neural networks.