Recognition Method for Crop Canopies Based on Thermal Infrared Image Processing Technology
MA Xiao-dan1*, LIU Meng1, GUAN Hai-ou1, WEN Feng-rui1, LIU Gang2
1. College of Information and Electrical Engineering, Heilongjiang Bayi Agricultural University, Daqing 163319, China
2. Key Laboratory of Modern Precision Agriculture System Integration Research, Ministry of Education, China Agricultural University,Beijing 100083, China
Abstract:In order to solve the problem that the gray level distribution of crop canopy thermal infrared image isuneven and has large noise, the traditional image segmentation method is difficult to realize the effective recognition of its target region. In this study, by the thermal infrared images of adzuki bean canopy’s in the seedling stage was taken as the research object, Combining fuzzy neural network and affine transformation, a crop canopy recognition model based on thermal infrared image processing technology was proposed. First, the adaptive characteristics of the five-layer linear normalized fuzzy neural network were used to select the Gaussian membership function to automatically calculate the inference rules for canopy visible light image recognition, effectively segmenting the canopy area in the visible light image. By analyzing three segmentation indexes and entropy, the canopy segmentation quality of visible light images was quantitatively evaluated. When the network iterates 38 times, the error precision was 0.000 952, and the visible light image of the crop canopy was obtained. The average effective partition rate of the algorithm was 96.13%, and the entropy value of the image source average information was 2.454 4~5.198, which was only 0.245 9 different from the entropy of the canopy image obtained by the standard algorithm. Then, using the effective area of the canopy to obtain the visible light image as a reference image, the affine transformation algorithm was used to adjust the image transformation factors such as optimal translation, rotation, and scaling. To register the raw thermal infrared image. And a canopy thermal infrared image recognition method based on affine transformation was proposed. For a crop thermal infrared image with an initial temperature range of 16.35~19.92, when the rotation amplitude was 1.0 and the zoom factor was 0.9, the maximum temperature difference of the target image obtained as the optimal registration parameter of the heterogeneous image was 3.17 ℃. Relative The average temperature of the original image decreased from 18.711 ℃ to 17.790 ℃, and the crop canopy recognition based on thermal infrared image processing technology was realized. Finally, mutual information was used as a monitoring index to evaluate the thermal infrared image recognition method of crop canopy. In the canopy thermal infrared image recognition method proposed in this study, the average mutual information between the acquired target image and the initial thermal infrared image was 4.368 7, while the average mutual information between the standard target image and the initial thermal infrared image was 3.981 8, and the difference between the two was only 0.486 9. At the same time, the average temperature difference between the two canopy thermal infrared images was 0.25 ℃, which effectively eliminates the background noise of the original thermal infrared images. The research results show that the effectiveness and practicability of this research method could provide a technical reference for the application of thermal infrared images to reflect the characteristic parameters of crop physiological and ecological information.
[1] Khanal S, Fulton J, Shearer S. Computers and Electronics in Agriculture, 2017, 139: 22.
[2] CHEN Xin-xin, LIU Zi-yi, LÜ Mei-qiao, et al(陈欣欣, 刘子毅, 吕美巧, 等). Spectroscopy and Spectral Analysis(光谱学与光谱分析), 2019, 39(3): 730.
[3] ZHANG Zhi-tao, XU Chong-hao, TAN Cheng-xuan, et al(张智韬, 许崇豪, 谭丞轩, 等). Transactions of The Chinese Society for Agricultural Machinery(农业机械学报), 2019, 50(8): 213.
[4] Tanner C B. Agronomy Journal, 1963, 55(2): 210.
[5] CHEN Yue-wei, PENG Dao-gang, XIA Fei, et al(陈跃伟, 彭道刚, 夏 飞, 等). Laser & Infrared(激光与红外), 2018, 48(3): 401.
[6] Qiao Y, Wei Z, Zhao Y. Sensors, 2017, 17(8): 1811.
[7] Osher S, Sethian J A. Journal of Computational Physics, 1988, 79(1): 12.
[8] Ma J, Chen C, Li C, et al. Information Fusion, 2016, 31: 100.
[9] Zhang R, Zhu S, Zhou Q. A. Sensors, 2016, 16(10): 1756.
[10] Coiras E, Santamaria J, Miravet C A. Optical Engineering, 2000, 39(1): 282.
[11] ZHANG Zhi-tao, BIAN Jiang, HAN Wen-ting, et al(张智韬, 边 江, 韩文霆, 等). Transactions of the Chinese Society for Agricultural Machinery(农业机械学报), 2018, 49(10): 250.
[12] GUAN Hai-ou, YI Shu-juan, JIAO Feng, et al(关海鸥, 衣淑娟, 焦 峰, 等). Transactions of the Chinese Society for Agricultural Machinery(农业机械学报), 2012, 43(5): 162.
[13] MA Lan, CHEN Xiao-yong, DENG Guo-chen(马 兰, 陈筱勇, 邓国臣). Science of Surveying and Mapping(测绘科学), 2017, 42(5): 87.
[14] HE Han-zhi, ZHU Hong, WANG Wei(何瀚志, 朱 红, 王 伟). Computer Engineering and Design(计算机工程与设计), 2019, 40(3): 856.