4 research outputs found
Explaining hyperspectral imaging based plant disease identification: 3D CNN and saliency maps
Our overarching goal is to develop an accurate and explainable model for
plant disease identification using hyperspectral data. Charcoal rot is a soil
borne fungal disease that affects the yield of soybean crops worldwide.
Hyperspectral images were captured at 240 different wavelengths in the range of
383 - 1032 nm. We developed a 3D Convolutional Neural Network model for soybean
charcoal rot disease identification. Our model has classification accuracy of
95.73\% and an infected class F1 score of 0.87. We infer the trained model
using saliency map and visualize the most sensitive pixel locations that enable
classification. The sensitivity of individual wavelengths for classification
was also determined using the saliency map visualization. We identify the most
sensitive wavelength as 733 nm using the saliency map visualization. Since the
most sensitive wavelength is in the Near Infrared Region(700 - 1000 nm) of the
electromagnetic spectrum, which is also the commonly used spectrum region for
determining the vegetation health of the plant, we were more confident in the
predictions using our model
Plant disease identification using explainable 3D deep learning on hyperspectral images
Background
Hyperspectral imaging is emerging as a promising approach for plant disease identification. The large and possibly redundant information contained in hyperspectral data cubes makes deep learning based identification of plant diseases a natural fit. Here, we deploy a novel 3D deep convolutional neural network (DCNN) that directly assimilates the hyperspectral data. Furthermore, we interrogate the learnt model to produce physiologically meaningful explanations. We focus on an economically important disease, charcoal rot, which is a soil borne fungal disease that affects the yield of soybean crops worldwide.
Results
Based on hyperspectral imaging of inoculated and mock-inoculated stem images, our 3D DCNN has a classification accuracy of 95.73% and an infected class F1 score of 0.87. Using the concept of a saliency map, we visualize the most sensitive pixel locations, and show that the spatial regions with visible disease symptoms are overwhelmingly chosen by the model for classification. We also find that the most sensitive wavelengths used by the model for classification are in the near infrared region (NIR), which is also the commonly used spectral range for determining the vegetative health of a plant.
Conclusion
The use of an explainable deep learning model not only provides high accuracy, but also provides physiological insight into model predictions, thus generating confidence in model predictions. These explained predictions lend themselves for eventual use in precision agriculture and research application using automated phenotyping platforms
Task-driven learned hyperspectral data reduction using end-to-end supervised deep learning
An important challenge in hyperspectral imaging tasks is to cope with the large number of spectral bins. Common spectral data reduction methods do not take prior knowledge about the task into account. Consequently, sparsely occurring features that may be essential for the imaging task may not be preserved in the data reduction step. Convolutional neural network (CNN) approaches are capable of learning the specific features relevant to the particular imaging task, but applying them directly to the spectral input data is constrained by the computational efficiency. We propose a novel supervised deep learning approach for combining data reduction and image analysis in an end-to-end architecture. In our approach, the neural network component that performs the reduction is trained such that image features most relevant for the task are preserved in the reduction step. Results for two convolutional neural network architectures and two types of generated datasets show that the proposed Data Reduction CNN (DRCNN) approach can produce more accurate results than existing popular data reduction methods, and can be used in a wide range of problem settings. The integration of knowledge about the task allows for more image compression and higher accuracies compared to standard data reduction methods