The proposal was developed based on preliminary research activities with promising results in the field. In 2019, we collected preliminary data from a 70 acre field in Carmi, IL in support of this project. A Matrice 210 UAV was used to collect imagery with a X5S RGB camera and an Altum Multispectral/Thermal camera. Yield data was collected, and the field was grid mapped and sampled on 1/3 acre grids resulting in 224 soil samples. Preliminary analysis reveal that SCN egg counts are correlated with colors and patterns in UAV images associated with greater plant stress. Figure 1 is a RGB image of a field in Carmi, one of the proposed locations. The areas highlighted in the image are suffering due to nematode feeding. Figure 2 shows the intensity of SCN infestation in this field ranging from a low of 1,300 eggs/100cc soil to over 40,000 eggs /100 cc soil. We are completing data analysis and refining our models and techniques with this test field for the proposed trials.
For this project, field trials will be established at two locations in southern Illinois. One location will be in Carmi, IL and second location will be identified that represents a different soil type.
These locations will serve as data generators for the predictive models. Two different experimental setups will be used in the field trials:
First experimental setup - The experiment will be conducted at two locations. Soybean will be planted in SCN infested fields. These fields will represent traditional fields in Illinois with varying infestation level of SCN.
Second experimental setup - The study will be conducted at two locations and will consist of 24 rows treated with a fungicide and ILEVO seed treatment and 24 rows with only a fungicide treatment. The two treatments (with nematicide and without) will be replicated 4 times.
Fields that have a range of low (0 - 500 eggs per 100 cc soil) to high (<15,000 eggs per 100 cc of soil) will be selected. The field sizes will range from 70 – 150 acres. The fields will be mapped into 1/3-acre grids, where soil samples will be collected. Soil samples will be collected at planting, midseason, and harvest to determine the population densities of SCN and for a soil fertility analysis. By determining the SCN population densities, we will be able to validate the areas in the field with high populations and relate those populations to the imagery. At the end of the season, soybeans will be harvested, and the yield data will be collected from yield monitors. Multispectral and visual UAV data will be collected at each field site, along with close-range spectral data collected with a spectroradiometer. Images will be collected using an X5S RGB camera and an Altum multispectral camera mounted to a UAV. In the strips where there are different treatments, 20 spectral readings will be collected using a spectroradiometer, and the data will then be averaged to represent each strip. Imagery and spectra will be collected 30, 45, 60, and 75 DAP (days after
planting).
Collection of multispectral data - Every flight mission will be designed with 80% forward overlaps and 80% side overlaps. All flight missions will be completed under the local drone related legislative and campus regulations. Each flight mission will cover the entirety of the trial area and the time of flight will be dependent on the size of the area. To ensure that the imagery will be correctly georeferenced, ground control points (GCPs) will be used to correctly project the imagery. The GCPs will be placed randomly throughout the study area and the GPS coordinates will be collected. RTK correction will be used to give sub-centimeter accuracy of the GCPs.
Development of AI algorithms - As shown in previous research that nematode (e.g., Heterodera schactii) activities present unique spectral and textual signatures on the phenological representation of soybean fields, we propose to develop agricultural artificial intelligence (Agro-AI) models to identify the areas with high nematode densities for better management of SCN. The Agro-AI models are especially suitable to identify these fuzzy-look features that are difficult to be examined with regular field or image inspection. This study will collect UAV image samples to develop deep learning models. The manually labeled UAV images containing both nematodeaffected and healthy areas will be randomly divided 2/3 and 1/3 as the training and testing samples respectively. Data augmentation techniques such as rotation, reflection, and Gaussian smoothing kernel filtering will be applied to the original training samples to increase the size of the training dataset and prevent the deep neural network from overfitting. The training samples will then be used to feed a deep learning convolutional neural network (CNN) algorithm. The Faster R-CNN detector can be trained by the layers with ten epochs. The haar-like cascading detector can be trained by 50 training stages with a 0.01 false alarming rate. Every testing sample is treated with the well-trained faster R-CNN detector and cascading detector. Their corresponding CNN feature and haar-like feature could be extracted. The likelihood of each feature reflecting nematodes infection are predicted by the detector. The non-maximum suppression (NMS) will be used to suppress redundant detections surrounding the same feature. Every training image is sliced to 9 equal size small images to enhance training efficiency. Final training samples used to train the model are enlarged to 5 times the original data. A valid detection is defined by an overlap ratio of the predicted bounding box and ground truth box that is beyond 0.4.
Testing samples will be labeled as three groups based on the severity of the nematodes’ destructive effects based on expert interpretation. Accuracy on every single image and corresponding
detection method will be recorded as the response and method factor, respectively. As the response is bounded from 0 to 1, a beta Generalized Linear Model (GLM) is fitted to interpret the treatment effect of the detection methods and environmental types on detection accuracy. The final precision and recall of the detection by every testing sample will be recorded. The F-score, a measure of a
test’s accuracy considering both the precision P and the recall R, is calculated to assess the performance of this model. The percentage of training samples varying from 10%-100% with 10%
interval can be used to determine the minimum training data size required for accurate detection.