Next Article in Journal
A Supervised Method for Nonlinear Hyperspectral Unmixing
Previous Article in Journal
Wetland Classification Based on a New Efficient Generative Adversarial Network and Jilin-1 Satellite Image
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Improving Field-Scale Wheat LAI Retrieval Based on UAV Remote-Sensing Observations and Optimized VI-LUTs

1
Key Laboratory of Ecosystem Network Observation and Modeling, Institute of Geographic Sciences and Natural Resources Research, Chinese Academy of Science, Beijing 100101, China
2
College of Resources and Environment, University of Chinese Academy of Sciences, Beijing 100049, China
3
Shandong Dongying Institute of Geographic Sciences, Institute of Geographic Sciences and Natural Resources Research, Chinese Academy of Science, Dongying 257000, China
4
CAS Engineering Laboratory for Yellow River Delta Modern Agriculture, Institute of Geographic Sciences and Natural Resources Research, CAS, Beijing 100101, China
5
State Key Laboratory of Resources and Environmental Information System, Institute of Geographic Sciences and Natural Resources Research, Chinese Academy of Science, Beijing 100101, China
6
Changchun Institute of Optics, Fine Mechanics and Physics, Chinese Academy of Sciences, Changchun 130033, China
7
Yusense Information Technology and Equipment (Qingdao) Ltd., Qingdao 266000, China
8
Research Center for UAV Applications and Regulation, Chinese Academy of Science, Beijing 100101, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2019, 11(20), 2456; https://0-doi-org.brum.beds.ac.uk/10.3390/rs11202456
Submission received: 28 August 2019 / Revised: 10 October 2019 / Accepted: 19 October 2019 / Published: 22 October 2019
(This article belongs to the Section Remote Sensing in Agriculture and Vegetation)

Abstract

:
Leaf area index (LAI) is a key biophysical parameter for monitoring crop growth status, predicting crop yield, and quantifying crop variability in agronomic applications. Mapping the LAI at the field scale using multispectral cameras onboard unmanned aerial vehicles (UAVs) is a promising precision-agriculture application with specific requirements: The LAI retrieval method should be (1) robust so that crop LAI can be estimated with similar accuracy and (2) easy to use so that it can be applied to the adjustment of field management practices. In this study, three UAV remote-sensing missions (UAVs with Micasense RedEdge-M and Cubert S185 cameras) were carried out over six experimental plots from 2018 to 2019 to investigate the performance of reflectance-based lookup tables (LUTs) and vegetation index (VI)-based LUTs generated from the PROSAIL model for wheat LAI retrieval. The effects of the central wavelengths and bandwidths for the VI calculations on the LAI retrieval were further examined. We found that the VI-LUT strategy was more robust and accurate than the reflectance-LUT strategy. The differences in the LAI retrieval accuracy among the four VI-LUTs were small, although the improved modified chlorophyll absorption ratio index-lookup table (MCARI2-LUT) and normalized difference vegetation index-lookup table (NDVI-LUT) performed slightly better. We also found that both of the central wavelengths and bandwidths of the VIs had effects on the LAI retrieval. The VI-LUTs with optimized central wavelengths (red = 612 nm, near-infrared (NIR) = 756 nm) and narrow bandwidths (~4 nm) improved the wheat LAI retrieval accuracy (R2 ≥ 0.75). The results of this study provide an alternative method for retrieving crop LAI, which is robust and easy use for precision-agriculture applications and may be helpful for designing UAV multispectral cameras for agricultural monitoring.

Graphical Abstract

1. Introduction

Crop growth varies between/within fields owing to differences in environmental conditions (e.g., climate, soil, and water), management (e.g., fertilization, irrigation, and pest/disease control) [1], and crop genotypes. Mapping the spatial variation of crop canopy traits and the factors affecting them is important for agriculture since these field-scale maps are helpful for precise agronomic management. The precise management is essential to improve crop production and light-utilization efficiency [2]. Satellite remote sensing has been widely applied in crop monitoring. However, its coarse spatial resolutions (e.g., moderate-resolution imaging spectroradiometer (MODIS) (1 km), Sentinel 1A (10 m), Landsat (30 m)) fail to fulfill the requirement of precise mapping at the field scale (~cm level) [3]. In contrast to satellite platforms, ground platforms are able to map at the centimeter scale, but they fail to be applied in relatively large areas [4]. However, unmanned aerial vehicles (UAVs) are able to solve these issues. UAVs are easy to use, have a low cost, and are able to capture data with a high spatial resolution (~cm); they provide a promising platform for monitoring the crop growth status and evaluating the results of field management. Multispectral cameras onboard UAVs capture the spectral reflectance of the vegetation canopy at specific bands. These remote-sensing observations can be used to obtain the leaf area index (LAI) which is a crucial biophysical parameter related to crop growth [3,5]. LAI data can be used to quantify crop variability [6], monitor crop growth [7], predict the crop yield [8], estimate the amount of aboveground biomass [9], and evaluate the effects of field management [10]. However, LAI retrieval methods based on UAV remotely sensed data must be robust and easy use so that the crop LAI will be estimated with similar accuracy to fulfill the requirement of practical application.
The PROSAIL, a radiative transfer model, which is run in a forward mode with appropriate parameter combinations (e.g., LAI, chlorophyll content, and leaf structure parameters) to simulate the spectral reflectance of vegetation canopy, is widely employed to retrieve LAI. The PROSAIL inputs (parameter combinations) and outputs (spectral reflectance) are used to generate look-up tables (LUTs) [11]. Thus, LAI, one of the PROSAIL inputs, can be retrieved by searching for the corresponding spectral reflectance in a given LUT using cost functions. The cost function is developed by simulated reflectance and measure reflectance. This LUT inversion method can reduce computational loads, lower the risk of converging to local minima, and help to avoid the ill-posed problems caused by uncertainties of the measurements and models [12].
The reflectance-LUT strategy, searching for the optimal simulated LAI by simulated reflectance and measured reflectance, is widely employed, particularly for traditional satellite remote sensing (Table 1). For UAV/airborne remote sensing, reflectance-LUTs are generated via multispectral or hyperspectral imaging. LAI retrieval based on UAV/airborne hyperspectral reflectance-LUTs can achieve high accuracy [13]. However, the multispectral cameras onboard UAVs typically have only three to six channels [14]. Spectral information for crops captured by multispectral cameras may be deficient owing to the limited bands. Therefore, the direct use of multispectral reflectance for LAI retrieval may be inaccurate. Generally, for multispectral UAV/airborne data, LAI estimation requires auxiliary methods, such as multi-angle observations for more vegetation information [12] and combinations with other algorithms (e.g., Bayesian networks [15], artificial neural networks, and random forests [16]). Multi-angle observations cost extra time and labor; machine learning requires ground sampling for the development of models. Therefore, these auxiliary methods may increase the difficulty of UAV multispectral remote sensing in practical application.
Vegetation indices (VIs) are calculated by different band combinations and mathematical formulas. Compared with spectral reflectance, vegetation indices (VIs) can provide more vegetation information, since VIs can offset parts of the errors from the surrounding environment and reveal implied plant information. Therefore, the VI-LUT strategy, searching for the optimal simulated LAI in a given LUT by simulated VIs (calculated from the PROSAIL outputs) and measured VIs, is an alternative strategy. Most of the previous studies related to VI-LUTs were based on satellite remote sensing [17]; no references were based on UAV/airborne remote sensing. Thus, we attempted to bridge this gap by investigating the robustness of the VI-LUT strategy for LAI retrieval using UAV multispectral remote sensing.
Table 1. Summary of LAI retrieval based on LUTs generated from the PROSAIL model.
Table 1. Summary of LAI retrieval based on LUTs generated from the PROSAIL model.
LUTsPlatforms SensorsReferences
Reflectance-LUTsUAV/AirborneMultispectral camera[12,15]
Hyperspectral camera[11,13]
SatelliteZY-3 MUX, GF-1 WFV, HJ-1 CCD[18]
Landsat[19,20,21]
Sentinel-2[22,23,24]
RapidEye[22]
CHRIS/PROBA[25]
MODIS[20,26]
AWiFS[27]
Ground-basedHyperspectral spectrometer [28,29,30]
VI-LUTssatelliteCHRIS/PROBA[31]
RapidEye[17]
Landsat[32]
LAI = Leaf area index; UAV = unmanned aerial vehicle; Reflectance-LUTs = reflectance-based lookup tables; VI-LUTs = Vegetation index-based lookup tables; ZY-3 MUX = the resources satellite No. 3 (ZY-3) with a multi-spectral camera (MUX); GF-1 WFV = the satellite GaoFen No. 1 (GF-1) with a wide field of view (WFV) camera; HJ-1 CCD = the environment satellite (HJ-1) with a charge-coupled device (CCD) sensor. CHRIS/PROBA = Compact High-Resolution Imaging Spectrometer/Project for On-Board Autonomy; MODIS = moderate-resolution imaging spectroradiometer; AWiFS = Advanced Wide-Field Sensor
The radiometric data acquired from various sensors have different spectral and spatial characteristics because of their sensor-dependence [3]. Previous studies have already demonstrated that normalized difference vegetation index (NDVI) is significantly affected by differences in bandwidth and spatial resolutions [33,34]. Owing to weight reduction and downsizing, spectral cameras onboard UAVs have technical differences in the central wavelengths and bandwidths compared with cameras onboard ground and satellite platforms. Most VIs were developed based on the satellite (e.g., enhanced vegetation index without a blue band (EVI2) [35], atmospherically resistant vegetation index (ARVI) [36]) and ground-based remote-sensing data (e.g., soil adjusted vegetation index (SAVI), modified soil adjusted vegetation index (MSAVI) [37]). Thus the calculation of these VIs using field-scale UAV images might cause uncertainties for crop parameter estimation. Red (R) and near-infrared (NIR) bands have been widely used for calculations of VIs because the spectral reflectance at R and NIR bands is sensitive to vegetation phenotypic traits (e.g., LAI and chlorophyll content) [38,39,40,41]. Previous studies have attempted to optimize the construction of VIs by evaluating the central wavelengths and bandwidths of R and NIR bands [39]. Therefore, this study followed this method to improve the performance of UAV-based VIs on LAI estimation at the field scale.
The objectives of this study were to (1) search for a feasible LUT strategy for field-scale LAI retrieval using multispectral UAV images, and (2) improve the easiness to use, robustness, and accuracy of this method so that it can provide precise agronomic information at field scale. To that purpose, the performance of reflectance-LUT and VI-LUT strategies generated by the PROSAIL model for LAI retrieval were investigated using multispectral UAV images; the central wavelengths and bandwidths of the red and NIR bands for VIs calculations on LAI retrieval were evaluated and further optimized using UAV hyperspectral images.

2. Methodology

2.1. Study Area and Long-Term Experimental Plots

UAV observations were conducted at the Yucheng Comprehensive Experiment Station (YCES) of the Chinese Academy of Sciences (36.83° N, 116.57° E), which is located along the north side of the Lower Yellow River (Figure 1). The study area has a warm temperate and a semi-humid monsoon climate with an approximate annual mean temperature of 13.4 °C and an average annual precipitation of 576.70 mm, concentrated between July and September. The YCES is dominated by a typical cropping system of winter wheat (October–June) and summer maize (July–September).
We selected six types of long-term experimental plots with different tillage treatments, soil water content, and nutrient levels for experiments (Figure 1 and Table 2). All the experimental plots had been operational for >10 years. Owing to the obvious differences in crop traits caused by the existing gradients of agricultural treatments, the crop LAI dynamic range was large. Thus, the study area was perfect for UAV remote-sensing experiments.

2.2. Ground Measurements and UAV Flight Missions

2.2.1. Ground Measurements of LAI

Leaves of wheat were collected via destructive sampling for LAI measurements in the laboratory (Figure 2). The row spacing of all experimental plots was 20 cm. In order to ensure the reasonability and typicality of the study, wheat with relatively uniform canopies outside the border areas of each plot was selected for the experiment. A row of wheat with a length of 0.5 m was cut for LAI measurements. The leaf area was measured using an LI-3000C Leaf Area Meter (Li-COR Biosciences, Lincoln, NE, USA). The average LAI of each plot was calculated according to
LAI (m2/m2) = TLA/(a × b)
where TLA (m2) represents the total leaf area of crops within each plot, and a (m) and b (m) represent the width and length of each plot, respectively. Because wheat is row crops, the TLA was calculated according to
TLA = 2 × LA × nrows × b
where LA (m2) represents the measured leaf area of crops harvested within each plot, nrows represents the number of rows for each plot. With UAV flights, three LAI datasets were collected for evaluation of the LAI retrieval accuracy (Table 4). Each dataset exhibited a high coefficient of variation (CV) owing to the wide range of agricultural treatments, which contributed to a good model fit between the measured and simulated LAI. Because the flight coverage of the multispectral UAV observations (all six fields) differed from that of hyperspectral UAV observations (see Figure 1 for the two red rectangles), the number of ground samples for the LAI estimation was different for 15 May 2018.

2.2.2. UAV Flight Missions and Data Pre-Processing

On 15 May 2018, and 16 May 2019, two multispectral UAV flight observations were conducted. On 15 May 2018, one hyperspectral UAV flight observation was conducted. For the UAV missions, two types of UAV platforms (see Figure 3 for the DJ M100 four-rotator UAV (SZ DJI Technology Co., Shenzhen, Guangzhou, China) and DJ M600 Pro six-rotator UAV (SZ DJI Technology Co., Shenzhen, Guangzhou, China) equipped with two sensors (Table 5)—Micasense RedEdge-M multispectral camera (MicaSense, Seattle, WA, USA) and Cubert S185 hyperspectral camera (Cubert GmbH, Ulm, Germany)—were used. The two multispectral flights covered all six fields; the hyperspectral flight covered Field B, D, and E (see Figure 1 for the two red rectangles).
The UAV flight time of the three flight missions was from 10:00 to 14:00. For radiation correction, the reflectance of a spectral panel was collected during each flight. The radiation correction, image mosaicking, and orthography of the UAV images were conducted using Agisoft Photoscan (Agisoft LLC, St. Petersburg, Russia) and Pix4D Mapper 3.1.22 (Pix4D, S.A., Lausanne, Switzerland). ENVI 5.1 (ESRI, RedLands, CA, USA) and Python 2.7 were used for further data analyses and programming, respectively.

2.3. Retrieving LAI from UAV Data Using PROSAIL Model

The PROSAIL model, along with the UAV remote-sensing data, was used to retrieve wheat LAI [42]; it is derived from the combination of the PROSPECT blade model and the SAIL canopy structure model [43]. The PROSPECT model simulates the optical properties of leaves, from 400 to 2500 nm, with four inputs: leaf structure parameter (N), chlorophyll content (Chl), leaf water mass per area (LMA), and blade equivalent thickness (EWT). The SAILH model is a radiative transfer model on the canopy scale, in which vegetation is treated as a mixed medium with an assumption that the blade azimuth distribution is uniform [44]. In total, fourteen parameters (N, Chl, EWT, LMA, LAI, leaf carotenoid content (caro), brown pigment content, soil brightness parameter (psoil), hot-spot size parameter (hot spot), solar zenith and azimuth angles, view zenith and azimuth angles, and average leaf angle (ALA)) are required for running the PROSAIL model, most of which are difficult to obtain (Table 6) (e.g., ALA, caro, and hot spot).
High-dimensional LUTs are often generated to retrieve LAI because most PROSAIL inputs are difficult to reach in practice [11,24]. The computation processes for high-dimensional LUTs are slow and complex, which may hinder the application of LUTs to parameter inversion. Sensitivity analysis of models is an alternative way to solve this issue. It can identify the sensitive model inputs and evaluate their sensitivity levels; the changes of these sensitive inputs in a certain range will lead to obvious variations of model outputs. Thus some unessential PROSAIL variables can be set as constants so that the dimension of the LUTs decreases [12]. Sensitivity analysis includes local and global analyses. The global analysis can reveal the effects of each input and interaction among inputs on model outputs. In this study, the extended Fourier amplitude sensitivity test (EFAST) was used for global sensitivity analysis. The EFAST is a quantitative method based on variance [45]. The variance produced by the change of model outputs reveals the sensitive values or levels of model inputs [12].
LAI retrieval from the UAV data comprised three steps (Figure 4). Firstly, select appropriate VIs for LAI retrieval based on the (EFAST) global sensitivity analyses of the PROSAIL model; secondly, generate reflectance-LUTs and VI-LUTs by running the PROSAIL model; thirdly, retrieve LAI through cost functions (see Section 2.3.3 for the formula of cost functions) based on the LUTs. In step 3, both of the multispectral and hyperspectral UAV datasets were used. The multispectral datasets were used to investigate the robustness of LUT strategies for LAI retrieval (step 3.1); the hyperspectral dataset was used to improve the LAI retrieval robustness and accuracy (step 3.2).

2.3.1. Selecting Optimal VIs for LAI Retrieval (Global Sensitivity Analysis)

The global sensitivity analyses were conducted to select VIs for LAI retrieval. First, 5000 model input datasets (Range (EFAST), Table 6) with uniform distributions were generated using the Simlab 2.2.1 software (JRC, Italy). These generated datasets were the simulated inputs of the PROSAIL model. Second, the PROSAIL model was run in a forward mode to obtain model outputs (spectral reflectance) using the PyProSAIL package of Python 2.7 (http://teledetection.ipgp.jussieu.fr/prosail/, accessed on 24 June 2019). Then VIs (Table 7) were calculated using the PROSAIL outputs—spectral reflectance. Thirdly, datasets of the PROSAIL inputs and outputs were used to conduct the EFAST global sensitivity analysis using Simlab 2.2.1.
The reflectance values of the blue (B), green (G), red (R), red-edge (E), and NIR bands were denoted as ρB, ρG, ρR, ρE, and ρNIR, respectively. We selected some VIs that were used for the estimation of crop traits in previous studies (Table 7, where ρ1 = ρNIR). These VIs were divided into three groups: R-VIs (calculated using ρR and ρNIR), G-VIs (calculated using ρG and ρNIR), and E-VIs (calculated using ρE and ρNIR). In this study, variables of VIs were changed; thus VIs with the same formula but different variables were denoted as m-VIs (e.g., modified atmospherically resistant vegetation index (m-ARVI) represents R-ARVI (ρ2 = ρR), G-ARVI (ρ2 = ρG), and E-ARVI (ρ2 = ρE) in the R-VI, G-VI, and E-VI groups, respectively).
Table 7. Bands and VIs used in this study.
Table 7. Bands and VIs used in this study.
m-VIsEquationsR-VIsG-VIsE-VIsRef.
ρ2 = ρRρ2 = ρEρ2 = ρE
m-ARVI(ρ1 − 2ρ2 + ρB)/(ρ1 + 2ρ2ρB)[3]
m-EVI2.5 ((ρ1ρ2)/(ρ1 + 6ρ2 − 7.5ρB + 1))[46]
m-WDRVI(0.12ρ1ρ2)/(0.12ρ1 + ρ2)[46]
m-MSR ( ρ 1 / ρ 2 1 ) / ( ρ 1 / ρ 2   +   1 ) [47,48]
m-MSAVI2 0.5 [ ( 2 ρ 1 + 1 ) ( 2 ρ 1 + 1 ) 2 8 ( ρ 1 ρ 2 ) ] [38]
m-MCARI2 1.5 [ 2.5 ( ρ 1 ρ 2 ) 1.3 ( ρ 1 ρ G ) ] / [ ( 2 ρ 1 + 1 ) 2 ( 6 ρ 1 5 ρ 2 ) 0.5 ] [48]
m-MTVI11.2(1.2(ρ1ρG) − 2.5(ρ2ρG)) [48]
m-TVI0.5(120(ρ1ρG)) − 200(ρ2ρG) [48]
m-NDVI(ρ1ρ2)/(ρ1 + ρ2)[28]
m-OSAVI1.16(ρ1ρ2)/(ρ1 + ρ2 + 0.16)[46]
m-RVIρ1/ρ2[3]
m-SAVI(1 + 0.5)((ρ1ρ2)/(ρ1 + ρ2 + 0.5))[3]
m-NRIρ2/(ρ2 + ρE + ρ1) [49]
m-EVI22.5(ρ1ρ2)/(ρ1 + 2.4ρ2 + 1)[35,46]
R = Red, G = Green, E = red-edge; m-ARVI = modified atmospherically resistant vegetation index; m-EVI = modified enhanced vegetation index; m-WDRVI = modified wide dynamic range vegetation index; m-MSR = modified simple ratio; m-MSAVI2 = modified secondary optimized soil adjusted vegetation index; m-MCARI2 = modified improved chlorophyll absorption ratio index; m-MTVI1 = modified improved triangular vegetation index; m-TVI = modified triangular vegetation index; m-NDVI = modified normalized difference vegetation index; m-OSAVI = modified optimized soil adjusted vegetation index; m-RVI = modified ratio index; m-SAVI = modified soil adjusted vegetation index; m-NRI = modified nitrogen ratio index; m-EVI2 = modified enhanced vegetation index without a blue band.

2.3.2. Generating Reflectance-LUTs and VI-LUTs

Since sensitivity analysis can identify sensitive inputs of a model and evaluate their sensitivity levels, some unessential inputs can be set as constant values for decreasing the dimension of LUTs. For simplifying the LAI retrieval to fulfill the requirement of practical application, we generated two-dimensional (2D) LUTs for LAI retrieval, as previous studies have done [12], which set chlorophyll content and LAI as variables while other parameters in PROSAIL model as constants.
The ranges of the model inputs were determined for wheat according to ground measurements, previous studies [23], and the LOPEX’93 database (http://opticleaf.ipgp.fr/index.php?page=database, accessed on 29 November 2018) (Value (LUT), Table 6). We generated two types of LUTs: the reflectance-LUTs and VI-LUTs, which retrieved the LAI through spectral reflectance and VIs, respectively.

2.3.3. Retrieving LAI through Cost Functions

The cost function was used to find the optimal LAI estimate
f RMSE   =   1 p i = 1 p ( R m R s ) 2
where p represents the number of input variables (the number of bands for reflectance-LUTs/p = 1 for VI-LUTs). Rm represents the measured spectral reflectance/VIs, and Rs represents the simulated reflectance/VIs from the PROSAIL model. When the cost-function value reaches the minimum in a given reflectance-LUT/VI-LUT, the simulated LAI is considered to be the optimal estimate.
In this section, both of the multispectral and hyperspectral UAV data were used for LAI retrieval; the multispectral data were used for investigating the robustness of LAI retrieval strategies. However, the hyperspectral data were further used for improving LAI retrieval accuracy and robustness. For clear statements, the optimization of LAI retrieval is explained in detail in Section 2.3.4.

2.3.4. Optimizing LAI Retrieval Using Hyperspectral Datasets

The hyperspectral UAV data were used for optimizing LAI retrieval. The central wavelengths and bandwidths of the R and NIR bands for LAI retrieval were assessed to optimize the construction of four UAV-based VIs (m-NDVI, m-MCARI2, m-ARVI, and m-NRI selected by global sensitivity analysis; see Section 3.1); the central wavelengths of the B, G, and E bands were set to constant values of 475, 560, and 717 nm, respectively, according to the specifications of the Micasense RedEdge-M (Table 5).
For evaluating the effects of the central wavelengths and bandwidth on LAI retrieval, four simulated datasets were generated as Table 8 shows. The bandwidths of Datasets 1 and 2 were broad; those of Datasets 3 and 4 were narrow (4 nm). The central wavelengths of R and NIR bands were optimized in Datasets 2 and 4.
The optimal central wavelengths of the NIR and R bands were determined via three steps. First, an autocorrelation analysis was performed between any two hyperspectral bands among 88 bands (600–950 nm, Cubert S185). We set p > 0.05 to find uncorrelated bands since the p value of correlated variables was < 0.05. Second, four selected m-VIs (m-NDVI, m-MCARI2, m-ARVI, and m-NRI; see Section 3.1) were calculated from the combinations of reflectance data determined above. Third, the optimal central wavelengths were determined by the highest Pearson correlation coefficient (r values) between the calculated VIs and the measured LAI.
Then, for evaluating the effects of the bandwidth on the LAI retrieval, the narrow bandwidths of the hyperspectral data were resampled into broad bandwidths (Table 8). Because the spectral resolution of the hyperspectral data (Cubert S185) was relatively coarse (4 nm), we only compared the LAI retrieval accuracy of VIs with a narrow bandwidth (4 nm) and broad bandwidths (equal to those of the RedEdge-M camera; see Table 5).

2.4. Statistical Analysis

Python 2.7 was used for statistical analyses, including four statistical indicators for evaluation of the LAI retrieval accuracy: the Pearson correlation coefficient (r), coefficient of determination (R2), root-mean-square error (RMSE), and mean relative error (MRE). The equations for calculating the R2, RMSE, and MRE were calculated according to
R 2 = { [ i = 1 n ( M i M ¯ ) ( E i E ¯ ) ] / [ i = 1 n ( M i M ¯ ) 2 i = 1 n ( E i E ¯ ) 2 ] } 2
RMSE = i = 1 n ( M i E i ) 2 / n
MRE = [ i = 1 n ( | M i E i | / M i ) ] / n
where i represents the sequence number of the array, Mi represents the LAI value of the measured LAI array, Ei represents the LAI value of the estimated LAI array, M ¯ represents the average value of the measured LAI array, and E ¯ represents the average value of the estimated LAI array.

3. Results

3.1. Optimal VIs Selected through Global Sensitivity Analyses

Results of the global sensitivity analyses indicate that the B, R, and NIR bands were more sensitive to LAI, whereas the G and E bands were more sensitive to Chl (Figure 5a). Additionally, E and NIR bands were quite sensitive to the average leaf angle (ALA), with values of 0.26 and 0.19, respectively; B and R were quite sensitive to soil-brightness parameter (psoil), with values of 0.35 and 0.27, respectively; G and E were partially sensitive to leaf-structure parameter (N), with values of 0.17 and 0.11, respectively. Compared with the multispectral bands, the average sensitivity values of the VIs to psoil and N decreased to 0.02 and 0.03, respectively. Thus, the LAI, Chl, and ALA were the three most sensitive PROSAIL inputs to the 39 VIs, with average sensitivity values of 0.72, 0.21, and 0.10, respectively. The average sensitivity values of the R-VIs to the LAI, Chl, and ALA were 0.86, 0.03, and 0.13, respectively. Those of the E-VIs were 0.60, 0.39, and 0.06, respectively, and those of the G-VIs were 0.71, 0.22, and 0.10, respectively. The foregoing results indicate that the R-VIs were the most sensitive to the LAI, the least sensitive to the Chl. Therefore, R-VIs were selected for LAI retrieval.
Within the R-VI group, the total sensitivity values of the following six R-VIs to the LAI reached 0.90: nitrogen ratio index in the R-VI group (R-NRI, 0.93), normalized difference vegetation index in the R-VI group (R-NDVI, 0.93), wide dynamic range vegetation index in the R-VI group (R-WDRVI, 0.91), modified improved chlorophyll absorption ratio index in the R-VI group (R-MCARI2, 0.91), atmospherically resistant vegetation index in the R-VI group (R-ARVI, 0.91), and optimized soil adjusted vegetation index in the R-VI group (R-OSAVI, 0.90). R-NDVI, R-WDRVI, and R-OSAVI were calculated using only the R and NIR bands, while R-NDVI was more commonly used and more sensitive to the LAI than R-WDRVI and R-OSAVI. Therefore, R-NDVI, R-NRI, R-MCARI2, and R-ARVI were selected for LAI retrieval. For simplicity, R-NDVI, R-NRI, R-MCARI2, and R-ARVI are denoted as NDVI, NRI, MCARI2, and ARVI, respectively.

3.2. LAI Retrieval Based on Two LUT Strategies Using Multispectral UAV Data

For assessment of the LAI retrieval performance of two LUT strategies, two-year UAV multispectral observations and were used. For reflectance-LUT, the R2 values were 0.42 and close to zero in 2018 and 2019, respectively, remarkably lower than those for the VIs-LUT (R2 > 0.74) (Table 9). The RMSE and MRE for the reflectance-LUT (RMSE ≥ 0.94, MRE ≥ 0.70 in 2018; RMSE ≥ 2.23, MRE ≥ 2.71 in 2019) were higher than those for the VIs-LUT (RMSE ≤ 0.51, MRE ≤ 0.30 in 2018; RMSE ≤ 0.47, MRE ≤ 0.31 in 2019). The foregoing results indicate that the VI-LUT strategy was more robust and accurate than the reflectance-LUT strategy, which offers an alternative method for LAI retrieval.
The LAI retrieval from four VI-LUTs corresponded to ground-based measurements (Figure 6 and Table 9). The R2, RMSE, and MRE values for LAI retrieval ranged between 0.74–0.83, 0.33–0.51, and 0.22–0.31, respectively. Overall, the LAI estimation accuracy was similar among the four VI-LUTs, although the NDVI-LUT and MCARI2-LUT performed slightly better. In 2018, the R2 values of the MCARI2-LUT and NDVI-LUT were close to 0.75, and the RMSE and MRE of MCARI2-LUT were 0.38 and 0.22, respectively (the lowest among the four VI-LUTs), which was consistent with the comparison between the regression line and the 1:1 line (Figure 6c). In 2019, the R2 value of the MCARI2-LUT was 0.83, higher than the other three VI-LUTs. The RMSE and MRE values of MCARI2-LUT were 0.33 and 0.30, respectively, which were the lowest among the four VI-LUTs. For ARVI-LUT and NRI-LUT, LAI values were slightly underestimated both in 2018 and in 2019. The foregoing results indicate that the VI-LUTs had different characteristics of LAI estimation.

3.3. Optimization of VI-LUTs for LAI Retrieval Using Hyperspectral UAV Data

3.3.1. Optimization of Central Wavelengths for VI Calculation

Figure 7a shows the results of the autocorrelation analyses (Pearson correlation analysis between any two bands). Figure 7b is the uncorrelated band ranges under the conditions of r < 0.20 and p > 0.05 (n = 89, see Table 4). As the wavelength that located in NIR ranges increases, the signal-noise ratio gradually becomes low and the reflectance will be more easily affected by the absorption of water [50]. Thus, the band ranges over 850 nm were excluded in this study. We selected two appropriate regions: 740–760 nm (ρ1 bands) and 600–723 nm (ρ2 bands), along with 740–850 nm (ρ1 bands) and 703–725 nm (ρ2 bands).
Then, m-NDVI, m-MCARI2, m-ARVI, and m-NRI were calculated using the combinations of ρ1 bands and ρ2 bands within the above-determined ranges. The central wavelengths of the ρ1 and ρ2 bands were refined via Pearson correlation analysis between the calculated VIs and the measured LAI (Figure 8). The specific bands of ρ1 and ρ2 were determined for four VIs by identifying the highest r value from the correlation plots in Figure 8. As shown in Table 10, m-NDVI, m-MCARI2, and m-ARVI exhibited significantly positive correlations with the measured LAI (r > 0.86, p < 0.01, n = 89), while m-NRI was significantly negatively correlated with the measured LAI (rm-NRI = −0.86, p < 0.01, n = 89). Correspondingly, the optimal central wavelengths of m-NDVI, m-ARVI, and m-NRI were all located at 752 nm (ρ1 band) and 672 nm (ρ2 band), while the values for m-MCARI2 were located at 756 nm (ρ1 band) and 612 nm (ρ2 band).

3.3.2. LAI Retrieval Based on Two LUT Strategies Using Hyperspectral UAV Data

A comparison between the measured and simulated LAI indicates that the VI-LUT strategy with UAV hyperspectral observations was more robust than the reflectance-LUT strategy (Table 11). The R2 values for the reflectance-LUTs (≤0.32) were significantly lower than those for the VIs-LUTs (≥0.75), and the RMSE and MRE values for the reflectance-LUTs (RMSE > 2.57, MRE ≥ 2.22) were higher than those for the VIs-LUTs (RMSE ≤ 0.66, MRE ≤ 0.40). Furthermore, among the four proposed VIs, the MCARI2-LUT was the most robust and accurate strategy for wheat LAI retrieval (R2 = 0.82, RMSE = 0.37, and MRE = 0.26).

3.3.3. Evaluation of Optimized VI-LUTs Using Hyperspectral Data for LAI Retrieval

Firstly, we compared datasets with the same bandwidths to reveal the effects of central wavelengths on LAI retrieval. For datasets with a broad bandwidth (Datasets 1 and 2, the dots; see Figure 9), the LAI retrieval robustness for the MCARI2-LUT was improved slightly after optimizing the central wavelengths of the VIs, as its RMSE decreased by 0.22. However, those of ARVI-LUT did not. For datasets with a narrow bandwidth (Datasets 3 and 4, the crosses), after optimizing central wavelengths, R2 increased obviously by 0.12, 0.10, 0.12, and 0.09 for the NDVI-LUT, ARVI-LUT, MCARI2-LUT, and NRI-LUT, respectively. RMSE decreased by 0.04 and 0.06, and MRE decreased by 0.03 and 0.06 for NDVI-LUT and MCARI2-LUT, respectively. Although RMSE of ARVI-LUT and NRI-LUT slightly increased by 0.02 and 0.01, respectively, overall, optimizing the central wavelengths of datasets with narrow bandwidths could improve the robustness of VI-LUTs for LAI retrieval.
Secondly, we compared datasets with the same central wavelengths to show the effects of bandwidth on LAI retrieval. For datasets with the optimized central wavelengths (Datasets 2 and 4, in red, see Figure 9), overall, the LAI retrieval for datasets with narrow bandwidths was more robust than datasets with broad bandwidths. However, for VIs without optimization of the central wavelengths (Datasets 1 and 3, in black; see Figure 9), overall, the effects of bandwidth on LAI retrieval remained unclear according to the results of this study. These foregoing results indicate that both of the central wavelength and bandwidth affected LAI retrieval.

4. Discussion

Our research indicates that low-dimensional VI-LUTs along with UAV multispectral images were feasible for wheat LAI retrieval, and more robust than reflectance-LUTs. The MCARI2-LUT and NDVI-LUT are recommended for LAI retrieval owing to their better performance. Furthermore, the central wavelengths and bandwidths of a commonly used multispectral camera (Micasense RedEdge-M), can be modified for specific applications. Regarding the LAI retrieval accuracy with two-year multispectral UAV images, the R2 values of the VI-LUT strategy were >0.74. The R2 values of the reflectance-LUT strategy were 0.42 in 2018 and close to zero in 2019. In wheat LAI retrieval with four simulated datasets, a dataset with central wavelengths of 756 nm for the NIR band and 612 nm for the R band and a narrow bandwidth (~4 nm) performed the most robustly; the R2 value of the MCARI2-LUT increased by 0.06, and the RMSE and MRE decreased by 0.18 and 0.10, respectively.

4.1. Analyses of LAI Retrieval Performance for Reflectance-LUTs and VI-LUTs

Results of the global sensitivity analyses for PROSAIL model indicate that the LAI was the dominant factor affecting B, R, and NIR bands and VIs, in agreement with previous studies [43,51]. The Chl and ALA were two additional variables that the spectral bands and VIs were sensitive to (see Figure 5). In this study, the VIs were divided into three groups to identify the optimal VI group for LAI retrieval. The results indicate that the R-VIs were the most sensitive to the LAI and could reduce the influences of other PROSAIL variables. The E-VIs and G-VIs were sensitive to the Chl but less influenced by the ALA. This study selected the R-VIs for LAI retrieval. However, E-VIs have also been widely applied in LAI retrieval owing to their capabilities of reducing the impact of the ALA of different crops [17]. But the interactions between crop biophysical (e.g., LAI, ALA) and biochemical parameters (e.g., Chl) may introduce uncertainties in parameter retrieval [28]. Additionally, the red-edge bandwidth is narrow, with a low signal-to-noise ratio and uncertainties in its spectral response [51]. Therefore, in UAV remote-sensing, replacing the traditional red bands (ρ2 = ρR) with red-edge bands (ρ2 = ρE) for VIs calculation on LAI retrieval must be considered in future studies.
The results based on two-year multispectral data indicate that the VI-LUT strategy was able to accurately estimate the crop LAI, and performed more robustly than the reflectance-LUT. In this study, UAV multispectral images were acquired during the wheat grain filling period in different years, which might be an effective way to demonstrate the robustness of LUT method for LAI retrieval. We speculate that because the VIs were expressed as a nonlinear mathematical combination of different bands (Table 7), the differences among sensed objects identified only by G, B, or R reflectance observations would be increased by VI calculations [52]. Previous studies have suggested that VIs can provide information on vegetation phenotyping and reduce some interferences from the soil, atmosphere, and shadows [17,51]. Accordingly, the VI-LUT performed more accurately and robustly than the reflectance-LUT for LAI retrieval. However, the reflectance-LUT strategy is also suitable for LAI retrieval when the spectral bands are not limited [11,18]. Most of the multispectral cameras applied in agriculture, such as the multiSPEC-4C and Micasense RedEdge-M cameras which are commonly used, only include three to six bands; thus, they do not provide sufficient information on spectral reflectance or reduce influences of environmental noises. Therefore, the VI-LUT strategy is a fine alternative method for LAI retrieval and is particularly recommended for precision-agriculture applications with multispectral cameras having limited bands.

4.2. Analyses of LAI Retrieval Performance for Different VI-LUTs

In this study, the differences in LAI retrieval among the four VI-LUTs were not obvious. This is because most of the bands used for the VI calculation were the same, leading to a strong correlation between the VIs. However, the VI-LUTs had specific characteristics regarding LAI estimation. The MCARI2-LUT and NDVI-LUT performed slightly better than the other VI-LUTs in this study (Table 9 and Table 11). We cannot explain the superiority of NDVI and MCARI2 for LAI retrieval based on the results of this study. However, previous studies have indicated that MCARI2 involving the G band is sensitive to vegetation coverage and types, as the G band can indicate canopy chlorophyll content that is strongly related to the LAI [53,54]. Therefore, MCARI2 is a robust index for LAI estimation [48]. Moreover, in order to make the issue—the better performance of NDVI and MCARI2 for LAI retrieval—more convincing, future studies can further investigate the different formulas and the characteristics of different bands for the calculations of VIs.
Moreover, some VIs, such as NDVI, tend to be saturated for high-density vegetation [14,55]. However, as shown in Figure 6, the saturation problem of VIs was not significant in this study. We speculate that the foregoing results were related to the LAI measurements. In this study, the LAI values were relatively low, compared with previous studies in which the LAI was measured using an LAI-2200, LAI-2000, or digital hemispherical photography (DHP) [56,57]. The LAI measured using an LAI-2200, LAI-2000, and DHP included leaves and other green parts such as stems and ears, whereas the LAI measured using an LI-3000C only includes leaves (no other green parts). We compared the LAI values for the LI-3000C and LAI-2200. The results indicated that the LAI measured using the LAI-2200 was higher than using the LI-3000C. The R2 value of the regression model was 0.82, indicating that the LAI measurements in this study were feasible. In this study, if the LAI was measured using an LAI-2200, VIs may have underestimated the LAI (with an increased RMSE and MRE); thus, the saturation problem of VIs might occur when the canopy achieves middle-to-high coverage.

4.3. Other Issues Regarding LAI Retrieval Accuracy

Both the VIs and LAI could characterize the functional traits of vegetation canopies, but the relationships between the LAI and VIs were nonlinear and varied between the different vegetation types [58]. The LAI retrieval accuracy for maize was lower than that for wheat for both two-year multispectral observations (please see Appendix A Figure A1 and Appendix A Table A1) and one-year hyperspectral observations (please see Appendix A Table A2). The main reason for this might be the different canopy structures. The remotely sensed LAI implicitly refers to green parts that could be either leaves or other green elements of a vegetation canopy [51]. Thus, the LAI is sensitive to plant structure, such as the clump level, leaf angle, and crop height [59,60]. The PROSAIL model was successfully used for various vegetation types. It treats the canopy as a collection of absorbing and scattering tissues randomly distributed in a horizontal layer. Hence, the PROSAIL model is recommended for application to homogenous crops [11,61]. Both maize and wheat are row crops. When they are at the early stage of growth, they have incomplete coverage and strong leaf clumping, and the background reflectance dominates the spectral signal and affects LAI retrieval [43]. At the middle and later growth stages, their canopies tend to be homogenous. Therefore, the PROSAIL retrieval performance was limited for hyperspectral observations of maize at the early stage of growth (R2 ≤ 0.39) (please see Appendix A Table A2), in agreement with previous studies [62]. Furthermore, the biophysical and biochemical characteristics of vegetation change during different stages. In this study, we did not focus on the effects of phenology on the LAI retrieval. However, continuous UAV observations of critical crop growth periods are required for further validations of the robustness and accuracy of VI-LUTs for LAI retrieval.
Statistical regression models (e.g., a simple linear regression model between vegetation parameters and VIs) and machine-learning models (e.g., random forest models and support vector models) have been widely employed for the retrieval of crop biophysical and biochemical parameters. However, one of the main drawbacks is that their applications are valid only in the areas for which they have been calibrated [63]. The PROSAIL model, as one of the radiative transfer models, can overcome this disadvantage and does not require field measurements in conjunction with remote observations for calibration and validation. Because various combinations of canopy parameters may yield similar spectral reflectance, the ill-posed retrieval problem of the PROSAIL model hinders LAI retrieval. Therefore, a priori information, i.e., ranges of model inputs for model parameterization, is essential. The a priori information is related to many factors, such as the crop types, canopy structure, and growth period of the crops. Thus, the low-dimensional VI-LUT should be modified according to the specific vegetation and growth periods.

4.4. Analyses of Optimization for LUT Strategies

In this study, the VI formula structure may also have affected VI calculation in LAI retrieval. Additionally, the central wavelength and bandwidth of VIs affected LAI retrieval. LAI retrieval with four simulated datasets indicate that the VI-LUTs with optimized narrow bands (4 nm) had the highest accuracy. However, the effects of different bandwidths on the accuracy of crop parameter estimation remain unclear. Some hyperspectral narrow bands are sensitive to certain crop parameters, but they might be insensitive to other crop parameters, which may increase the computational loads and distort the accuracy of LAI retrieval [41]. Observational data with broad bands can yield strong reflectance signals, which may reduce the influence of the surrounding environment. Therefore, when selecting the bandwidth in practice, it is essential to achieve a tradeoff between the retrieval sensitivity and the signal-noise ratio.
In summary, this study demonstrates that the VI-LUT strategy based on the PROSAIL model is an alternative method for LAI retrieval with higher accuracy and robustness than the reflectance-LUT strategy. It is valuable for monitoring the crop growth status in precision-agriculture owing to its advantages of simplicity, robustness, accuracy, and saving time and labor. Moreover, the optimization of the central wavelengths and bandwidths of VIs may be helpful for designing multispectral cameras according to specific applications.

5. Conclusions

The objective of this study was to develop a crop LAI retrieval strategy with accuracy, robustness, and ease of use for UAV remote sensing applications in precision-agriculture at the field scale, which would assist the design of UAV multispectral cameras for agronomic monitoring. The study was conducted using UAV remotely sensed images (two-year multispectral data and one-year hyperspectral data) and the PROSAIL model simulation approach. It was concluded that the low-dimensional VI-LUTs were easy to use and more robust (R2 ≥ 0.74, RMSE ≤ 0.51, MRE ≤ 0.31 for multispectral datasets) than reflectance-LUTs (R2 ≤ 0.42, RMSE ≥ 0.94, MRE ≥ 0.70 for multispectral datasets) in crop LAI retrieval based on UAV data. The differences in LAI retrieval among four VI-LUTs were not obvious. However, the MCARI2-LUT and NDVI-LUT performed slightly better than the other VI-LUTs and thus are recommended for crop LAI retrieval. Moreover, both the central wavelengths and bandwidths of the VIs affected LAI retrieval. The hyperspectral UAV data with central wavelengths of 756 nm for the NIR band and 612 nm for the R band and a narrow bandwidth (~4 nm) improved the performance of the MCARI2-LUT for LAI retrieval.
The VI-LUT strategy based on the PROSAIL model, as an alternative method for LAI retrieval at the field scale, is recommended for crop growth monitoring with multispectral cameras having limited bands. The optimized central wavelengths and bandwidths of VIs, as well as corresponding methods of VI optimization, might contribute to improving the design of multispectral cameras onboard UAVs for retrieving vegetation traits in precision-agriculture.

Author Contributions

Z.S., Y.H. and X.L. acquired the funding for UAV flight missions and ground measurements. W.Z., Z.S. and Y.H. developed the workflow for this study. W.Z., B.Y. and J.Z. conducted UAV flight missions and processed UAV data. J.L. (Jing Li), B.L. and S.L. developed a part of experimental plots with different treatments. W.Z., J.L. (Jianbin Lai), K.Z., Y.L. and B.L. measured LAI in situ. W.Z. and Z.S. analyzed the data and wrote the paper.

Funding

This study was supported by National Natural Science Foundation of China (31570472, 31870421, 41771388), Key Projects of the Chinese Academy of Sciences (KFZD-SW-319), the Science and Technology Service Network Initiative of the Chinese Academy of Sciences (KFJ-STS-ZDTP-049), the Strategic Priority Research Program of the Chinese Academy of Sciences (XDA19040303), and the National Key Research and Development Program of China (2017YFC0503805).

Acknowledgments

Thanks for assistance provided by Guicang Ma, Xuezhi Yue, Zhenmin Liu, Ruiqiang Wu, and Wei Jiang for flight tests, and Yinxia Zhu for assistance with LAI measurements in this study.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Figure A1. Comparisons of two-year measured LAI and estimated LAI values for wheat and maize based on four VIs-LUTs. (a) NDVI-LUT, (b) MCARI2-LUT, and (c) NRI-LUT in 2016 (multispectral data captured from multiSPEC-4C camera); (d) NDVI-LUT, (e) MCARI2-LUT, (f) NRI-LUT, and (g) ARVI-LUT in 2018 (multispectral data captured from Micasense RedEdge-M camera). R12 represents the coefficient of determination for a linear regression model between the measured and estimated LAI for wheat. R22 represents the coefficient of determination for a linear regression model between the measured and estimated LAI for maize.
Figure A1. Comparisons of two-year measured LAI and estimated LAI values for wheat and maize based on four VIs-LUTs. (a) NDVI-LUT, (b) MCARI2-LUT, and (c) NRI-LUT in 2016 (multispectral data captured from multiSPEC-4C camera); (d) NDVI-LUT, (e) MCARI2-LUT, (f) NRI-LUT, and (g) ARVI-LUT in 2018 (multispectral data captured from Micasense RedEdge-M camera). R12 represents the coefficient of determination for a linear regression model between the measured and estimated LAI for wheat. R22 represents the coefficient of determination for a linear regression model between the measured and estimated LAI for maize.
Remotesensing 11 02456 g0a1
Table A1. R2, RMSE, and MRE for linear regression between simulated LAI and estimated LAI based on reflectance-LUTs/VI-LUTs using multispectral UAV data in 2016 and 2018.
Table A1. R2, RMSE, and MRE for linear regression between simulated LAI and estimated LAI based on reflectance-LUTs/VI-LUTs using multispectral UAV data in 2016 and 2018.
CropsYearReflectance-LUTsVIs-LUTs
fRMSE InputR2RMSEMREfRMSE InputR2RMSEMRE
Wheat2016R,NIR0.412.491.23NDVI0.700.470.16
n = 72R,NIR,G0.412.491.23MCARI0.750.830.34
R,NIR,E0.412.491.23NRI0.741.040.40
2018R,NIR0.420.940.70NDVI0.760.440.25
n = 107R,NIR,B0.420.940.70ARVI0.740.510.30
R,NIR,G0.420.940.70MCARI0.750.380.22
R,NIR, E0.420.940.70NRI0.750.460.25
Maize2016R,NIR0.001.480.42NDVI0.680.730.17
n = 40R,NIR,G0.001.480.42MCARI0.730.580.16
R,NIR,E0.001.480.42NRI0.611.560.40
2018R,NIR0.002.210.87NDVI0.610.620.18
n = 113R,NIR,B0.002.210.87ARVI0.590.660.19
R,NIR,G0.002.210.87MCARI0.711.190.42
R,NIR, E0.002.210.87NRI0.620.660.20
Table A2. LAI retrieval accuracy (R2, RMSE, and MRE) based on reflectance-LUTs/VIs-LUTs using hyperspectral UAV data in 2018 (n = 89).
Table A2. LAI retrieval accuracy (R2, RMSE, and MRE) based on reflectance-LUTs/VIs-LUTs using hyperspectral UAV data in 2018 (n = 89).
Crops Reflectance-LUTs VIs-LUTs
BandsR2RMSEMREVIsR2RMSEMRE
WheatR, NIR0.272.5762.22NDVI0.800.550.31
R, NIR, B0.272.5762.22ARVI0.760.660.40
R, NIR, G0.272.5792.22MCARI20.820.370.26
R, NIR, E0.322.6032.24NRI0.750.580.32
MaizeR, NIR0.072.5992.56NDVI0.220.750.36
R, NIR, B0.122.5502.50ARVI0.280.900.45
R, NIR, G0.122.5502.50MCARI20.390.440.17
R, NIR, E0.112.5472.50NRI0.330.780.37

References

  1. Swain, K.C. Suitability of low-altitude remote sensing images for estimating nitrogen treatment variations in rice cropping for precision agriculture adoption. J. Appl. Remote Sens. 2007, 1, 013547. [Google Scholar] [CrossRef]
  2. Zhang, C.; Kovacs, J.M. The application of small unmanned aerial systems for precision agriculture: A review. Precis. Agric. 2012, 13, 693–712. [Google Scholar] [CrossRef]
  3. Soudani, K.; François, C.; Le Maire, G.; Le Dantec, V.; Dufrêne, E. Comparative analysis of IKONOS, SPOT, and ETM+ data for leaf area index estimation in temperate coniferous and deciduous forest stands. Remote Sens. Environ. 2006, 102, 161–175. [Google Scholar] [CrossRef] [Green Version]
  4. Ni, J.; Yao, L.; Zhang, J.; Cao, W.; Zhu, Y.; Tai, X. Development of an unmanned aerial vehicle-borne crop-growth monitoring system. Sensors 2017, 17, 502. [Google Scholar] [CrossRef] [PubMed]
  5. Potgieter, A.B.; George-Jaeggli, B.; Chapman, S.C.; Laws, K.; Suárez Cadavid, L.A.; Wixted, J.; Watson, J.; Eldridge, M.; Jordan, D.R.; Hammer, G.L. Multi-spectral imaging from an unmanned aerial vehicle enables the assessment of seasonal leaf area dynamics of sorghum breeding lines. Front. Plant Sci. 2017, 8, 1532. [Google Scholar] [CrossRef] [PubMed]
  6. Mueller, N.D.; Gerber, J.S.; Johnston, M.; Ray, D.K.; Ramankutty, N.; Foley, J.A. Closing yield gaps through nutrient and water management. Nature 2012, 490, 254–257. [Google Scholar] [CrossRef]
  7. Duveiller, G.; Weiss, M.; Baret, F.; Defourny, P. Retrieving wheat Green Area Index during the growing season from optical time series measurements based on neural network radiative transfer inversion. Remote Sens. Environ. 2011, 115, 887–896. [Google Scholar] [CrossRef]
  8. Geipel, J.; Link, J.; Claupein, W. Combined spectral and spatial modeling of corn yield based on aerial images and crop surface models acquired with an unmanned aircraft system. Remote Sens. 2014, 6, 10335–10355. [Google Scholar] [CrossRef]
  9. Yue, J.; Yang, G.; Li, C.; Li, Z.; Wang, Y.; Feng, H.; Xu, B. Estimation of winter wheat above-ground biomass using unmanned aerial vehicle-based snapshot hyperspectral sensor and crop height improved models. Remote Sens. 2017, 9, 708. [Google Scholar] [CrossRef]
  10. Walter, A.; Finger, R.; Huber, R.; Buchmann, N. Opinion: Smart farming is key to developing sustainable agriculture. Proc. Natl. Acad. Sci. USA 2017, 114, 6148–6150. [Google Scholar] [CrossRef]
  11. Duan, S.-B.; Li, Z.-L.; Wu, H.; Tang, B.-H.; Ma, L.; Zhao, E.; Li, C. Inversion of the PROSAIL model to estimate leaf area index of maize, potato, and sunflower fields from unmanned aerial vehicle hyperspectral data. Int. J. Appl. Earth Obs. Geoinf. 2014, 26, 12–20. [Google Scholar] [CrossRef]
  12. Roosjen, P.P.J.; Brede, B.; Suomalainen, J.M.; Bartholomeus, H.M.; Kooistra, L.; Clevers, J.G.P.W. Improved estimation of leaf area index and leaf chlorophyll content of a potato crop using multi-angle spectral data—Potential of unmanned aerial vehicle imagery. Int. J. Appl. Earth Obs. Geoinf. 2018, 66, 14–26. [Google Scholar] [CrossRef]
  13. Locherer, M.; Hank, T.; Danner, M.; Mauser, W. Retrieval of seasonal Leaf Area Index from simulated EnMAP data through optimized LUT-Based inversion of the PROSAIL model. Remote Sens. 2015, 7, 10321–10346. [Google Scholar] [CrossRef]
  14. Yao, X.; Wang, N.; Liu, Y.; Cheng, T.; Tian, Y.; Chen, Q.; Zhu, Y. Estimation of wheat LAI at middle to high levels using unmanned aerial vehicle narrowband multispectral imagery. Remote Sens. 2017, 9, 1304. [Google Scholar] [CrossRef]
  15. Xu, X.Q.; Lu, J.S.; Zhang, N.; Yang, T.C.; He, J.Y.; Yao, X.; Cheng, T.; Zhu, Y.; Cao, W.X.; Tian, Y.C. Inversion of rice canopy chlorophyll content and leaf area index based on coupling of radiative transfer and Bayesian network models. ISPRS J. Photogramm. Remote Sens. 2019, 150, 185–196. [Google Scholar] [CrossRef]
  16. Wang, Y.; Zhang, K.; Tang, C.; Cao, Q.; Tian, Y.; Zhu, Y.; Cao, W.; Liu, X. Estimation of rice growth parameters based on linear mixed-effect model using multispectral images from Fixed-Wing Unmanned Aerial Vehicles. Remote Sens. 2019, 11, 1371. [Google Scholar] [CrossRef]
  17. Dong, T.; Liu, J.; Shang, J.; Qian, B.; Ma, B.; Kovacs, J.M.; Walters, D.; Jiao, X.; Geng, X.; Shi, Y. Assessment of red-edge vegetation indices for crop leaf area index estimation. Remote Sens. Environ. 2019, 222, 133–143. [Google Scholar] [CrossRef]
  18. Zhao, J.; Li, J.; Liu, Q.; Wang, H.; Chen, C.; Xu, B.; Wu, S. Comparative analysis of Chinese HJ-1 CCD, GF-1 WFV and ZY-3 MUX sensor data for leaf area index estimations for maize. Remote Sens. 2018, 10, 68. [Google Scholar] [CrossRef]
  19. Ding, Y.; Zhang, H.; Li, Z.; Xin, X.; Zheng, X.; Zhao, K. Comparison of fractional vegetation cover estimations using dimidiate pixel models and look- up table inversions of the PROSAIL model from Landsat 8 OLI data. J. Appl. Remote Sens. 2016, 10, 036022. [Google Scholar] [CrossRef]
  20. Mridha, N.; Sahoo, R.N.; Sehgal, V.K.; Krishna, G.; Pargal, S.; Pradhan, S.; Gupta, V.K.; Kumar, D.N. Comparative evaluation of inversion approaches of the radiative transfer model for estimation of crop biophysical parameters. Int. Agrophys. 2015, 29, 201–212. [Google Scholar] [CrossRef]
  21. Su, W.; Huang, J.; Liu, D.; Zhang, M. Retrieving Corn Canopy Leaf Area Index from Multitemporal Landsat Imagery and Terrestrial LiDAR Data. Remote Sens. 2019, 11, 572. [Google Scholar] [CrossRef]
  22. Darvishzadeh, R.; Wang, T.; Skidmore, A.; Vrieling, A.; O’Connor, B.; Gara, T.; Ens, B.; Paganini, M. Analysis of Sentinel-2 and RapidEye for Retrieval of Leaf Area Index in a Saltmarsh Using a Radiative Transfer Model. Remote Sens. 2019, 11, 671. [Google Scholar] [CrossRef]
  23. Richter, K.; Hank, T.B.; Vuolo, F.; Mauser, W.; D’Urso, G. Optimal exploitation of the Sentinel-2 spectral capabilities for crop leaf area index mapping. Remote Sens. 2012, 4, 561–582. [Google Scholar] [CrossRef]
  24. Xie, Q.; Dash, J.; Huete, A.; Jiang, A.; Yin, G.; Ding, Y.; Peng, D.; Hall, C.C.; Brown, L.; Shi, Y.; et al. Retrieval of crop biophysical parameters from Sentinel-2 remote sensing imagery. Int. J. Appl. Earth Obs. Geoinf. 2019, 80, 187–195. [Google Scholar] [CrossRef]
  25. Danner, M.; Berger, K.; Wocher, M.; Mauser, W.; Hank, T. Retrieval of biophysical crop variables from multi-angular canopy spectroscopy. Remote Sens. 2017, 9, 726. [Google Scholar] [CrossRef]
  26. Fei, Y.; Jiulin, S.; Hongliang, F.; Zuofang, Y.; Jiahua, Z.; Yunqiang, Z.; Kaishan, S.; Zongming, W.; Maogui, H. Comparison of different methods for corn LAI estimation over northeastern China. Int. J. Appl. Earth Obs. Geoinf. 2012, 18, 462–471. [Google Scholar] [CrossRef]
  27. Nigam, R.; Bhattacharya, B.K.; Vyas, S.; Oza, M.P. Retrieval of wheat leaf area index from AWiFS multispectral data using canopy radiative transfer simulation. Int. J. Appl. Earth Obs. Geoinf. 2014, 32, 173–185. [Google Scholar] [CrossRef]
  28. Jay, S.; Maupas, F.; Bendoula, R.; Gorretta, N. Retrieving LAI, chlorophyll and nitrogen contents in sugar beet crops from multi-angular optical remote sensing: Comparison of vegetation indices and PROSAIL inversion for field phenotyping. Field Crops Res. 2017, 210, 33–46. [Google Scholar] [CrossRef] [Green Version]
  29. Tripathi, R.; Sahoo, R.N.; Sehgal, V.K.; Tomar, R.K.; Chakraborty, D.; Nagarajan, S. Inversion of PROSAIL model for retrieval of plant biophysical parameters. J. Indian Soc. Remote Sens. 2012, 40, 19–28. [Google Scholar] [CrossRef]
  30. Wang, S.; Gao, W.; Ming, J.; Li, L.; Xu, D.; Liu, S.; Lu, J. A TPE based inversion of PROSAIL for estimating canopy biophysical and biochemical variables of oilseed rape. Comput. Electron. Agric. 2018, 152, 350–362. [Google Scholar] [CrossRef]
  31. Lin, J.; Pan, Y.; Lyu, H.; Zhu, X.; Li, X.; Dong, B.; Li, H. Developing a two-step algorithm to estimate the leaf area index of forests with complex structures based on CHRIS/PROBA data. For. Ecol. Manag. 2019, 441, 57–70. [Google Scholar] [CrossRef]
  32. Quan, X.; He, B.; Li, X.; Liao, Z. Retrieval of grassland live fuel moisture content by parameterizing radiative transfer model with interval estimated LAI. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2016, 9, 910–920. [Google Scholar] [CrossRef]
  33. Teillet, P. Effects of spectral, spatial, and radiometric characteristics on remote sensing vegetation indices of forested regions. Remote Sens. Environ. 1997, 61, 139–149. [Google Scholar] [CrossRef]
  34. Steven, M.D.; Malthus, T.J.; Baret, F.; Xu, H.; Chopping, M.J. Intercalibration of vegetation indices from different sensor systems. Remote Sens. Environ. 2003, 88, 412–422. [Google Scholar] [CrossRef]
  35. Jiang, Z.; Huete, A.; Didan, K.; Miura, T. Development of a two-band enhanced vegetation index without a blue band. Remote Sens. Environ. 2008, 112, 3833–3845. [Google Scholar] [CrossRef]
  36. Kaufman, Y.J.; Tanre, D. Atmospherically resistant vegetation index (ARVI) for EOS-MODIS. IEEE Trans. Geosci. Remote Sens. 1992, 30, 261–270. [Google Scholar] [CrossRef]
  37. Qi, J.; Chehbouni, A.; Huete, A.R.; Kerr, Y.H.; Sorooshian, S. A modified soil adjusted vegetation index. Remote Sens. Environ. 1994, 48, 119–126. [Google Scholar] [CrossRef]
  38. Xue, J.; Su, B. Significant remote sensing vegetation indices: A review of developments and applications. J. Sens. 2017, 2017, 1–17. [Google Scholar] [CrossRef]
  39. Chen, H.; Huang, W.; Li, W.; Niu, Z.; Zhang, L.; Xing, S. Estimation of LAI in winter wheat from multi-angular hyperspectral VNIR data: Effects of view angles and plant architecture. Remote Sens. 2018, 10, 1630. [Google Scholar] [CrossRef]
  40. Broge, N.H.; Leblanc, E. Comparing prediction power and stability of broadband and hyperspectral vegetation indices for estimation of green leaf area index and canopy chlorophyll density. Remote Sens. Environ. 2001, 76, 156–172. [Google Scholar] [CrossRef]
  41. Liu, K.; Zhou, Q.; Wu, W.; Xia, T.; Tang, H. Estimating the crop leaf area index using hyperspectral remote sensing. J. Integr. Agric. 2016, 15, 475–491. [Google Scholar] [CrossRef] [Green Version]
  42. Nilson, T.; Kuusk, A. A reflectance model for the homogeneous plant canopy and its inversion. Remote Sens. Environ. 1989, 27, 157–167. [Google Scholar] [CrossRef]
  43. Berger, K.; Atzberger, C.; Danner, M.; D’Urso, G.; Mauser, W.; Vuolo, F.; Hank, T. Evaluation of the PROSAIL model capabilities for future hyperspectral model environments: A review study. Remote Sens. 2018, 10, 85. [Google Scholar] [CrossRef]
  44. Zhang, L.; Guo, C.L.; Zhao, L.Y.; Zhu, Y.; Cao, W.X.; Tian, Y.C.; Cheng, T.; Wang, X. Estimating wheat yield by integrating the WheatGrow and PROSAIL models. Field Crops Res. 2016, 192, 55–66. [Google Scholar] [CrossRef]
  45. Saltelli, A.; Tarantola, S.; Chan, K.P.-S. A quantitative model-independent method for global sensitivity analysis of model output. Technometrics 1999, 41, 39–56. [Google Scholar] [CrossRef]
  46. Zou, X.; Mottus, M. Sensitivity of Common Vegetation Indices to the Canopy Structure of Field Crops. Remote Sens. 2017, 9, 994. [Google Scholar] [CrossRef]
  47. Chen, J.M. Valuation of vegetation indices and a modified simple ratio for boreal applications. Can. J. Remote Sens. 1996, 22, 229–242. [Google Scholar] [CrossRef]
  48. Haboudane, D. Hyperspectral vegetation indices and novel algorithms for predicting green LAI of crop canopies: Modeling and validation in the context of precision agriculture. Remote Sens. Environ. 2004, 90, 337–352. [Google Scholar] [CrossRef]
  49. Schleicher, T.D.; Bausch, W.C.; Delgado, J.A.; Ayers, P.D. Evaluation and refinement of the nitrogen reflectance index (NRI) for site-specific fertilizer management. In Proceedings of the American Society of Agricultural and Biological Engineers, Sacramento, CA, USA, 29 July–1 August 2001. [Google Scholar]
  50. Gao, B. NDWI—A normalized difference water index for remote sensing of vegetation liquid water from space. Remote Sens. Environ. 1996, 58, 257–266. [Google Scholar] [CrossRef]
  51. Verger, A.; Vigneau, N.; Cheron, C.; Gilliot, J.-M.; Comar, A.; Baret, F. Green area index from an unmanned aerial system over wheat and rapeseed crops. Remote Sens. Environ. 2014, 152, 654–664. [Google Scholar] [CrossRef]
  52. Burkart, A.; Aasen, H.; Alonso, L.; Menz, G.; Bareth, G.; Rascher, U. Angular dependency of hyperspectral measurements over wheat characterized by a novel UAV based goniometer. Remote Sens. 2015, 7, 725–746. [Google Scholar] [CrossRef]
  53. Gitelson, A.A. Remote estimation of canopy chlorophyll content in crops. Geophys. Res. Lett. 2005, 32. [Google Scholar] [CrossRef] [Green Version]
  54. Hunt, E.R.; Horneck, D.A.; Spinelli, C.B.; Turner, R.W.; Bruce, A.E.; Gadler, D.J.; Brungardt, J.J.; Hamm, P.B. Monitoring nitrogen status of potatoes using small unmanned aerial vehicles. Precis. Agric. 2018, 19, 314–333. [Google Scholar] [CrossRef]
  55. Gitelson, A.A. Wide dynamic range vegetation index for remote quantification of biophysical characteristics of vegetation. J. Plant Physiol. 2004, 161, 165–173. [Google Scholar] [CrossRef]
  56. Fang, H.; Ye, Y.; Liu, W.; Wei, S.; Ma, L. Continuous estimation of canopy leaf area index (LAI) and clumping index over broadleaf crop fields: An investigation of the PASTIS-57 instrument and smartphone applications. Agric. For. Meteorol. 2018, 253–254, 48–61. [Google Scholar] [CrossRef]
  57. Luo, S.; Wang, C.; Xi, X.; Nie, S.; Fan, X.; Chen, H.; Yang, X.; Peng, D.; Lin, Y.; Zhou, G. Combining hyperspectral imagery and LiDAR pseudo-waveform for predicting crop LAI, canopy height and above-ground biomass. Ecol. Indic. 2019, 102, 801–812. [Google Scholar] [CrossRef]
  58. Kross, A.; McNairn, H.; Lapen, D.; Sunohara, M.; Champagne, C. Assessment of RapidEye vegetation indices for estimation of leaf area index and biomass in corn and soybean crops. Int. J. Appl. Earth Obs. Geoinf. 2015, 34, 235–248. [Google Scholar] [CrossRef] [Green Version]
  59. Chianucci, F.; Disperati, L.; Guzzi, D.; Bianchini, D.; Nardino, V.; Lastri, C.; Rindinella, A.; Corona, P. Estimation of canopy attributes in beech forests using true colour digital images from a small fixed-wing UAV. Int. J. Appl. Earth Obs. Geoinf. 2016, 47, 60–68. [Google Scholar] [CrossRef] [Green Version]
  60. Pisek, J.; Buddenbaum, H.; Camacho, F.; Hill, J.; Jensen, J.L.R.; Lange, H.; Liu, Z.; Piayda, A.; Qu, Y.; Roupsard, O.; et al. Data synergy between leaf area index and clumping index Earth Observation products using photon recollision probability theory. Remote Sens. Environ. 2018, 215, 1–6. [Google Scholar] [CrossRef]
  61. Darvishzadeh, R.; Matkan, A.A.; Ahangar, A.D. Inversion of a radiative transfer model for estimation of rice canopy chlorophyll content using a lookup-table approach. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2012, 5, 1222–1230. [Google Scholar] [CrossRef]
  62. Richter, K.; Atzberger, C.; Vuolo, F.; D’Urso, G. Evaluation of sentinel-2 spectral sampling for radiative transfer model based LAI estimation of wheat, sugar beet, and maize. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2011, 4, 458–464. [Google Scholar] [CrossRef]
  63. Becker-Reshef, I.; Vermote, E.; Lindeman, M.; Justice, C. A generalized regression-based model for forecasting winter wheat yields in Kansas and Ukraine using MODIS data. Remote Sens. Environ. 2010, 114, 1312–1323. [Google Scholar] [CrossRef]
Figure 1. Location of Yucheng Comprehensive Experiment Station (YCES) and overview of the long-term experimental plots. (The yellow rectangles are the boundaries of plots. The red rectangles are flight coverage of the UAV hyperspectral mission; two UAV multispectral missions covered all six fields. The red spots are ground sampling plots for Fields A, C, and F in 2018).
Figure 1. Location of Yucheng Comprehensive Experiment Station (YCES) and overview of the long-term experimental plots. (The yellow rectangles are the boundaries of plots. The red rectangles are flight coverage of the UAV hyperspectral mission; two UAV multispectral missions covered all six fields. The red spots are ground sampling plots for Fields A, C, and F in 2018).
Remotesensing 11 02456 g001
Figure 2. Wheat in experimental plots and wheat leaf destructive sampling for LAI measurements on 15 May 2018: (a) N0-80%fc in Field D, (b) N140-80%fc in Field, (c) N280-80%fc in Field D, and (d) Destructive sampling.
Figure 2. Wheat in experimental plots and wheat leaf destructive sampling for LAI measurements on 15 May 2018: (a) N0-80%fc in Field D, (b) N140-80%fc in Field, (c) N280-80%fc in Field D, and (d) Destructive sampling.
Remotesensing 11 02456 g002
Figure 3. UAV platforms used in this study: (a) the DJ M600 Pro six-rotor UAV; (b) the DJ M100 four-rotor UAV.
Figure 3. UAV platforms used in this study: (a) the DJ M600 Pro six-rotor UAV; (b) the DJ M100 four-rotor UAV.
Remotesensing 11 02456 g003
Figure 4. Flowchart of crop LAI retrieval using multispectral and hyperspectral UAV datasets. ARVI = atmospherically resistant vegetation index; MCARI2 = modified improved chlorophyll absorption ratio index; NDVI = modified normalized difference vegetation index; NRI = nitrogen ratio index; NIR = near-infrared band; R = red band; fRMSE = cost function.
Figure 4. Flowchart of crop LAI retrieval using multispectral and hyperspectral UAV datasets. ARVI = atmospherically resistant vegetation index; MCARI2 = modified improved chlorophyll absorption ratio index; NDVI = modified normalized difference vegetation index; NRI = nitrogen ratio index; NIR = near-infrared band; R = red band; fRMSE = cost function.
Remotesensing 11 02456 g004
Figure 5. Results of global sensitivity analyses for the (a) five-band reflectance, (b) R-VIs, (c) G-VIs, and (d) E-VIs to PROSAIL inputs using the EFAST method.
Figure 5. Results of global sensitivity analyses for the (a) five-band reflectance, (b) R-VIs, (c) G-VIs, and (d) E-VIs to PROSAIL inputs using the EFAST method.
Remotesensing 11 02456 g005
Figure 6. Comparisons of two-year measured LAI and simulated wheat LAI values based on four VIs-LUTs (a) NDVI-LUT in 2018, (b) ARVI-LUT in 2018, (c) MCARI2-LUT in 2018, and (d) NRI-LUT in 2018; (e) NDVI-LUT in 2019, (f) ARVI-LUT in 2019, (g) MCARI2-LUT in 2019, and (h) NRI-LUT in 2019. R2 represents the coefficient of determination for a linear regression model between the measured LAI and simulated LAI.
Figure 6. Comparisons of two-year measured LAI and simulated wheat LAI values based on four VIs-LUTs (a) NDVI-LUT in 2018, (b) ARVI-LUT in 2018, (c) MCARI2-LUT in 2018, and (d) NRI-LUT in 2018; (e) NDVI-LUT in 2019, (f) ARVI-LUT in 2019, (g) MCARI2-LUT in 2019, and (h) NRI-LUT in 2019. R2 represents the coefficient of determination for a linear regression model between the measured LAI and simulated LAI.
Remotesensing 11 02456 g006
Figure 7. Autocorrelation analyses between any ρ1 and ρ2 bands (600–950 nm). (a) the r values; (b) the p values (> 0.05).
Figure 7. Autocorrelation analyses between any ρ1 and ρ2 bands (600–950 nm). (a) the r values; (b) the p values (> 0.05).
Remotesensing 11 02456 g007
Figure 8. Pearson correlation coefficients (r values) between the simulated LAI and m-VIs derived from the refined ρ1 and ρ2. (a) m-NDVI, (b) m-NRI, (c) m-MCARI2, and (d) m-ARVI.
Figure 8. Pearson correlation coefficients (r values) between the simulated LAI and m-VIs derived from the refined ρ1 and ρ2. (a) m-NDVI, (b) m-NRI, (c) m-MCARI2, and (d) m-ARVI.
Remotesensing 11 02456 g008
Figure 9. (a) R2, (b) RMSE, and (c) MRE for linear regression between measured LAI and simulated LAI of four datasets based on VI-LUTs (Datasets with optimized central wavelengths are in red: Datasets 2 and 4; datasets without optimizing central wavelengths are in black: Datasets 1 and 3. Datasets with broad bandwidths are denoted as dots: Datasets 1 and 2; datasets with narrow bandwidths are denoted as the crosses: Datasets 3 and 4).
Figure 9. (a) R2, (b) RMSE, and (c) MRE for linear regression between measured LAI and simulated LAI of four datasets based on VI-LUTs (Datasets with optimized central wavelengths are in red: Datasets 2 and 4; datasets without optimizing central wavelengths are in black: Datasets 1 and 3. Datasets with broad bandwidths are denoted as dots: Datasets 1 and 2; datasets with narrow bandwidths are denoted as the crosses: Datasets 3 and 4).
Remotesensing 11 02456 g009
Table 2. Treatments of the six experimental fields.
Table 2. Treatments of the six experimental fields.
Field (Plot Number, Plot Size)Treatments 1 (Plot Number)
Field A (16, 2×2 m2)CT-RR 2 (3), NT-RR (3), CT-S (3), NT-S (4), CT-1/2S (3)
Field B (25, 5×6 m2)CK (4), NP (4), NK (4), PK (4), NPK (7), NPK-S (2)
Field C (6, 10×10 m2)CT (6)
Field D (32, 5×10 m2)N0-60%fc (3), N70-60%fc (3), N140-60%fc (3), N210-60%fc (3) N280-60%fc (4), N0-80%fc (3), N70-80%fc (3), N140-80%fc (3) N210-80%fc(3), N280-80%fc(4)
Field E (32, 5×10 m2)0%fc (8), 40%fc (8), 60%fc (8), 80%fc (8)
Field F (18, 8×40 m2)NT-S-F1 (3), NT-S-F2 (3), NT-RR-F1 (3), NT-RR-F2 (3), CT-S-F1 (3), CT-S-F2 (3)
1 Tillage treatments: see Table 3; 2 The symbol “-” represents different treatments are implemented together in the plot.
Table 3. Abbreviations of agronomic treatments in Table 2.
Table 3. Abbreviations of agronomic treatments in Table 2.
Abbr.Definitions
Tillage treatmentCTconventional tillage
NTnon-tillage
Residual treatmentsRRresidual removal
Sreturning the whole straw of each plot to soil
1/2S returning half straw of each plot to soil
Fertilizer applicationsCKno N, P, K fertilizer
Nnitrogenous fertilizer
Kpotash fertilizer
Pphosphate fertilizer
N0, N70, N140, N210, N280N7070 kg N ha−1 for each crop season
0%fc, 40%fc, 60%fc, 80%fc60%fcirrigation to 60% of the field water capacity
Fertilization methodsF1only base fertilizer application
F2base fertilizer along with two topdressings
Table 4. Statistical summary of the in situ measured LAI (m2/m2).
Table 4. Statistical summary of the in situ measured LAI (m2/m2).
CamerasDateField (Samples)MinMeanMaxStandard DeviationCV (%) *
Multi-15 May 2018A (4)1.651.781.850.105.40
B (25)0.511.272.300.5946.34
C (6)2.082.593.050.3714.18
D (32)0.311.833.150.7742.20
E (32) 0.631.371.850.3022.16
F (8)1.782.062.500.2511.92
Total (107)0.311.623.150.6540.03
16 May 2019B (25)0.211.252.560.7963.30
D (32)0.461.512.930.7851.73
Total (57)0.211.402.930.8157.70
Hyper-15 May 2018B (25)0.511.272.300.5946.34
D (32)0.311.833.150.7742.20
E (32)0.631.371.850.3022.16
Total (89)0.311.513.150.6341.85
* CV represents the coefficient of variation.
Table 5. Specifications of the two cameras used in this study.
Table 5. Specifications of the two cameras used in this study.
SensorSpectral Channels (Central Wavelength/Spectral Ranges)Spatial ResolutionSpectral Resolution (nm)
Cubert S185450–950 nm1 cm4 nm
RedEdge-MB475, G560, R668, E717, NIR8404 cmB (20), G (20), R (10), E (10), NIR (40)
Table 6. Ranges of input variables for the PROSAIL model.
Table 6. Ranges of input variables for the PROSAIL model.
VariableAbbr.UnitValue (LUT) *Range (EFAST) **
Leaf structure parameterNUnitless1.51–2
Leaf chlorophyll contentChlμg·cm−220–70 (step = 0.2)20–70
Leaf carotenoid contentcaroμg·cm−2103–30
Brown pigment content-arbitrary units00
Blade equivalent thicknessEWTcm0.010.005–0.03
Leaf water mass per areaLMAg·cm−20.0050.004–0.007
Soil brightness parameterpsoilUnitless0.10.01–0.3
Leaf area indexLAIm2 m−20.1–6 (step = 0.01)0.1–6
Hot-spot size parameterhot spotm m−10.20.05-1
Solar zenith angle-degrees2020
Solar azimuth angle-degrees185185
View zenith angle-degrees00
View azimuth angle-degrees00
Average leaf angleALAdegrees7030–70
* Value (LUT) is used for generating lookup tables in Section 2.3.2; ** Range (Fourier amplitude sensitivity test (EFAST)) was used for running the global sensitivity analyses of PROSAIL model in Section 2.3.1.
Table 8. Four UAV datasets used for LAI retrieval.
Table 8. Four UAV datasets used for LAI retrieval.
DatasetValue of Central Wavelengths (Bandwidth) (nm)
BGRENIR
Dataset 1475/20560/20668/10717/10840/40
Dataset 2475/20560/20672 or 612/10 *717/10752 or 756/10 **
Dataset 3475/4560/4668/4717/4840/4
Dataset 4475/4560/4672 or 612/10 *717/4752 or 756/10 **
* The central wavelengths of R bands were 672 nm for m-NDVI, m-NRI, and m-ARVI, and 612 nm for m-MCARI2. Please see Table 10 for the selection of optimal central wavelengths. ** The central wavelengths of NIR bands were 752 nm for m-NDVI, m-NRI, and m-ARVI, and 756 nm for m-MCARI2. Please see Table 10 for the selection of optimal central wavelengths.
Table 9. R2, RMSE, and MRE for linear regression between measured LAI and simulated LAI based on reflectance-LUTs/VI-LUTs using multispectral UAV data.
Table 9. R2, RMSE, and MRE for linear regression between measured LAI and simulated LAI based on reflectance-LUTs/VI-LUTs using multispectral UAV data.
YearReflectance-LUTsVI-LUTs
BandsR2RMSEMREVIsR2RMSEMRE
2018 n = 107R,NIR0.420.940.70NDVI0.760.440.25
R,NIR,B0.420.940.70ARVI0.740.510.30
R,NIR,G0.420.940.70MCARI20.750.380.22
R,NIR, E0.420.940.70NRI0.750.460.25
2019 n = 57R,NIR0.012.232.71NDVI0.780.380.27
R,NIR,B0.012.232.71ARVI0.740.470.23
R,NIR,G0.012.232.71MCARI20.830.330.30
R,NIR, E 0.012.232.71NRI0.740.430.31
R2 represents the coefficient of determination for a linear regression model between the measured LAI and simulated LAI. RMSE = root-mean-square error, MRE = mean relative error.
Table 10. Specific optimal bands of ρ1 and ρ2 for the four m-VI calculations.
Table 10. Specific optimal bands of ρ1 and ρ2 for the four m-VI calculations.
CropsVIsρ1 Bands (nm)ρ2 Bands (nm)r Value
Wheatm-NDVI7526720.86 **
m-NRI752672−0.86 **
m-MCARI27566120.87 **
m-ARVI7526720.86 **
** indicates p < 0.01.
Table 11. R2, RMSE, and MRE for linear regression between simulated LAI and estimated LAI based on reflectance-LUTs/VI-LUTs using hyperspectral UAV data in 2018 (n = 89).
Table 11. R2, RMSE, and MRE for linear regression between simulated LAI and estimated LAI based on reflectance-LUTs/VI-LUTs using hyperspectral UAV data in 2018 (n = 89).
Reflectance-LUTsVIs-LUTs
BandsR2RMSEMREVIsR2RMSEMRE
R, NIR0.272.582.22NDVI0.800.550.31
R, NIR, B0.272.582.22ARVI0.760.660.40
R, NIR, G0.272.582.22MCARI20.820.370.26
R, NIR, E0.322.602.24NRI0.750.580.32

Share and Cite

MDPI and ACS Style

Zhu, W.; Sun, Z.; Huang, Y.; Lai, J.; Li, J.; Zhang, J.; Yang, B.; Li, B.; Li, S.; Zhu, K.; et al. Improving Field-Scale Wheat LAI Retrieval Based on UAV Remote-Sensing Observations and Optimized VI-LUTs. Remote Sens. 2019, 11, 2456. https://0-doi-org.brum.beds.ac.uk/10.3390/rs11202456

AMA Style

Zhu W, Sun Z, Huang Y, Lai J, Li J, Zhang J, Yang B, Li B, Li S, Zhu K, et al. Improving Field-Scale Wheat LAI Retrieval Based on UAV Remote-Sensing Observations and Optimized VI-LUTs. Remote Sensing. 2019; 11(20):2456. https://0-doi-org.brum.beds.ac.uk/10.3390/rs11202456

Chicago/Turabian Style

Zhu, Wanxue, Zhigang Sun, Yaohuan Huang, Jianbin Lai, Jing Li, Junqiang Zhang, Bin Yang, Binbin Li, Shiji Li, Kangying Zhu, and et al. 2019. "Improving Field-Scale Wheat LAI Retrieval Based on UAV Remote-Sensing Observations and Optimized VI-LUTs" Remote Sensing 11, no. 20: 2456. https://0-doi-org.brum.beds.ac.uk/10.3390/rs11202456

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop