The document describes two research activities (RAs) on county-level corn yield prediction using deep learning methods. RA-1 uses a Bayesian neural network to provide both yield predictions and uncertainty estimates, outperforming other models. RA-2 addresses the lack of transferability of models across regions through unsupervised domain adaptation strategies like adversarial domain adaptation networks, which improve predictions in data-scarce regions without labeled data. The proposed adaptive and Bayesian domain adaptation networks achieve better spatial transferability than compared methods.
Beyond the EU: DORA and NIS 2 Directive's Global Impact
County-Level Corn Yield Prediction with GeoAI.pdf
1. County-Level Corn Yield Prediction with
Deep Learning
Yuchi Ma, Ph.D.
Earth Systems Science
Stanford University
August 2023
2. Outline
• Introduction
• Materials
• Research Activities (RAs)
• Bayesian Neural Network for Corn Yield Prediction and Uncertainty
Analysis
• Adversarial Domain Adaptation on Corn Yield Prediction
2
3. Introduction – Background
Credit: UN, USDA
3
• Corn
• The U.S. is the biggest corn producer and exporter.
• Corn is the most valuable crop in America.
• Corn plays a significant role in American diet and food
industries.
• Corn Yield Prediction
• Food security monitoring
• Food market planning
• Farm resource management
• Risk assessment
4. Introduction – Background
Credit: USDA
4
• United States Department of Agriculture National Agricultural
Statistics Service (USDA-NASS)
• Publish monthly corn yield prediction at the state level from August to November
through the NASS Crop Production Progress Report.
• Publish corn yield statistics at the county level in February of the next year through
the NASS Quick Stats.
• Limitations:
• Large-scale field surveys are required, which is labor-intensive and time-
consuming
• Low spatial and temporal resolution
• Cannot meet the needs for in-season county-level yield prediction
5. Introduction – Remote Sensing + Machine Learning
5
Machine learning
Model
Ridge Regression, Random
Forests, Neural Networks …
Multi-Source Input
Predictors
Vegetation Indices, Weather
variables …
Predicted Yield
Traditional Machine Learning (ML)
• Associate input predictors X with response variable y
• No need for explicit programming or knowledges for
physiological mechanisms on individual crops.
• Ridge Regression, Random Forest (RF), etc.
Deep Learning (DL)
• Simulate neurons in human brain
• Automatic feature learning
• Multiple Layer Perceptron (MLP), Recurrent
Neural Network (RNN), etc.
6. Introduction – Challenges
# How to quantify predictive uncertainty and increase robustness for
supervised deep learning models on corn yield prediction?
6
Bottlenecks of point-estimate neural networks:
➢ A huge amount of training data is necessary
➢ Prone to overfitting
➢ Lack of uncertainty information
Point estimates
7. Introduction – Challenges
# How to improve DL models’ spatial transferability through unsupervised
domain adaptation strategies?
7
Bottlenecks of supervised machine learning:
➢ Untrainable for regions without reference yield records
➢ DL models are region-specific (or domain-specific)
➢ Lack of transferability due to domain shift
Data-abundant
source region/domain
Data-scarce
target region/domain
Domain Shift
8. Materials – Study Area
• U.S. Corn Belt
• The top corn-producing region
• Abundant county-level yield records
• Yield records from USDA NASS
8
9. Materials – Feature Variables
9
Category Variables Unit
Related
properties
Spatial
Resolution
Temporal
Resolution
Source Latency
Satellite
Imagery
Enhanced Vegetation Index (EVI)
Plant vigor 500 m Daily
MODIS One day
Green Chlorophyll Index (GCI)
Normalized Difference Water Index (NDWI)
Daytime Land Surface Temperature (LSTday)
Kelvin
Heat stress
1 km Daily
Nighttime Land Surface Temperature
(LSTnight)
Climate
Mean Temperature (Tmean)
°C
4 km Daily PRISM One day
Maximum Temperature (Tmax)
Minimum Temperature (Tmin)
Total Precipitation (PPT) (mm) mm
Water stress
Maximum Vapor Pressure Deficit (VPDmax)
hPa
Minimum Vapor Pressure Deficit (VPDmin)
GLDAS Water Stress (GLDASws) 0.25 arc
degree
3-hourly GLDAS
One
month
Evapotranspiration (ET) mm
Soil
Available Water Holding Capacity (AWC) cm
Soil water
uptake
30 m N/A SSURGO N/A
Soil Organic Matter (SOM) kg/m2 Soil nutrient
uptake
Cation Exchange Capacity (CEC) cmol/kg
Others
Year
County-level N/A
USDA
NASS
N/A
Historical average yield t/ha
10. Materials – Data Preprocessing
10
Start
MODIS PRISM GLDAS
…
Spatial Filtering
Spatial Aggregation
Temporal Aggregation
Pair with Yield Records
End
• Spatial Filtering
• Filtering out feature variables on non-corn areas
• Spatial Aggregation
• Aggregate feature variables into the county level
by calculating the mean value in each county
• Make sure all feature variables have the same
spatial resolution
• Temporal Aggregation
• Aggregate time-series feature variables into the
16-day interval by calculating the mean value in
each 16-day time period.
11. RA-1: Methodology
11
Traditional Neural Networks:
• Point estimates of weights
• Output only predicted values
• Prone to overfitting
Bayesian Neural Networks:
• Estimate posterior distributions of weights
• Output both target values and uncertainty
• A special form of ensemble learning →
Robust to overfitting
Traditional vs Bayesian Neural Network
Point estimates Posterior estimates
12. RA-1: Methodology
The proposed BNN model:
• A feature extraction net to extract features from input predictors
• Two independent nets, the yield net and the uncertainty net, to predict the mean ො
𝑦
and the standard deviation ො
𝜎 for the predictive yield distribution.
• The uncertainty net outputs ො
𝜎 as a measurement of predictive uncertainty.
12
Ma et al. 2021a
13. RA-1: Experimental Setup
• Experimental years: 2001-2019
• Testing years: 2010-2019
• Training strategy: using one year for testing and all preceding years for training
• Comparison methods:
• Ridge Linear Regression (Ridge)
• Random Forest (RF)
• Support Vector Regressor (SVR)
• Multilayer Perceptron (MLP)
• Recurrent Neural Network with Long Short-Term Memory (LSTM) cell
• Evaluation:
• End-of-season: using all non-sequential and sequential predictors and making
predictions on Oct 4th
• In-season: using all non-sequential and part of sequential predictors and
making predictions during the growing season 13
15. RA-1: Evaluation Results
In-season evaluation:
• Poor performance during the early season (i.e., before July)
• Performance increases as more informative features included
• Optimal performance achieved in August, which is two month before the harvest
15
Ma et al. 2021a
18. RA-1: Conclusion
RA-1: Bayesian Neural Network for Corn Yield Prediction and
Uncertainty Analysis
• The proposed BNN model outperforms other ML&DL models.
• Accurate corn yield prediction can be made in August, which is two
month ahead of the harvest.
• The predictive uncertainty have strong correlation with the prediction
error.
18
19. RA-2: Background
19
Bottlenecks of supervised ML&DL:
• Labeled data are required for model training → impossible to train
models for regions without yield records.
• Due to domain shift, supervised learning models tend to be location-
specific → models trained in the label-rich region (source domain)
would lose the validity when directly applied to the label-scarce region
(target domain).
20. RA-2: Background
20
Fine-tuning-based transfer learning:
• Idea: Pretrain a source model with abundant labeled source samples;
Fine-tune the source model with a few labeled target samples.
• Issue: A limited number of labeled target samples are still needed.
21. RA-2: Methodology
21
Unsupervised Domain Adaptation (UDA):
• Idea: Reduce domain shift by aligning feature distributions in the
source domain and the target domain.
• Advantage: No need of labeled samples from the target domain.
Source Domain Target Domain
UDA
Source Domain
Predictor
Cross-Domain
Predictor
Misprediction
Misprediction
UDA
Ma and Zhang. 2022
:High yield samples
:Low yield samples
22. RA-2: Methodology
Domain Adversarial Neural Network (DANN)
• Feature Extractor 𝐺𝑓: extract task-informative and domain-invariant features
• Yield Predictor 𝐺𝑦: make yield predictions
• Domain Classifier 𝐺𝑑: classify the domain labels
• Gradient Reversal Layer (GRL): Reverse the gradient during backpropagation
22
Ma et al. 2021b
23. RA-2: Methodology
• Training objectives:
1. Collaboratively train the feature extractor 𝐺𝑓 and the yield predictor 𝐺𝑦 to
minimize the yield prediction loss 𝐿𝑦 → extract task-informative features
(Eq. (2.2))
2. Adversarially train the feature extractor 𝐺𝑓 and the domain classifier 𝐺𝑑 via
GRL to maximize the domain loss 𝐿𝑑 → extract domain-invariant features
(Eq. (2.3))
• Loss function:
𝐿 = 𝐿𝑦 − 𝜆𝐿𝑑 (2.1)
𝐿𝑦 =
1
𝑛𝑆
σ𝑖=1
𝑛𝑆
𝑦𝑖 − ො
𝑦𝑖
2 (2.2)
𝐿𝑑 = −
1
𝑛𝑆+𝑛𝑇
σ𝑖=1
𝑛𝑆+𝑛𝑇
𝑑𝑖 log መ
𝑑𝑖 + (1 − 𝑑𝑖) log 1 − መ
𝑑𝑖 (2.3)
23
𝜆: weighting parameter
24. RA-2: Methodology
24
Adaptative Domain Adversarial Neural Network (ADANN)
• Issue: the yield prediction loss (in the form of MSE) can have a quite different
magnitude with the domain classification loss (in the form of Cross Entropy).
• Solution: the weighting parameter 𝜆𝑖 is adaptatively determined according to the
learning progress 𝑝𝑖 and the normalization term 𝑟𝑖:
𝑝𝑖 =
𝑖
𝑁
(2.4)
𝑟𝑖 =
𝐿𝑦𝑖
𝐿𝑑𝑖
(2.5)
𝜆𝑖 = 𝑟𝑖(
2
1+exp −𝑝𝑖
− 1) (2.6)
25. RA-2: Methodology
25
Bayesian Domain Adversarial Neural Network (BDANN)
• Issue: Many source samples is required to train a reliable ADANN.
• Solution: Applying Bayesian Inference and proposed BDANN with the purpose to
improve the model performance on small training datasets.
26. RA-2: Experimental Setup
Study area - Two ecological regions based on Environmental
Protection Agency (EPA):
• Eastern Temperate Forests (ETF)
• Warm, humid, and temperate
• Mild winters and humid summers
• High plant biodiversity
• Great Plains (GP)
• Semiarid with high winds
• Harsh winters and hot summers
• Lack of forests and rainfall
26
Ma et al. 2021a
27. RA-2: Experimental Setup
• Experimental years: 2006-2019
• Testing years: 2016-2019
• Training strategy: using one year for testing and all preceding years
for training
• Comparison methods:
• Random Forest (RF)
• Full-connected Deep Neural Network (DNN)
• DANN with a fixed weighting parameter
• Evaluation:
• GP → ETF: models are trained in GP and tested in ETF.
• ETF → GP: models are trained in ETF and tested in GP.
27
29. 0.5 1.0 1.5 2.0
0 250 500
Absolute Error (t/ha)
km
RF
RA-2: Absolute Error Maps
DNN DANN ADANN BDANN
Experiment: GP → ETF
Experiment: ETF → GP
30. RA-2: Conclusion
RA-2: Adversarial Domain Adaptation on Corn Yield Prediction
1. The UDA strategy was used for corn yield prediction.
2. Two adversarial domain adaptation models were proposed for county-level corn
yield prediction.
3. The proposed ADANN and BDANN outperformed RF, DNN, and DANN with
better spatial transferability across spatial domains.
4. The BDANN model could generalize well on small training sets.
30
31. Selected Publications – Related
RA-1:
• Ma, Y., Zhang, Z., Kang, Y. and Özdoğan, M., 2021a. Corn yield prediction and uncertainty analysis based
on remotely sensed variables using a Bayesian neural network approach. Remote Sensing of Environment,
259, p.112408.
• Chen, S., Liu, W., Feng, P., Ye, T., Ma, Y. and Zhang, Z., 2022. Improving Spatial Disaggregation of Crop
Yield by Incorporating Machine Learning with Multisource Data: A Case Study of Chinese Maize Yield.
Remote Sensing, 14(10), p.2340.
RA-2:
• Ma, Y., Zhang, Z., Yang, H.L. and Yang, Z., 2021b. An adaptive adversarial domain adaptation approach for
corn yield prediction. Computers and Electronics in Agriculture, 187, p.106314.
• Ma, Y. and Zhang, Z., 2022. A Bayesian Domain Adversarial Neural Network for Corn Yield Prediction.
IEEE Geoscience and Remote Sensing Letters, 19, pp 1-5.
31