Back to YieldSAT homepage

Complete Benchmark Results

Full field-level and subfield-level benchmarks for CV10, LORO, and LOYO across all country-crop combinations.

This page contains the complete benchmark table for YieldSAT. It includes 10-fold cross-validation (CV10), leave-one-region-out (LORO), and leave-one-year-out (LOYO) evaluation for the temporal, spatial-temporal, and multimodal fusion baselines discussed on the main project page.

On smaller screens, scroll horizontally to inspect the full table.

Experiment Setup Modalities Fusion Method Model Field-Level Subfield   (Pixel)-Level
ARG-C ARG-S ARG-W BRA-C BRA-S BRA-W GER-R GER-W URG-S ARG-C ARG-S ARG-W BRA-C BRA-S BRA-W GER-R GER-W URG-S
R2 RMSE R2 RMSE R2 RMSE R2 RMSE R2 RMSE R2 RMSE R2 RMSE R2 RMSE R2 RMSE R2 RMSE R2 RMSE R2 RMSE R2 RMSE R2 RMSE R2 RMSE R2 RMSE R2 RMSE R2 RMSE
CV10 Sentinel-2 X Transformer[1] 0.771.350.730.630.870.780.790.820.750.350.620.490.750.670.561.26 0.720.560.612.330.620.940.761.170.442.160.380.940.221.390.441.250.322.430.381.24
LSTM[5] 0.791.290.720.640.850.840.750.880.680.390.620.490.620.83-0.872.60 0.660.620.592.400.600.960.741.210.422.200.340.980.221.390.361.33-0.323.380.371.26
3D-ConvLSTM[2] 0.841.130.790.550.920.620.820.740.760.340.730.420.810.580.651.12 0.770.510.652.200.650.900.791.100.452.130.390.940.231.380.491.200.342.400.411.22
3D-LSTM[4] 0.741.440.770.580.890.700.820.740.760.340.630.480.820.570.541.28 0.730.560.592.370.650.900.791.080.462.130.390.930.241.370.481.200.302.460.391.23
Sentinel-2 + ADM Input Fusion Transformer[1] 0.691.550.720.640.830.870.790.810.710.370.610.490.760.650.611.19 0.730.550.522.570.580.980.701.300.442.170.370.960.221.390.451.240.382.320.391.23
LSTM[5] 0.761.390.720.640.850.830.810.780.640.420.580.520.810.580.631.16 0.720.560.582.410.590.980.751.200.432.180.330.990.211.400.471.210.352.380.391.23
3D-ConvLSTM[2] 0.122.630.820.520.890.720.830.720.760.340.720.420.780.630.701.03 0.800.480.083.580.680.870.781.110.462.120.380.940.231.380.421.270.392.300.421.20
3D-LSTM[4] 0.402.170.760.590.870.770.840.710.750.350.720.420.810.590.621.17 0.770.520.362.970.640.910.781.130.462.120.410.930.241.370.491.200.372.340.401.22
Feature Fusion AFF[4] 0.841.120.840.490.920.600.840.700.800.310.710.430.800.600.740.96 0.820.460.702.030.730.790.840.960.462.120.440.900.241.370.491.200.442.200.431.19
MMGF[3] 0.791.300.820.510.890.690.760.860.800.310.590.510.750.680.770.90 0.750.530.652.200.700.840.801.070.422.190.420.920.201.400.441.260.442.210.401.22
LORO Sentinel-2 X Transformer[1] 0.551.880.590.770.711.160.351.430.330.570.370.630.171.220.011.890.630.65 0.432.800.521.050.611.490.272.470.211.070.111.480.091.590.102.790.311.31
LSTM[5] 0.472.050.640.720.701.180.391.390.280.590.270.68-0.161.45-2.713.660.630.65 0.442.790.551.030.611.480.272.460.161.100.111.48-0.051.71-0.783.930.341.28
3D-ConvLSTM[2] 0.681.590.690.670.800.950.551.190.370.550.470.580.251.160.141.76 0.680.610.542.520.571.010.691.330.332.350.191.080.161.440.121.570.072.840.361.26
3D-LSTM[4] 0.492.000.700.670.810.920.591.130.410.530.480.570.261.16-0.081.980.650.63 0.392.910.590.980.711.300.342.330.221.060.201.410.171.520.022.920.361.26
Sentinel-2 + ADM Input Fusion Transformer[1] 0.442.100.670.700.631.300.521.220.200.620.360.640.181.22-0.342.20 0.590.680.382.930.531.040.531.630.312.390.151.110.121.470.131.56-0.022.970.301.31
LSTM[5] 0.432.120.660.710.611.340.381.390.290.580.260.690.201.20-1.392.94 0.560.710.482.690.541.030.531.640.282.450.121.130.111.480.051.63-0.363.440.321.30
3D-ConvLSTM[2] 0.332.300.730.630.810.930.621.090.510.480.490.570.231.180.121.78 0.600.670.213.310.600.970.701.320.342.340.261.040.171.430.071.610.102.790.331.29
3D-LSTM[4] 0.232.460.660.700.741.090.651.050.470.500.480.580.031.32-0.162.05 0.620.660.392.900.571.000.641.440.372.290.261.040.171.430.031.64-0.022.970.351.27
Feature Fusion AFF[4] 0.531.920.780.570.870.780.551.190.630.420.520.550.071.29-1.853.21 0.600.680.492.640.650.900.781.120.342.350.350.970.181.420.151.55-0.583.700.341.28
MMGF[3] 0.591.800.650.710.781.000.101.680.300.580.230.70-0.561.68-0.952.65 0.540.730.502.630.541.040.681.350.132.690.201.070.071.51-0.131.77-0.343.410.311.31
LOYO Sentinel-2 X Transformer[1] 0.571.830.520.840.780.990.191.590.280.59-0.090.830.331.100.141.760.600.68 0.422.830.481.100.671.380.192.600.161.100.051.530.141.550.132.740.331.29
LSTM[5] 0.462.060.500.850.771.020.191.590.060.67-0.050.82-0.071.39-1.743.140.610.67 0.392.900.461.120.611.490.152.660.041.180.081.50-0.051.71-0.683.820.331.29
3D-ConvLSTM[2] 0.611.750.650.710.781.000.201.580.360.55-0.060.820.431.020.231.67 0.630.650.482.680.541.040.671.380.152.670.191.080.061.520.211.490.152.710.341.28
3D-LSTM[4] 0.531.930.660.700.771.020.361.420.120.65-0.330.920.411.030.001.900.620.66 0.472.700.551.030.681.350.212.560.131.120.031.540.181.520.012.930.351.27
Sentinel-2 + ADM Input Fusion Transformer[1] 0.332.300.650.720.751.080.251.54-0.090.720.110.750.470.980.021.880.560.71 0.243.240.491.090.601.510.222.540.081.150.111.480.251.450.062.860.301.32
LSTM[5] 0.422.130.620.740.741.080.291.490.260.590.120.750.580.870.331.560.540.73 0.442.780.491.090.611.490.192.600.171.090.091.500.311.380.182.670.301.32
3D-ConvLSTM[2] -0.012.820.670.690.771.030.451.320.340.56-0.150.860.490.960.461.39 0.560.71-0.023.750.561.010.641.430.272.470.191.080.041.540.221.470.222.600.311.31
3D-LSTM[4] 0.182.540.620.740.780.990.281.500.450.52-0.070.820.261.150.151.750.520.74 0.253.230.531.050.681.340.222.550.221.060.061.520.131.560.152.720.301.32
Feature Fusion AFF[4] 0.332.300.770.590.810.930.421.350.380.540.140.740.441.000.191.71 0.530.730.352.990.660.890.721.270.292.430.291.020.111.480.251.450.162.700.321.30
MMGF[3] 0.312.330.660.700.361.710.391.380.380.55-0.080.830.151.24-0.031.920.480.77 0.303.100.571.000.282.030.232.530.241.050.061.520.051.630.062.860.281.34

ARG-C: Argentina Corn, ARG-S: Argentina Soybean, ARG-W: Argentina Wheat, BRA-C: Brazil Corn, BRA-S: Brazil Soybean, BRA-W: Brazil Wheat, GER-R: Germany Rapeseed, GER-W: Germany Wheat, URG-S: Uruguay Soybean. Blank or X indicates no fusion method is used for Sentinel-2-only models in the corresponding setup. Under Experiment Setup, CV10 refers to 10-fold cross-validation, LORO refers to Leave-One-Region-Out, and LOYO refers to Leave-One-Year-Out. Superscript references next to model names link to the benchmark papers listed below.

Benchmark References

[1] P. Helber et al., “Crop yield prediction: An operational approach to crop yield modeling on field and subfield level with machine learning models,” in IGARSS - IEEE international geoscience and remote sensing symposium, IEEE, 2023, pp. 2763–2766.

[2] P. Helber, B. Bischke, C. Packbier, P. Habelitz, and F. Seefeldt, “An operational approach to large-scale crop yield prediction with spatio-temporal machine learning models,” in IGARSS 2024-2024 IEEE international geoscience and remote sensing symposium, IEEE, 2024, pp. 4299–4302.

[3] F. Mena et al., “Adaptive fusion of multi-modal remote sensing data for optimal sub-field crop yield prediction,” Remote Sensing of Environment, vol. 318, p. 114547, 2025.

[4] M. Miranda, D. Pathak, M. Nuske, and A. Dengel, “Multi-modal fusion methods with local neighborhood information for crop yield prediction at field and subfield levels,” in IGARSS 2024-2024 IEEE international geoscience and remote sensing symposium, IEEE, 2024, pp. 4307–4311.

[5] D. Pathak et al., “Predicting crop yield with machine learning: An extensive analysis of input modalities and models on a field and sub-field level,” in IGARSS- IEEE international geoscience and remote sensing symposium, 2023, pp. 2767–2770. doi: 10.1109/IGARSS52108.2023.10282318.