Articles | Volume 22, issue 15
https://doi.org/10.5194/bg-22-3965-2025
https://doi.org/10.5194/bg-22-3965-2025
Research article
 | 
14 Aug 2025
Research article |  | 14 Aug 2025

On the added value of sequential deep learning for the upscaling of evapotranspiration

Basil Kraft, Jacob A. Nelson, Sophia Walther, Fabian Gans, Ulrich Weber, Gregory Duveiller, Markus Reichstein, Weijie Zhang, Marc Rußwurm, Devis Tuia, Marco Körner, Zayd Hamdi, and Martin Jung
Abstract

Estimating ecosystem–atmosphere fluxes such as evapotranspiration (ET) in a robust manner and at a global scale remains a challenge. Methods based on machine learning (ML) have shown promising results in achieving such upscaling, providing a complementary methodology that is independent from process-based and semi-empirical approaches. However, a systematic evaluation of the skill and robustness of different ML approaches is an active field of research that requires more investigation. Concretely, deep learning approaches in the time domain have not been explored systematically for this task.

In this study, we compared instantaneous (i.e., non-sequential) models (extreme gradient boosting (XGBoost) and a fully connected neural network (FCN)) with sequential models (a long short-term memory (LSTM) model and a temporal convolutional network (TCN)) for the modeling and upscaling of ET. We compared different types of covariates (meteorological without precipitation, precipitation, remote sensing, and plant functional types) and their impact on model performance at the site level in a cross-validation setup.

When using only meteorological covariates, we found that the sequential models (LSTM and TCN) performed better (each with a Nash–Sutcliffe efficiency (NSE) of 0.73) than the instantaneous models (FCN and XGBoost), both with an NSE of 0.70, in site-level cross-validation at the hourly scale. The advantage of the sequential models diminished with the inclusion of remote-sensing-based predictors (NSE of 0.75 to 0.76 versus 0.74). On the anomaly scale, the sequential models consistently outperformed the non-sequential models across covariate setups, with an NSE of 0.36 (LSTM) and 0.38 (TCN) versus 0.33 (FCN) and 0.32 (XGBoost) when using all covariates.

For the upscaling from site to global coverage, we input the two best-performing combinations of covariates – (a) meteorological and remote sensing observations and (b) precipitation and plant functional types in addition – with globally available gridded data. To evaluate and compare the robustness of the modeling approaches, we generated a cross-validation-based ensemble of upscaled ET, compared the ensemble mean and variance among models, and contrasted it with independent global ET data. In particular, we investigate three questions regarding the performance of the sequential models compared to the non-sequential models in the context of spatial upscaling: (a) whether they lead to more realistic and robust global and regional ET, (b) whether they are able to capture the temporal dynamics of ET better, and (c) how robust they are to the covariate setup and training data subsets.

The generated patterns of global ET variability were relatively consistent across the ML models overall, but in regions with low data support via eddy covariance (EC) stations, we observed substantial biases across models and covariate setups and large ensemble uncertainties. The sequential models better capture the temporal dynamics of ET when upscaled to global coverage, especially when using precipitation as additional input, and they seem to be more robust to covariate setups, particularly the LSTM model. However, together with the non-temporal FCN model, they exhibited a larger ensemble spread than XGBoost, and they yielded lower global ET estimates than what is currently understood. XGBoost showed a smaller ensemble spread compared to neural networks, in particular when conditions were poorly represented in the training data, but it was more sensitive to the covariate setup. Plant functional types were useful at the site level for the improved representation of spatial patterns but had a significant leverage on upscaling results – i.e., having a disproportionate impact on the spatial patterns, especially for XGBoost, but less for the LSTM model.

Our findings highlight non-linear model responses to biases in the training data and underscore the need for improved upscaling methodologies, which could be achieved by increasing the amount and quality of training data or by the extraction of more-targeted features representing spatial variability. The neural networks seem to yield more-realistic ensemble uncertainty compared to XGBoost. Approaches such as transfer learning, knowledge-guided ML, or hybrid modeling, which encourage physically consistent results while harnessing the efficiency of ML, should be further investigated. Deep learning for flux upscaling holds great promise, while remedies for its vulnerability to training data distribution changes still need consideration by the community.

Share
1 Introduction

Measurements of land–atmosphere fluxes of gases, such as water vapor or carbon dioxide, are crucial for understanding the interactions between climate and ecosystems. Instruments at eddy covariance (EC) stations measure such fluxes integrated over a time span of 30 or 60 min and a small spatial footprint, spanning a couple of hundred meters to over a kilometer depending on the instrument height, terrain roughness, and wind conditions. The measurement is performed at the ecosystem level, as it represents the integral biotic and abiotic processes across scales (Baldocchi et al.2001). While EC stations provide a crucial source of data to measure these fluxes, they come with challenges. For instance, their representativeness and applicability for regional to global analysis may be restricted due to the sparsity of EC sites in the geographic and climate space (Fig. 1a and b).

https://bg.copernicus.org/articles/22/3965/2025/bg-22-3965-2025-f01

Figure 1Overview of eddy covariance (EC) sites used in this work. (a) Distribution of EC sites (white points) and map grid cells (background color) within the global climate in terms of mean temperature and annual precipitation. (b) Geographic EC site locations in different gray shading according to the number of hourly observations of evapotranspiration. The map color corresponds to the representativeness of a geographic location by the EC station sites. It is the average Euclidean distance in climate space (mean and standard deviation of normalized 15-daily temperature, precipitation, and radiation) to the 10 closest stations. A lower representativeness (red) means a given location is further away from EC sites in climate space.

Evapotranspiration (ET) is the combined flux of water vapor via evaporation from bare surfaces and plant transpiration. The ET flux is of high relevance for modeling and understanding the earth system because it links the water, carbon, and energy cycles (Jung et al.2010; Nelson et al.2018). However, the modeling of ET is challenging due to the highly dynamic nature and modulation of ecosystems. Their behavior depends on past system exposure via so-called dynamic memory effects (Ogle et al.2015; Besnard et al.2019; Kraft et al.2019, 2021). Among other factors, ET depends on soil moisture, which is primarily driven by the recent past rather than by instantaneous weather conditions. Other processes impacting ET that depend on past meteorology are related to vegetation states, such as leaf area and phenology (Migliavacca et al.2012).

To consider such complex memory effects, a model must incorporate past system exposure, such as temperature or precipitation. Alternatively, the model can be fed with states that represent past exposure, such as leaf area index (LAI) and soil moisture observations, or aggregations of past meteorology, like temperature or precipitation sums. However, the observation of ecosystem states is challenging and often not possible. In situ measurements, e.g., of soil moisture, are not consistently available at all EC stations and may not always precisely coincide with the eddy covariance measurements in space or time, limiting the applicability for across-site modeling. As an alternative, remotely sensed observations can serve as proxies of ecosystem states, like vegetation indices for foliage or phenology. These observations alone can only partially explain EC measurements, as they represent structural or optical properties of the canopy rather than plant physiology or subsurface water states; in particular, optical observations tend to saturate with dense vegetation (Huete et al.2002). Therefore, it may be beneficial to learn the non-observable states for the modeling of land–atmosphere fluxes as non-linear functions of available covariates. Here, sequential machine learning (ML) models may offer a unique opportunity, as they are able to extract dynamic proxies from temporal data (Rußwurm and Körner2017; Kraft et al.2019).

ET can be quantified at large scales by employing process-based paradigms, i.e., land surface models, or semi-empirical approaches based on inputs from remote sensing observations and predefined empirical relationships (e.g., the Global Land Evaporation Amsterdam Model (GLEAM); Martens et al.2017). As a complementary approach, the data-driven upscaling, i.e., the generalization from the irregularly distributed EC stations to a regular spatio-temporal field, can provide independent insights into ecosystem processes (Jung et al.2017). The upscaling is achieved by training an ML model at the EC sites with covariates that are also available as spatio-temporal fields (Jung et al.2009). The optimized model is then fed with the contiguous covariates to generate regional- to global-scale products.

Due to the availability of long-term records of both eddy covariance data and remote sensing products, increased computational capacities, and a higher acceptance of ML approaches in the geosciences (Camps-Valls et al.2021), data-driven approaches to model ecosystem–atmosphere fluxes have gained momentum in the past decade (Tramontana et al.2016; Jung et al.2011; Nelson et al.2024; Zhu et al.2024). Today, ML is widely used to model and upscale EC data, but the field is still dominated by non-sequential modeling (i.e., an instantaneous model that does not learn memory effects), such as decision trees or fully connected neural networks.

An ensemble of global harmonized products of upscaled EC fluxes from different ML algorithms (tree, kernel, regression splines, and neural-network-based methods) was released by the FLUXCOM initiative (FLUXCOM2017), founded on previous work by Beer et al. (2010), Jung et al. (2010, 2011), and Tramontana et al. (2016). These products are built upon non-sequential models, and they account for memory via manually designed features, such as seasonal amplitudes or water availability indices, and remote-sensing-based ecosystem state proxies, like vegetation indices (Huete et al.2002). The FLUXCOM products of energy (Jung et al.2019) and carbon (Jung et al.2020) are utilized in contemporary land–atmosphere interaction studies and function as benchmarks for earth system models. To improve the temporal resolution and resolve the diurnal cycle, Bodesheim et al. (2018) upscaled 30 min fluxes of carbon and energy using randomized decision forests (Breiman2001) with a non-sequential modeling approach. Xiao et al. (2014) upscaled daily carbon and water fluxes in North America using moderate imaging spectroradiometer (MODIS) data with non-sequential ML approaches. Xu et al. (2018) evaluated different non-sequential ML methods to upscale ET with high-resolution features available regionally in China. Both Zhao et al. (2019) and ElGhawi et al. (2023) used a non-sequential physics-constrained neural network approach to model ET, which has the potential to yield physically consistent and partially interpretable models. Recently, Nelson et al. (2024) published an hourly upscaling product of carbon and energy fluxes (X-BASE), built upon a novel framework (FLUXCOM-X) that enables the testing and application of different data streams and ML methods for upscaling in a flexible manner. They use a non-sequential model based on boosted regression trees (XGBoost; Chen and Guestrin2016) and account for memory effects via remote sensing state proxies.

Non-sequential ML approaches, however, cannot represent temporal variable interactions beyond the observable state proxies in contrast to, for instance, recurrent neural networks (RNNs; Lipton et al.2015). For time series regression, the long short-term memory (LSTM) network (Hochreiter and Schmidhuber1997) is a widely used architecture based on the RNN paradigm (Van Houdt et al.2020). Such sequential approaches have been evaluated for EC flux modeling at the site level. Reichstein et al. (2018) applied RNNs to model the weekly net ecosystem exchange (NEE) of carbon from nine European flux stations with meteorological forcing and showed the relevance of temporal information via a permutation test. Besnard et al. (2019) employed an LSTM architecture to model monthly NEE at EC sites and achieved better performance than with a non-sequential random forest. But still, they reported a poor representation of temporal dynamics in terms of both interannual variability and anomalies, the deviations from the mean seasonal cycle.

In the domain of deep learning, different model architectures are capable of processing sequential data. In the earth sciences, the LSTM has become the de facto standard, even though other architectures have been developed, such as the temporal convolutional network (TCN; Oord et al.2016; Bai et al.2018). The TCNs use sparse convolution along the temporal dimension to consider long-term effects more efficiently. More recently, models employing self-attention (Vaswani et al.2017) have shown noteworthy performance in many domains. These sequential models could also hold potential for EC flux modeling, as has been shown by Armstrong et al. (2022) and Nakagawa et al. (2023). While conceptually apparent, there is little systematic evidence of whether such sequential deep learning methods provide an advantage over non-sequential approaches for the upscaling of EC fluxes and of how these models respond to other issues inherent in upscaling, such as limited and unevenly sampled training data and distribution shift from the local point data to gridded fields.

In this study, we provide a systematic comparison of different ML approaches to the modeling of site-level ET fluxes and subsequently their upscaling to the global scale using the FLUXCOM-X modeling framework (Nelson et al.2024). A simple linear model, XGBoost, and a feed-forward fully connected neural network serve as baselines for non-sequential models. Two sequential models, one based on the LSTM architecture and another based on a TCN, account for temporal effects. We chose these models as they are conceptually different but both commonly used for time series simulation and forecasting tasks, and we acknowledge that other architectures could be used as well. We compare the model performances at the site level in a cross-validation setup and assess the relevance of dynamical memory effects for ET modeling. For each model, we conduct a feature ablation experiment, where we drop feature groups. The groups considered in addition to meteorology are precipitation, which we obtained from reanalysis data and not from site-level observations due to large gaps; dynamic state representations, based on remotely sensed observations; and plant functional types (PFTs), which are static descriptors of site characteristics. We provide and investigate cross-validation-based upscaling ensembles from the independent cross-validation models to test for robustness. To assess the impact of the model architecture on upscaling, we contrast our products globally to a set of land surface model simulations and to a semi-empirical approach (GLEAM). We investigate, in the context of upscaling, (a) whether the sequential models lead to more realistic and robust global and regional ET compared to independent estimates, (b) whether they are able to capture the temporal dynamics of ET better, and (c) how robust they are to the covariate setup and training data subsets compared to the non-sequential models.

The key contributions of this study are the following:

  • a systematic comparison of the effectiveness of different ML methods for site-level land–atmosphere ET flux modeling,

  • an assessment and discussion of the relevance of different covariates in the context of ecological memory effects for ET,

  • a characterization and comparison of an ensemble of upscaled ET estimates generated with different ML models.

2 Data sources and processing

We used hourly EC data from 2001 to 2020, processed by the ONEFLUX pipeline (Pastorello et al.2020). Only sites available under the CC BY 4.0 license were included in this analysis, i.e., FLUXNET 2015 (Pastorello et al.2020), ICOS Drought 2018 (Drought 2018 Team and ICOS Ecosystem Thematic Centre2020), ICOS Warm Winter 2020 (Warm Winter 2020 Team and ICOS Ecosystem Thematic Centre2022), or the more-recent ICOS or Ameriflux releases when present. In total, we used 287 sites with approximately 19 million hourly observations of ET and meteorological conditions. The approach by Jung et al. (2024) was used for quality flagging. We used latent heat energy as the target flux and converted it to ET, assuming a constant latent heat of vaporization of 2.45 MJ mm−1. The following meteorological covariates were considered: near-surface air temperature (Tair), vapor pressure deficit (Δe), shortwave irradiation (Rin), potential shortwave irradiation (Rin,pot), and the time derivative of potential shortwave irradiation (ΔRin,pot). Note that the time derivative, which is the difference between potential shortwave irradiation values for 2 consecutive hours, is intended to help the non-sequential models to discern the diurnal cycle. As precipitation (P) observation at the site level is often missing, we used the hourly ERA5 reanalysis instead (Hersbach et al.2020), extracted from the nearest pixel to the site. In addition, we used remote sensing observations from the moderate imaging spectroradiometer (MODIS) sensor on board both the Terra and Aqua satellite platforms (collection 006). These include the enhanced vegetation index (EVI; Huete et al.2002) and the normalized difference water index (NDWI; Gao1996), both retrieved at the site level from the MCD43A4 product and quality filtered based on the MCD43A2 product (spatial resolution of 500 mSchaaf and Wang2015a, b) and from the MCD43C4 product for the global data runs (Wan et al.2015a; spatial resolution of 0.05°). Additionally, the land surface temperature (LST) was obtained from MOD11A1 at the site level (Wan et al.2015a; spatial resolution of 1 km) and from MOD11C1 globally (Wan et al.2015b; spatial resolution of 0.05°). Although the MCD43A4 product for the reflectances uses observations from a period of 16 d to characterize and invert the bidirectional reflectance distribution function of a given pixel for the day at the center of the period, this operation is done over a temporally moving window at daily time steps, resulting in output data with daily frequency. The processing of the datasets, cutouts at the sites, and quality control correspond to the setup used in the FLUXCOM-X-BASE dataset (Nelson et al.2024; Walther et al.2022; Jung et al.2024). As an optional covariate, we use the plant functional type (PFT), available for all EC station sites. The nine PFTs were one-hot-encoded and repeated in time to match the hourly time series. One-hot encoding represents categorical variables as binary values, assigning a unique binary digit to each category. The sample time series of the covariates and ET are shown in Fig. 2.

https://bg.copernicus.org/articles/22/3965/2025/bg-22-3965-2025-f02

Figure 2A 1-year time series from the Hainich site (DE-Hai) in Germany. Meteorological covariates (hourly): near-surface air temperature (Tair), vapor pressure deficit (Δe), shortwave irradiation (Rin), potential shortwave irradiation (Rin,pot), time derivative of potential shortwave irradiation (ΔRin,pot), and precipitation (P). Remote sensing (daily): enhanced vegetation index (VEVI), normalized difference water index (VNDWI), and land surface temperature (Tsurf). Land–atmosphere target flux (hourly): evapotranspiration (yET).

Download

For upscaling, we used global meteorological data from the ERA5 reanalysis (Hersbach et al.2020), corresponding to the site-level variables. The hourly data were spatially resampled to a resolution of 0.05° using bilinear interpolation. These data were also used to fill gaps in site-level meteorological observations.

For the evaluation of the upscaling results, due to the lack of direct and spatially contiguous observations of ET, we used the Global Land Evaporation Amsterdam Model (GLEAM) v3 (Martens et al.2017) and global sums of yearly ET from 14 land surface modes (TRENDY v6; values extracted from Pan et al.2020) as reference. Note that these reference data sources do not represent the ground truth but are estimates derived using different approaches, independent from the data-driven upscaling performed here.

3 Methods

3.1 Experimental setup

We evaluate a set of sequential and non-sequential ML models at the site level in a leave-sites-out cross-validation setup. The models are trained with different types of covariates: meteorological site-level observations without precipitation (met), precipitation from ERA5 (prec), remote sensing (rs), and PFTs (pft). These experiments with different sets of variables as model inputs are summarized in Table 1. In total, six covariate setups were tested and combined with five ML models; i.e., 30 models were trained and evaluated at the site level. For the evaluation, we use the Nash–Sutcliffe modeling efficiency (Nash and Sutcliffe1970).

(1) NSE = 1 - t = 1 T y t - y ^ t 2 t = 1 T y t - y 2 ,

where yt is the observed ET, y^t is the predicted ET at time t, and y represents the mean of the observations. The NSE is calculated per site and can take values from −∞ to 1, reflecting the model performance relative to the mean of the observations. Values above 0 indicate better prediction than using the mean observations, and 1 is a perfect prediction. Note that for the evaluation of spatial patterns, the NSE was not computed per site but across sites, which corresponds to the R2.

Table 1The ablation experiment with different covariate groups: meteorological without precipitation (met, hourly), precipitation (prec, hourly), remote-sensing-based (rs, daily), and plant functional type (pft, constant). Each item corresponds to a unique covariate setup.

met: near-surface air temperature Tair, vapor pressure deficit Δe, shortwave irradiation Rin, potential shortwave irradiation Rin,pot, time derivative of the potential shortwave irradiation ΔRin,pot; prec: precipitation P; rs: enhanced vegetation index VEVI, normalized difference water index VNDWI, land surface temperature Tsurf; pft: plant functional type SPFT.

Download Print Version | Download XLSX

3.2 Modeling approach

With the goal of evaluating model performance at EC station locations and afterwards upscaling to the global scale, we tested a number of ML algorithms in a site-level cross-validation setup. We denote the modeling problem as

(2) y ^ s , t = f θ ( X s , t - K : t , c s ) .

Here, Xs,t-K:tR(K+1)×D is the D dynamic input covariates with up to K antecedent time steps, and cs∈ℝM is the M static (constant) input features. The target flux of ET is represented as y^s,tR at site s and time step t. Note that K=0 with only instantaneous covariates Xs,t is a special case where no antecedent time steps are considered (i.e., a non-sequential model). We aim to find the parameters

(3) θ = arg min θ L f θ ( X s , t - K : t , c s ) , y t

of function fθ, which minimize the loss function , given by the mean squared error (MSE).

As baselines, we used a linear regression (linearreg) as well as two non-sequential models, a fully connected feed-forward neural network (fcn), and extreme gradient boosting (xgboost). The latter was also used in the recent state-of-the-art global upscaling product xbase (Nelson et al.2024). The setup for these models was largely consistent with xbase; i.e., the same covariates were used here plus precipitation. The remote sensing and PFT covariates were repeated in time for every hour to obtain uniform inputs; i.e., the former were constant over a day and the latter were constant over the entire time series. Although the remote sensing covariates do, in theory, vary on a sub-daily basis, these variables are not available at the hourly resolution, and the diurnal variations are driven primarily by hourly-varying meteorological variables, although they interact with satellite-based features that change only on a daily to weekly basis. In addition to these non-sequential models, we used two sequential models: a simple LSTM architecture (a model able to learn temporal dynamics via its built-in memory processing mechanism) and a TCN model (which applies 1-D convolutions in time). Those sequential layers were stacked so as to achieve the extraction of complex temporal features. While the LSTM has, conceptually, an unlimited receptive field, the temporal context considered by the TCN depends on its hyperparameters. The neural-network-based models use the building blocks illustrated in Fig. 3 and were implemented in PyTorch (Paszke et al.2019) v1.13.

https://bg.copernicus.org/articles/22/3965/2025/bg-22-3965-2025-f03

Figure 3The three neural network layers used in this study: (a) a feed-forward neural network; (b) a temporal convolutional network (TCN), which applies causal (i.e., does not consider future time steps) 1-D convolutions in the time dimension; and (c) a long short-term memory (LSTM) model, which uses recursion for information flow in the time dimension. The model inputs (xs,t) at site s and time t are mapped to the output y^s,t.

Download

3.3 Model training

To identify models with the capacity to generalize well to unseen sites, we trained them following an eight-fold cross-validation scheme, for which the data splitting between sites was kept identical across different models and covariate setups. To decrease the dependency between the sets, we ensure that sites in close spatial proximity (below 0.05° distance) are part of the same set using the clustering of coordinates. The site groups are provided in the Appendix (Table B1). For each of the eight folds, six of the cross-validation sets were used for training (75 %), one for validation (12.5 %), and one for testing (12.5 %), such that each site appeared in the testing set once. The training and validation sets were used for model tuning with the early stopping algorithm: the model parameters were optimized on the training set, while the validation set was used to evaluate the generalizability regularly (10 times in each training epoch). Once the validation loss converged over a given number of validation steps (the “patience”), model training was halted, and the best parameters were restored. With these parameters, the model was applied to the independent test set. This approach yielded independent predictions for each site, which we then used to evaluate the model's performance on a site-level basis. For a speedup of the training, the model was iteratively fed with randomly selected sequences of 2 years. The first year was used for providing a temporal context similar to the “spinup” in dynamic process models, while the second was used for tuning. Note that the 2-year sequence was randomly sampled in every epoch, ensuring that all observations were, potentially, used for training with high likelihood.

We used a random search over a predefined set of hyperparameters. For each model, 20 parameter sets were sampled uniformly with replacement. The sets are reported in Table A1 in the Appendix. Note that we selected hyperparameter ranges based on prior experiments; i.e., we excluded values that performed consistently badly in order to obtain a denser sampling of the sensitive ranges. With this protocol, we tuned hyperparameters independently for each model except for linearreg, which has no hyperparameters. The same hyperparameter set was used throughout the cross-validation for each model setup. Thus, the cross-validation ensemble is composed of models with the same hyperparameters but trained on different subsets of the data.

To quantify model uncertainty on the site level, we performed the cross-validation for the six best-performing models (in terms of validation MSE) from hyperparameter tuning for each model and setup individually. For later analysis and upscaling, only the best-performing model was used.

3.4 Upscaling

To achieve global coverage, we fed the models with harmonized and gridded data from 2001 to 2021 with 0.05° spatial and hourly temporal resolution. Due to the high computational demands of the upscaling, we decided to use only two well-performing covariate setups across all models. We selected the met+rs and the met+prec+rs+pft (full) setups as they were among the best-performing setups. The linearreg model was excluded, as it performed significantly worse than the non-linear algorithms in the site-level cross-validation. For each of the four remaining ML models, we compute an ensemble of eight upscaling products. The members, hereafter referred to as the “cross-validation ensemble”, correspond to the models obtained from the cross-validation folds, i.e., each fold yielded one model which was trained and evaluated on an independent set of sites. Note that this differs from the X-BASE setup (Nelson et al.2024), where the cross-validation was used exclusively for model evaluation, and the upscaling was done with a single model trained again on additional sites without holding out a test set. This method does not yield an ensemble and is, therefore, not suited for the evaluation of within-model upscaling robustness. The upscaled products are then evaluated by an ML model intercomparison and by contrasting global yearly sums and regional cross-validation ensemble mean and variability to independent products.

4 Results and discussion

4.1 Site-level modeling of evapotranspiration (ET)

In this section, the EC site-level prediction of ET is evaluated based on the cross-validation setup. We aim to understand the impact of different covariate types and ML approaches on performance at different temporal scales and to assess the relevance of sequential model architectures on reproducing the ET observed at EC sites.

4.1.1 Model performances across scales

Figure 4 presents site-level ET accuracy in terms of NSE for different ML models and covariate groups across scales. We now focus on the best-performing models (solid lines in Fig. 4). Overall, model outcomes were more influenced by the choice of covariates than by the ML algorithm used. Notably, a significant interaction between ML models and covariates was observed.

https://bg.copernicus.org/articles/22/3965/2025/bg-22-3965-2025-f04

Figure 4Site-level evaluation for modeled evapotranspiration: median Nash–Sutcliffe efficiency (NSE) across sites for different models (lines) and covariates (x axis) for various scales (panels). The shaded area represents the range of the top six models from hyperparameter tuning per model setup, with the solid line indicating the best-performing model (top one). This can be interpreted as model uncertainty. Note that the model selection during hyperparameter tuning was based on the validation set and the mean squared error. Consequently, the best model is not necessarily the top performer in terms of NSE. The scales shown are raw for hourly, daily for daily aggregates, seasonal for daily seasonal, anom for daily anomalies (daily minus seasonal), and iav for interannual variability. The spatial scale reflects the NSE across site mean values, indicating the ability of the models to capture spatial variability.

Download

In general, the ML models outperformed linear regression by a substantial margin. On the raw and daily timescales, sequential models exhibited the best performance, maintaining stable NSE values of 0.70–0.75 (raw) and 0.60–0.65 (daily) across data setups. Non-sequential models performed worse when using only meteorological covariates but showed a significant improvement when remote sensing covariates were included, achieving a similar performance to the sequential models.

On a seasonal scale, the sequential models outperformed the others in the absence of remote sensing covariates. However, when remote sensing covariates were included, the performance differences between models became less pronounced. For anomalies, adding remote sensing covariates enhanced the model performance across all setups, with sequential models consistently outperforming non-sequential models. Notably, this was the only scale where precipitation had a clear positive impact across all models compared to the met+rs+pft setup.

For interannual variability, model performance was generally poor across all setups, with no clear patterns, and the small y-axis range underscores this. Adding PFTs and precipitation did not improve and, on the contrary, in some cases, reduced performance. On the spatial scale, the model performance was similar across models, although the sequential models exhibited a slight advantage with the met and met+prec setups. Including PFT covariates notably improved the performance of all models on this scale.

Regarding model uncertainty (shaded areas in Fig. 4), the sequential models generally showed lower robustness than non-sequential models, likely due to their added complexity. The sequential models consistently outperformed the non-sequential ones on the raw scale. On the daily and seasonal scales, the differences between the sequential models were primarily driven by model uncertainty; i.e., the uncertainty ranges displayed in Fig. 4 overlapped strongly on the spatial scale. On the anomaly scale, the sequential models reliably outperformed the non-sequential models across the top six setups. Given the small performance range and large model uncertainties on the iav scale, caution is needed when interpreting these differences. Similarly, the differences on the spatial scale were minimal, and the relative uncertainties were substantial.

Next, we discuss the broader implications of these findings. The linear models (linearreg) consistently underperformed because evapotranspiration is governed by complex interactions and non-linear functions, making the advantages of ML methods particularly evident. While the sequential models showed only marginal improvement when adding covariates related to the ecosystem state, the non-sequential models exhibited more substantial improvements. This serves as a sanity check for the sequential models: they were able to extract additional information from the temporal meteorological covariates, as expected. However, incorporating remote sensing covariates improved and stabilized their performance. Conversely, this suggests that remote sensing covariates serve as useful proxies for ecological memory: while the sequential models could extract additional information from antecedent covariates, most of this information appeared to be contained in the remote sensing covariates, allowing the non-sequential models to achieve a similar performance.

On the anomaly scale, we observed a more noticeable performance increase when considering remote sensing, even for the sequential models. This is important, as anomalies are crucial for studying and quantifying ecosystem responses to uncommon or extreme conditions. This performance increase could be linked to processes observable by remote sensing but not directly derivable from meteorology, such as the effect of management on crops and forests and natural disturbances. Here, adding precipitation improved the performance of all models. This is not surprising, as precipitation is a key driver of ET through its influence on soil moisture (Nelson et al.2020). However, the improvement was comparatively small, which may be due to the use of precipitation reanalysis data that may not fully represent local conditions. The low performance on the iav scale was also reported by Jung et al. (2019) and Nelson et al. (2024).

It is worth noting that adding PFTs as covariates did not improve, and in some cases even reduced, model performance on the temporal scales. On the spatial scale, however, their inclusion was beneficial, highlighting the potential importance of spatial covariates for modeling EC fluxes and upscaling. PFTs have long been criticized for not accurately representing the continuous characteristics of ecosystems (Reichstein et al.2014; Kattge et al.2011). Our experiments suggest that while adding PFTs provides additional information for representing spatial patterns, they may also harm extrapolation on other scales, arguably due to the inflated covariate space; indeed, each of the nine PFTs introduces another input dimension due to one-hot encoding. We therefore recommend exploring alternative spatially continuous variables, such as soil properties or plant traits, that could summarize ecosystem functional properties.

These findings advocate for comprehensive feature selection to identify more relevant static features, thereby avoiding unnecessary inflation of the input dimensionality. Alternatively, or in addition, location embeddings, such as SatCLIP (Klemmer et al.2025), could improve model generalizability by providing a condensed representation of land surface characteristics.

4.1.2 Memory effects matter

As previously mentioned, the performance gap between the non-sequential and sequential models decreased when remote sensing observations were incorporated as covariates. We explore these differences in Fig. 5. As shown in panel (a) (Fig. 5a), which displays the absolute error difference between the lstm and xgboost models, the non-sequential models performed worse when high incoming radiation (> 200 W m−2) was paired with either low or high observed ET. To represent these conditions, the models must implicitly learn about water availability. It appears that the sequential model was able to learn proxies of wetness from the meteorological time series, whereas the non-sequential model was unable to do so. Instead, the non-sequential model seems to have learned an average behavior, which performed well in most situations.

https://bg.copernicus.org/articles/22/3965/2025/bg-22-3965-2025-f05

Figure 5Comparison of a sequential (lstm) and a non-sequential (xgboost) model in terms of absolute error and mean predicted ET in the space of observed evapotranspiration × shortwave irradiation: panels (a) and (b) show the difference in absolute error between the lstm and xgboost models, with panel (a) showing the difference when only meteorological covariates are used and panel (b) showing the difference when precipitation, remote sensing, and PFT covariates (full setup) are included. Here, magenta indicates cases where the sequential model outperforms the non-sequential model, and green is vice versa. Panels (c) and d) show the difference in mean predicted ET between the lstm and xgboost models for the respective covariate setups. In these panels, red colors indicate an underestimation of ET by xgboost compared to lstm, and blue indicates the opposite. The histograms represent the marginal data distribution.

Download

When remote sensing covariates were added as well as PFTs and precipitation, the differences in performance were reduced but not entirely eliminated (Fig. 5b). This interpretation is further supported by panels (c) and (d) in Fig. 5, which show the difference in mean predicted ET between the sequential and non-sequential models. With access to only meteorological covariates (Fig. 5c), the non-sequential model overestimated ET, with high incoming radiation but low observed ET – representing dry conditions that the model failed to recognize. In contrast, high observed ET was underestimated by the non-sequential model, which likely corresponds to wet conditions. After incorporating remote sensing, PFTs, and precipitation covariates (Fig. 5d), the performance differences were substantially reduced. This comparison underscores the importance of memory effects in ET modeling and illustrates how remote sensing covariates, while useful, are not perfect proxies for ecological memory.

Furthermore, including precipitation may be beneficial for the sequential models in principle, as these models could learn soil moisture dynamics. However, as noted earlier, the performance increase was not as pronounced as expected (Fig. 4), possibly due to the use of reanalysis data that may not fully capture local conditions.

4.2 Scaling evapotranspiration to global coverage

With the models optimized at the site level, we create global ensembles of ET estimates for the met+rs and the full (met+prec+rs+pft) covariate setups. The ensemble members were trained with different subsets of the training data within the cross-validation scheme. At the site level, the differences between ML models were small when considering remotely sensed observations or PFTs as covariates. However, when scaling globally, data distribution shifts can (and will) affect different model types in different ways. These shifts arise from the different scales of the measurements (point at EC site versus grid globally), the different data products used (direct observation of meteorological variables at EC site versus reanalysis globally), and the spatial extrapolation into different ecoclimatological conditions from irregularly and sparsely sampled locations. In this section, we consider the performance of the different ML approaches and covariate setups while scaling out of the EC station locations.

https://bg.copernicus.org/articles/22/3965/2025/bg-22-3965-2025-f06

Figure 6Global annual evapotranspiration (ET) per model. The violin plots represent the density of independent cross-validation runs (black dots), with their mean values across runs displayed as white dots. For the machine learning models, the met+rs setup (meteorological and remote sensing covariates) is shown in light gray, while the full setup (meteorological, precipitation, remote sensing, and PFT covariates) is shown in dark gray. The data from a number of land surface models (lsm), the GLEAM product (gleam), and FLUXCOM X-Base (xbase) are added as a reference.

Download

4.2.1 Global patterns of evapotranspiration

Figure 6 shows global annual ET estimates for all ML models and the two selected covariate setups. With the met+rs setup, the non-sequential models (xgboost, fcn) estimated about 65×103km3 yr−1 global annual ET, which is close to the land surface model (lsm) ensemble mean (64.7±6.9×103km3 yr−1), while the sequential models (tcn, lstm) predicted roughly 8 % lower values (59.8 and 61.8×103km3 yr−1, respectively). With the full setup, xgboost showed the strongest increase in global ET (+10 %), while fcn, lstm, and tcn increased modestly (1 %, 0.5 %, and 5 %, respectively). xgboost was closest to gleam and xbase, whereas neural networks aligned more with lsm estimates.

https://bg.copernicus.org/articles/22/3965/2025/bg-22-3965-2025-f07

Figure 7Spatial model evaluation and comparison for the met+rs and full (meteorological, precipitation, remote sensing, and PFT covariates) setup. Grid-level cross-validation ensemble mean ET for each model is shown in panel (a) for met+rs and panel (b) for full. The first column represents xgboost, and the remaining columns show the difference in xgboost. Panel (c) shows the difference met+rs minus full. The color scale is consistent across panels and reflects mean ET in mm d−1. Inset histograms show the distribution of values weighted by grid cell area; the median is indicated by a dashed black line.

Figure 7 displays spatial patterns of ensemble mean ET per model and covariate setup. The models aligned well spatially in the met+rs setup (Fig. 7a), with lower ET by the sequential models, especially in regions in the Southern Hemisphere, which have sparse coverage by EC stations. The full setup (Fig. 7b) revealed stronger divergences across models, particularly in arid zones such as the Sahara and Arabian Peninsula and generally in the subtropical zones. Figure 7c shows that changes in ET estimates due to covariate setup differences were low in temperate zones but more pronounced in arid and tropical regions. lstm appeared relatively robust across setups.

https://bg.copernicus.org/articles/22/3965/2025/bg-22-3965-2025-f08

Figure 8Temporal correlation of monthly ET anomalies among models and covariate setups. (a) Correlation between monthly ET anomaly time series from the met+rs and full setups. (b) Correlation between monthly ET anomalies of met+rs and the GLEAM product. (c) Difference in temporal correlation with GLEAM between the full and met+rs setups. Inset histograms show area-weighted value distributions, with dashed black lines indicating the median.

A temporal anomaly correlation analysis (Fig. 8) reveals high agreement (r>0.8) between covariate setups in most areas, except for arid zones, and only the tcn maintained high consistency globally. A comparison with gleam (Fig. 8b) showed moderate correlations (mean r≈0.5), with discrepancies mainly in tropical and arid regions. Figure 8c highlights where models improved alignment with GLEAM when using additional covariates: tcn improved slightly, while lstm improved most consistently across dry regions.

https://bg.copernicus.org/articles/22/3965/2025/bg-22-3965-2025-f09

Figure 9Median absolute deviation (MAD) of the cross-validation ensemble for each model setup, representing spatial uncertainty in mm d−1. (a) MAD for the met+rs setup. (b) MAD for the full setup. Inset histograms show the distribution of values weighted by grid cell area; the dashed black line marks the median.

Figure 9 shows uncertainty via median absolute deviation (MAD) across ensemble members based on monthly values. xgboost had the lowest spread overall, with modest increases in tropical zones and more pronounced increases in arid zones in the full setup. The neural networks showed a higher ensemble spread: in the tropics and subtropics for met+rs and in arid regions for the full setup. In the latter, uncertainty patterns aligned with ecosystem boundaries, likely reflecting PFT transitions.

Table 2Global evapotranspiration (ET) correlation across model ensemble members. The correlation is based on global ET sums across cross-validation members trained on the same sites. The upper triangle (bold) shows the Pearson correlation, and the lower triangle shows the Spearman rank correlation.

Download Print Version | Download XLSX

We saw that models showed biases in terms of global ET sums, but apart from xgboost, the values were relatively robust to changes in covariate sets. From the spatial and temporal patterns (Figs. 7, 8, and 9), it becomes evident that the uncertainty originates mainly from tropical and arid to semi-arid regions. Finally, we want to investigate if the differences are related to the technical setup (architecture and covariates) or to the variability in the training data used in the cross-validation. Therefore, we investigate the alignment of members shown in Fig. 6: we compute the linear (r) and rank (ρ) correlation between global sums of ET between all models and covariate setups. In other words, we quantify if using the same data subset in cross-validation leads to consistent upscaling behavior in terms of global ET. The results are shown in Table 2. Global ET estimates from xgboost were highly correlated between setups (r=0.90, ρ=0.76), suggesting consistent behavior across training data subsets. Neural networks exhibited a lower correlation (r=0.32–0.75), indicating greater variability. Overall, consistency in upscaling behavior was not strongly tied to model type or covariate setup.

4.2.2 Model biases across architectures and covariate setups

Most models, particularly the sequential ones, estimated global ET – averaged across members – at the lower end of current estimates from independent methods of about 70±5×103km3 yr−1 (Jung et al.2019) and compared to gleam and xbase. Energy balance non-closure at EC sites could depress ET by up to 20 % (Jung et al.2019), yet recent findings suggest that the closure gap may contribute only moderately to the global ET bias, as the closure gap very likely stems primarily from sensible rather than latent heat fluxes, i.e., ET (Zhang et al.2024; Mauder et al.2024). Because both xgboost-full and the xbase product would share the same closure issue but still reach the benchmark range, the deficit of the other models must originate elsewhere; options are training data subsets, the ML approaches, covariate setup, or data distribution shift due to extrapolation into underconstrained regions and product types (site level versus gridded).

Disentangling these possible causes is difficult, but we can draw some conclusions from the results. We have shown that different training subsets did not lead to consistent upscaling behavior (Table 2). In other words, taking the same training data subset for model training did not lead to similar behavior across ML models in terms of global ET. This suggests that the ensemble variance at the global scale is not driven by training data but rather by how the ML models extrapolate out of the training data distribution. It indicates that, when evaluating the global sums alone, neither neural networks versus xgboost nor sequential versus non-sequential models can be regarded as inherently more consistent – or therefore more reliable – for global upscaling.

A further complication is the change from EC site-observed meteorology to reanalysis grids. At EC sites, meteorological variables are locally observed, whereas global inputs are derived from reanalysis grids (with a spatial resolution of 0.25°) with assimilation and spatial smoothing (Hersbach et al.2020; Parker2016; Grusson and Barron2022; Valmassoi et al.2023). Sequential networks, which exploit fine-scale temporal structure, are theoretically more sensitive to such a distribution shift than tree-based or feed-forward models. Nevertheless, our results show that spatial patterns are robust in data-rich regions, both across architectures (ensemble means, Fig. 7) and within architectures (ensemble spread, Fig. 9). Because every model and covariate set converges where observational support is strong, the station-to-grid shift cannot be the dominant source of the spatial or global sum discrepancies. Instead, the residual differences arise from the extrapolation into data-sparse regions, with different behavior across models and covariate setups.

The model runs without PFTs (Figs. 7a and 9a) showed a better agreement with each other (Fig. 7a versus Fig. 7b) and within ensemble members (Fig. 9a versus Fig. 9b). This suggests that excluding the PFTs from the covariates improves the model robustness. However, we saw at the site level that the PFTs improved the model performance, particularly at the spatial scale (Fig. 4). At the same time, we observed unrealistic spatial patterns with the full setup: in particular, xgboost and fcn with the full setup estimated large ET in arid regions like the Sahara and Arabian Peninsula. These large ET estimates align poorly with our understanding of ET processes, and xbase estimates in these areas have also been shown to be too high (Nelson et al.2024). Thus, while fcn, and particularly xgboost with the full covariate setup, appears close to the global mean ET benchmark, this is likely for the wrong reasons, i.e., overestimation in arid regions. The sequential models also yielded higher global ET when including PFTs, likewise originating from arid regions, but the increase was less pronounced. lstm showed the greatest robustness across setups. For all models, we observed an increase in the ensemble spread in arid and also tropical regions when adding PFTs (Fig. 9). It is possible that in the site-level cross-validation, these effects were not visible because of a lack of EC stations in arid and tropical regions.

While we cannot answer which model in combination with which covariate setup yields the “best” upscaling results, we can conclude that non-temporal models yielded more realistic global ET estimates, even if this is largely due to covariate-driven artifacts in data-scarse regions rather than genuinely improved process representation. The sequential models were also sensitive to covariate setup, but lstm was more robust compared to the other models.

4.2.3 Consistent temporal dynamics

Despite variability in absolute ET magnitudes, the models showed strong internal consistency in their temporal dynamics (Fig. 8). The high correlation between covariate setups (Fig. 8a) indicates that models captured similar temporal variations and responded consistently to climatic forcing.

The temporal correlation with gleam was moderate overall, as discussed previously. The GLEAM ET product is based on conceptual models driven by remote sensing inputs and meteorological forcing. We do not treat it as ground truth, but we assume it provides spatially consistent temporal patterns due to its incorporation of prior knowledge. In contrast, our ML models rely solely on data-driven representations and may be sensitive to data shifts due to generalization beyond data support and data types. Given this, the overall moderate agreement with GLEAM is not surprising and does not imply poor model performance. Rather, we interpret divergences – especially in tropical and arid regions – as a consequence of low data availability and known challenges in these environments.

The temporal correlation with gleam improved notably for the lstm model in arid and some tropical regions when additional covariates were included (Fig. 8c). This improvement was not observed for xgboost, which showed a slight degradation, particularly in arid climates, or for fcn, which remained largely unchanged. We attribute the improved performance of sequential models primarily to the inclusion of precipitation, consistent with their enhanced performance on the anomaly scale at the site level when this covariate was added (Fig. 4).

It is unclear why tcn did not show a similar improvement in arid regions. At the site level (Fig. 5), tcn actually outperformed lstm on the anomaly scale, albeit with higher model uncertainty. This could suggest that lstm learns a more robust representation of temporal dynamics in these regions, which is not evident from the site-level cross-validation as those regions are largely absent from the training data. In contrast, tcn may be more sensitive to distribution shifts. There is evidence that lstm architectures are better suited to capturing hydrologically relevant temporal patterns compared to convolutional models (Kraft et al.2025). While lstm can theoretically access the full temporal context, tcn is constrained to a fixed temporal window of approximately 8 d in both the met+rs and full setups (Table A1). Note that this temporal context is, for the tcn architecture, related to the hyperparameters that yielded the best model performance. This limited context may explain its reduced ability to generalize temporal patterns in data-sparse regions.

In summary, temporal dynamics remained relatively consistent across methods. Sequential models, especially lstm, benefited from richer covariate input, particularly precipitation, improving alignment with an independent product in arid regions. This highlights the added value of using sequential models for representing temporal dynamics in ET modeling and upscaling, yet the interaction between model architectures and covariates, particularly in data-sparse regions, also suggests that a profound covariate selection is necessary to identify best-working setups.

4.3 Lessons learned and outlook

4.3.1 Site-level performance: small but consistent advantages for sequential models

At the site level, sequential models consistently outperformed non-sequential models for ET flux prediction, especially when evaluated on the anomaly scale (Fig. 4). This is in line with previous findings for NEE modeling (Besnard et al.2019), suggesting that temporal dependencies, such as ecosystem memory effects, are captured more effectively by models with recurrent or convolutional memory. The inclusion of remote sensing variables (e.g., vegetation indices) reduced the performance gap between sequential and non-sequential models by providing state proxies, although the advantage for sequential models persisted. Adding precipitation from reanalysis further improved performance at the anomaly scale, particularly for the sequential models, confirming the value of dynamic hydrological information. However, these improvements were relatively modest, indicating that data availability and quality, rather than model architecture, remain the primary bottlenecks. More-advanced architectures, such as temporal transformers, may offer further gains, but recent work in gross primary production (GPP) modeling has shown only marginal improvements (Nakagawa et al.2023), providing further evidence that EC flux modeling is still a data-limited problem.

4.3.2 Global upscaling: both model architecture and covariates drive uncertainty

Small changes in the covariate setup led to moderate differences in the global ET means (Fig. 6), especially for xgboost with the full setup, which showed an increase of 10 % in global ET, supposedly due to the strong leverage effects of PFTs in sparsely sampled regions. The moderate differences were partially due to error compensation. While we found no clear evidence that any architecture yields more realistic absolute ET estimates, our results point to systematic biases that depend on both model design and the covariates used. This highlights the importance of methodological choices, including architecture, covariate selection, and cross-validation design, in driving upscaling uncertainty.

4.3.3 Spatio-temporal patterns: robust dynamics, variable magnitudes

Despite differences in magnitude, all models showed consistent temporal ET patterns and strong alignment in anomaly correlations, both among themselves and, to a lesser extent, also with the GLEAM product (Fig. 8), yet the robustness drastically decreased outside of the training domain. With the inclusion of precipitation and PFTs in the full setup, the sequential models – particularly lstm – showed improved temporal agreement with GLEAM in arid regions, supporting their value for representing memory effects related to soil moisture. However, these benefits were largely limited to the temporal dimension; biases in global means and ensemble spread remained pronounced, especially for data-sparse regions where additional covariates such as PFTs introduce high leverage. Given their lower computational complexity and more stable behavior, we find that non-sequential models like xgboost currently offer a pragmatic and robust solution for global-scale ET upscaling, especially when paired with carefully selected meteorological and remote sensing inputs.

4.3.4 Outlook: addressing biases and uncertainties in data-scarce regions

Moving forward, reducing model biases and uncertainty in upscaling requires both methodological advances and improved data. Rather than prioritizing complex model architectures, future efforts should focus on four complementary pathways.

  • i.

    Feature selection and additional data constraints: Deep learning presents promise for enhancing flux modeling and upscaling and offers advanced computational techniques capable of managing complex non-linear interactions within ecosystems. However, to maximize the effectiveness of deep learning in such a data-limited setting, it is essential to implement additional constraints and to integrate richer data sources (Reichstein et al.2019). The accuracy of deep learning models relies heavily on the quality and diversity of the input data (Karpatne et al.2019). Enhancing these models with additional covariates that accurately reflect ecological and atmospheric conditions can significantly improve their predictive power. Additionally, expanding the network of flux stations and sharing the data for scientific applications would enhance the data pool to cover more diverse ecological conditions and climate zones, thereby enriching the training data used for model calibration and validation. Furthermore, applying constraints at a regional level, akin to the approach by Upton et al. (2024), who used an ensemble of atmospheric inversions of NEE as large-scale guidance for flux upscaling, could be used to reduce biases. For ET modeling, large-scale water balance could be used as a regional constraint, for example.

  • ii.

    Additional data sources. To further refine the performance of deep learning approaches in EC measurement upscaling, it could be beneficial to tap into additional data sources. Approaches such as transfer learning can be particularly effective (Caruana1997; Pan and Yang2010). By applying knowledge gained from one region to another region, or from related and richer datasets, models can achieve better generalization, especially in data-sparse areas. To deal with shifts in covariates due to the various reasons discussed previously, domain adaptation (He et al.2023) could provide a useful toolbox to reduce upscaling biases.

  • iii.

    Physical constraints. As a complementary pathway, incorporating prior scientific knowledge into deep learning models could help to address challenges associated with the data extrapolation and distribution shifts encountered in upscaling (Reichstein et al.2019; Kraft et al.2022). Such integration aids in aligning model outputs with established physical laws and ecological principles, thereby improving the reliability of the predictions (Reichstein et al.2022). Physics-informed and hybrid physics/ML approaches represent a cutting-edge direction in the field of flux modeling, as they merge the empirical strengths of deep learning with the deterministic nature of physical models. For upscaling into undersampled regions, such constraints can nudge the model outputs towards physically more plausible solutions. As an example, encoding simple relationships between precipitation and evaporation, or vegetation and transpiration, could help in reducing ET estimates in arid regions where EC stations are lacking. Although challenging, more comprehensive physical process parameterizations, such as the Penman–Monteith equations, can be combined with ML to estimate ET (Zhao et al.2019; ElGhawi et al.2023). This could, in principle, reduce the widely reported regional biases in upscaling EC fluxes with ML, which we identified as currently the main challenge in flux upscaling.

  • iv.

    Systematic benchmarking. Finally, systematic benchmarking frameworks, such as FLUXCOM-X (Nelson et al.2024), are essential for disentangling the complex interactions between models, covariates, and evaluation setups. Such frameworks enable controlled ablation studies and targeted diagnostics, helping to build a more-transparent understanding of uncertainty sources in ML-based flux upscaling.

In conclusion, while deep learning provides valuable tools for modeling land–atmosphere interactions, the key to better global ET estimates lies in more comprehensive observational data, thoughtful covariate selection, targeted physical constraints, and methods to reduce extrapolation bias – rather than in model complexity alone.

5 Conclusions

In this study, we assessed different covariate setups and machine learning approaches for modeling ET fluxes at eddy covariance sites through cross-validation and evaluated the robustness and quality of globally upscaled ET estimates. From our site-level analysis, we conclude that sequential deep learning models can outperform non-sequential models for ET flux modeling, particularly on the anomaly scale. The sequential models captured memory effects related to water availability, which improved their ability to represent the temporal dynamics of ET. However, this advantage diminished when remote sensing covariates were included, as these effectively act as proxies for ecosystem memory. Adding precipitation as a covariate led to small performance improvements, especially for sequential models and on the anomaly scale, confirming its value in representing hydrological constraints. The inclusion of PFTs increased the ability of the models to capture spatial variability across sites, yet their discrete nature raises concerns about their broader utility. We therefore recommend the further exploration of alternative static variables or targeted feature selection to maintain a parsimonious covariate set.

Globally, sequential and non-sequential models produced small but systematic differences in mean ET, suggesting that different model types learn distinct representations and respond differently and randomly when extrapolating out of the training distribution. When using PFTs as an additional covariate, both divergence of spatial means of ET among models and ensemble spread within models increased, particularly in arid regions where data support is limited. This highlights the large leverage of the PFTs on upscaling results when extrapolating out of the covariate space and underscores the need for a better understanding of the role of covariates in upscaling. While the sequential models suffered from biases similar to those of their non-sequential counterparts, the sequential models achieved better temporal alignment with the GLEAM product, particularly when precipitation was included.

In the context of our experiments, we can answer the research questions posed in the introduction as follows.

  • a.

    Do the sequential models lead to more realistic and robust global and regional ET estimates?
    No, the sequential models yielded lower global ET estimates than expected from independent estimates. However, the larger estimates of the non-sequential models were likely due to an overestimation of ET in arid regions. Nonetheless, the sequential models, especially the LSTM, were more robust to covariate shifts, supposedly due to their lower dependency on PFT due to their ability to link spatial variability to temporal features in the covariates.

  • b.

    Do the sequential models capture the temporal dynamics of ET better?
    Yes, the sequential models captured temporal dynamics better than the non-sequential models, at both the site and upscaling level, particularly when using precipitation as an additional input and with the LSTM architecture.

  • c.

    How stable are the sequential models in relation to the covariate setup and training data subsets compared to the non-sequential models?
    The sequential models showed a similar ensemble spread to the non-sequential neural network, indicating that robustness towards training data subsets is more linked to the model type (decision tree versus neural network) than to the sequential nature of the models.

Considering the added complexity and computational cost of sequential neural networks, the relatively modest performance gain at the site level, and their underestimation of global ET alongside larger ensemble spreads, we conclude that these models are not strictly required for global ET upscaling. Non-sequential ML approaches, such as XGBoost, can deliver similarly robust estimates across scales when supported by high-quality meteorological and remote sensing covariates. Yet, XGBoost overestimated ET in arid regions, which raises some concerns, but these are also not exclusive to this model type. In particular, the LSTM architecture was robust towards covariate setups and able to learn subtle temporal dynamics, which could be beneficial in specific applications. These findings highlight the importance of structural model diversity in ensemble setups in assessing the robustness and uncertainty of upscaling results.

The potential for upscaling ET to the global scale using modern ML methods is constrained by the information content and representativeness of the EC training data. As a result, seemingly small changes in covariate setups can exert a great influence over upscaling results, particularly in data-sparse regions. To enhance the robustness and physical consistency of global ET products, we recommend pursuing complementary strategies: integrating richer covariates, increasing EC site coverage, applying regional constraints, leveraging transfer learning, and embedding prior scientific knowledge into ML architectures. By following these pathways, deep learning has the potential to produce more accurate, robust, and physically grounded predictions of terrestrial evapotranspiration across diverse environmental settings.

Appendix A

Table A1 provides an overview of the hyperparameter search space and the best-performing combination of settings for each model and covariate setup. We sampled 20 random hyperparameter configurations for each model, evaluating them via early stopping on a validation set to determine the final selections. For the tcn model, the temporal context length considered (temp. contex) depends on the hyperparameters of the model, leading to a range of context lengths from about 4 d (91 h) to about 19 d (451 h). The number of parameters in each model (# parameters) was derived from the chosen architecture, highlighting the fact that the complexity increases substantially with deeper layers or larger hidden sizes.

Table A1Hyperparameter search space and the best-performing hyperparameter per model and setup. The best combination was found by evaluating 20 random samples based on early stopping validation loss. The xgboost and fcn models are non-sequential, the lstm model has a minimum of 1 year (9000 hourly time steps) and a maximum of 2 years of theoretical context (18 000 hourly time steps), and the tcn model's temporal context depends on the hyperparameters (reported in the row temp. context), ranging from 90 to 450 h, i.e., about 4 to 19 d. The number of parameters per model is reported in the row # parameters.

Download Print Version | Download XLSX

Appendix B

Table B1 displays the eight cross-validation groups, each column representing one group, with its corresponding eddy covariance site IDs. We used this grouping to ensure that the training and validation sets differ systematically across folds, minimizing spatial autocorrelation effects at the site level. The sites within each group span a range of climatic and ecophysiological conditions, allowing for a more robust evaluation of model generalization.

Table B1The eddy covariance site groups used for cross-validation. The eight groups correspond to columns, and the items correspond to site IDs.

Download Print Version | Download XLSX

Data availability

The upscaled ET fields generated for this study are available from the corresponding author upon reasonable request.

Author contributions

BK implemented the model architectures and data pipeline for the neural networks and conducted the analysis. He also took the lead in writing the article. JN, SW, FG, MJ, BK, and ZH contributed to the FLUXCOM-X framework, which this study builds upon. SW, JN, FG, UW, and MJ provided, processed, and cleaned the datasets used. BK, JN, SW, MJ, GD, MRe, and WZ contributed to the scientific evaluation of the results. MK, MRu, and DT contributed their expertise to the data-scientific and machine learning aspects of the study. All co-authors contributed to the paper.

Competing interests

The contact author has declared that none of the authors has any competing interests.

Disclaimer

Publisher's note: Copernicus Publications remains neutral with regard to jurisdictional claims made in the text, published maps, institutional affiliations, or any other geographical representation in this paper. While Copernicus Publications makes every effort to include appropriate place names, the final responsibility lies with the authors.

Acknowledgements

We acknowledge funding from the European Research Council for this study. We thank the Max Planck Institute for Biogeochemistry for providing the computational and data infrastructure for conducting this research. We appreciate the financial support from the Max Planck Society to publish this paper as open access.

Financial support

This research has been supported by the European Research Council (ERC) under the European Union's Horizon 2020 research and innovation program (grant agreement no. 855187).

The article processing charges for this open-access publication were covered by the Max Planck Society.

Review statement

This paper was edited by Nicolas Brüggemann and reviewed by Simon Besnard and Wenli Zhao.

References

Armstrong, S., Khandelwal, P., Padalia, D., Senay, G., Schulte, D., Andales, A., Breidt, F. J., Pallickara, S., and Pallickara, S. L.: Attention-based convolutional capsules for evapotranspiration estimation at scale, Environ. Model. Softw., 152, 105366, https://doi.org/10.1016/j.envsoft.2022.105366, 2022. a

Bai, S., Kolter, J. Z., and Koltun, V.: An Empirical Evaluation of Generic Convolutional and Recurrent Networks for Sequence Modeling, arXiv [preprint], https://doi.org/10.48550/arXiv.1803.01271, 4 March 2018. a

Baldocchi, D., Falge, E., Gu, L., Olson, R., Hollinger, D., Running, S., Anthoni, P., Bernhofer, C., Davis, K., Evans, R., Fuentes, J., Goldstein, A., Katul, G., Law, B., Lee, X., Malhi, Y., Meyers, T., Munger, W., Oechel, W., U, K. T. P., Pilegaard, K., Schmid, H. P., Valentini, R., Verma, S., Vesala, T., Wilson, K., and Wofsy, S.: FLUXNET: A New Tool to Study the Temporal and Spatial Variability of Ecosystem-Scale Carbon Dioxide, Water Vapor, and Energy Flux Densities, B. Am. Meteorol. Soc., 82, 2415–2434, https://doi.org/10.1175/1520-0477(2001)082<2415:FANTTS>2.3.CO;2, 2001. a

Beer, C., Reichstein, M., Tomelleri, E., Ciais, P., Jung, M., Carvalhais, N., Rödenbeck, C., Arain, M. A., Baldocchi, D., Bonan, G. B., Bondeau, A., Cescatti, A., Lasslop, G., Lindroth, A., Lomas, M., Luyssaert, S., Margolis, H., Oleson, K. W., Roupsard, O., Veenendaal, E., Viovy, N., Williams, C., Woodward, F. I., and Papale, D.: Terrestrial Gross Carbon Dioxide Uptake: Global Distribution and Covariation with Climate, Science, 329, 834–838, https://doi.org/10.1126/science.1184984, 2010. a

Besnard, S., Carvalhais, N., Arain, M. A., Black, A., Brede, B., Buchmann, N., Chen, J., Clevers, J. G. P. W., Dutrieux, L. P., Gans, F., Herold, M., Jung, M., Kosugi, Y., Knohl, A., Law, B. E., Paul-Limoges, E., Lohila, A., Merbold, L., Roupsard, O., Valentini, R., Wolf, S., Zhang, X., and Reichstein, M.: Memory Effects of Climate and Vegetation Affecting Net Ecosystem CO2 Fluxes in Global Forests, PLOS ONE, 14, e0211510, https://doi.org/10.1371/journal.pone.0211510, 2019. a, b, c

Bodesheim, P., Jung, M., Gans, F., Mahecha, M. D., and Reichstein, M.: Upscaled diurnal cycles of land–atmosphere fluxes: a new global half-hourly data product, Earth Syst. Sci. Data, 10, 1327–1365, https://doi.org/10.5194/essd-10-1327-2018, 2018. a

Breiman, L.: Random forests, Mach. Learn., 45, 5–32, https://doi.org/10.1023/A:1010933404324, 2001. a

Camps-Valls, G., Tuia, D., Zhu, X. X., and Reichstein, M.: Deep Learning for the Earth Sciences: A Comprehensive Approach to Remote Sensing, Climate Science and Geosciences, 1st edn., Wiley, Hoboken, NJ, ISBN 978-1-119-64614-3, 2021. a

Caruana, R.: Multitask Learning, Mach. Learn., 28, 41–75, https://doi.org/10.1023/A:1007379606734, 1997. a

Chen, T. and Guestrin, C.: XGBoost: A Scalable Tree Boosting System, in: Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, KDD'16, San Francisco, USA, 13–17 August 2016, Association for Computing Machinery, New York, NY, USA, ISBN 978-1-4503-4232-2, 785–794, https://doi.org/10.1145/2939672.2939785, 2016. a

Drought 2018 Team and ICOS Ecosystem Thematic Centre: Drought-2018 Ecosystem Eddy Covariance Flux Product for 52 Stations in FLUXNET-Archive Format, ICOS [data set], https://doi.org/10.18160/YVR0-4898, 2020. a

ElGhawi, R., Kraft, B., Reimers, C., Reichstein, M., Körner, M., Gentine, P., and Winkler, A. J.: Hybrid Modeling of Evapotranspiration: Inferring Stomatal and Aerodynamic Resistances Using Combined Physics-Based and Machine Learning, Environ. Res. Lett., 18, 034039, https://doi.org/10.1088/1748-9326/acbbe0, 2023. a, b

FLUXCOM: FLUXCOM Global Energy and Carbon Fluxes, FLUXCOM, https://fluxcom.org/ (last access: 5 August 2025), 2017. a

Gao, B.-c.: NDWI – A Normalized Difference Water Index for Remote Sensing of Vegetation Liquid Water from Space, Remote Sens. Environ., 58, 257–266, https://doi.org/10.1016/S0034-4257(96)00067-3, 1996. a

Grusson, Y. and Barron, J.: Challenges in reanalysis products to assess extreme weather impacts on agriculture: Study case in southern Sweden, PLOS climate, 1, e0000063, https://doi.org/10.1371/journal.pclm.0000063, 2022. a

He, H., Queen, O., Koker, T., Cuevas, C., Tsiligkaridis, T., and Zitnik, M.: Domain Adaptation for Time Series Under Feature and Label Shifts, in: Proceedings of the 40th International Conference on Machine Learning, Honolulu, Hawaii, USA, 23–29 July 2023, edited by: Krause, A., Brunskill, E., Cho, K., Engelhardt, B., Sabato, S., and Scarlett, J., vol. 202 of Proceedings of Machine Learning Research, PMLR, 12746–12774, https://proceedings.mlr.press/v202/he23b.html (last access: 5 August 2025), 2023. a

Hersbach, H., Bell, B., Berrisford, P., Hirahara, S., Horányi, A., Muñoz-Sabater, J., Nicolas, J., Peubey, C., Radu, R., Schepers, D., Simmons, A., Soci, C., Abdalla, S., Abellan, X., Balsamo, G., Bechtold, P., Biavati, G., Bidlot, J., Bonavita, M., Chiara, G. D., Dahlgren, P., Dee, D., Diamantakis, M., Dragani, R., Flemming, J., Forbes, R., Fuentes, M., Geer, A., Haimberger, L., Healy, S., Hogan, R. J., Hólm, E., Janisková, M., Keeley, S., Laloyaux, P., Lopez, P., Lupu, C., Radnoti, G., de Rosnay, P., Rozum, I., Vamborg, F., Villaume, S., and Thépaut, J.-N.: The ERA5 Global Reanalysis, Q. J. Roy. Meteor. Soc., 146, 1999–2049, https://doi.org/10.1002/qj.3803, 2020. a, b, c

Hochreiter, S. and Schmidhuber, J.: Long Short-Term Memory, Neural Comput., 9, 1735–1780, https://doi.org/10.1162/neco.1997.9.8.1735, 1997. a

Huete, A., Didan, K., Miura, T., Rodriguez, E. P., Gao, X., and Ferreira, L. G.: Overview of the Radiometric and Biophysical Performance of the MODIS Vegetation Indices, Remote Sens. Environ., 83, 195–213, https://doi.org/10.1016/S0034-4257(02)00096-2, 2002. a, b, c

Jung, M., Reichstein, M., and Bondeau, A.: Towards global empirical upscaling of FLUXNET eddy covariance observations: validation of a model tree ensemble approach using a biosphere model, Biogeosciences, 6, 2001–2013, https://doi.org/10.5194/bg-6-2001-2009, 2009. a

Jung, M., Reichstein, M., Ciais, P., Seneviratne, S. I., Sheffield, J., Goulden, M. L., Bonan, G., Cescatti, A., Chen, J., de Jeu, R., Dolman, A. J., Eugster, W., Gerten, D., Gianelle, D., Gobron, N., Heinke, J., Kimball, J., Law, B. E., Montagnani, L., Mu, Q., Mueller, B., Oleson, K., Papale, D., Richardson, A. D., Roupsard, O., Running, S., Tomelleri, E., Viovy, N., Weber, U., Williams, C., Wood, E., Zaehle, S., and Zhang, K.: Recent Decline in the Global Land Evapotranspiration Trend Due to Limited Moisture Supply, Nature, 467, 951–954, https://doi.org/10.1038/nature09396, 2010. a, b

Jung, M., Reichstein, M., Margolis, H. A., Cescatti, A., Richardson, A. D., Arain, M. A., Arneth, A., Bernhofer, C., Bonal, D., Chen, J., Gianelle, D., Gobron, N., Kiely, G., Kutsch, W., Lasslop, G., Law, B. E., Lindroth, A., Merbold, L., Montagnani, L., Moors, E. J., Papale, D., Sottocornola, M., Vaccari, F., and Williams, C.: Global Patterns of Land-Atmosphere Fluxes of Carbon Dioxide, Latent Heat, and Sensible Heat Derived from Eddy Covariance, Satellite, and Meteorological Observations, J. Geophys. Res.-Biogeo., 116, G00J07, https://doi.org/10.1029/2010JG001566, 2011. a, b

Jung, M., Reichstein, M., Schwalm, C. R., Huntingford, C., Sitch, S., Ahlström, A., Arneth, A., Camps-Valls, G., Ciais, P., Friedlingstein, P., Gans, F., Ichii, K., Jain, A. K., Kato, E., Papale, D., Poulter, B., Raduly, B., Rödenbeck, C., Tramontana, G., Viovy, N., Wang, Y.-P., Weber, U., Zaehle, S., and Zeng, N.: Compensatory Water Effects Link Yearly Global Land CO2 Sink Changes to Temperature, Nature, 541, 516–520, https://doi.org/10.1038/nature20780, 2017. a

Jung, M., Koirala, S., Weber, U., Ichii, K., Gans, F., Camps-Valls, G., Papale, D., Schwalm, C., Tramontana, G., and Reichstein, M.: The FLUXCOM Ensemble of Global Land-Atmosphere Energy Fluxes, Scientific Data, 6, 1–14, https://doi.org/10.1038/s41597-019-0076-8, 2019. a, b, c, d

Jung, M., Schwalm, C., Migliavacca, M., Walther, S., Camps-Valls, G., Koirala, S., Anthoni, P., Besnard, S., Bodesheim, P., Carvalhais, N., Chevallier, F., Gans, F., Goll, D. S., Haverd, V., Köhler, P., Ichii, K., Jain, A. K., Liu, J., Lombardozzi, D., Nabel, J. E. M. S., Nelson, J. A., O'Sullivan, M., Pallandt, M., Papale, D., Peters, W., Pongratz, J., Rödenbeck, C., Sitch, S., Tramontana, G., Walker, A., Weber, U., and Reichstein, M.: Scaling carbon fluxes from eddy covariance sites to globe: synthesis and evaluation of the FLUXCOM approach, Biogeosciences, 17, 1343–1365, https://doi.org/10.5194/bg-17-1343-2020, 2020. a

Jung, M., Nelson, J., Migliavacca, M., El-Madany, T., Papale, D., Reichstein, M., Walther, S., and Wutzler, T.: Technical note: Flagging inconsistencies in flux tower data, Biogeosciences, 21, 1827–1846, https://doi.org/10.5194/bg-21-1827-2024, 2024. a, b

Karpatne, A., Ebert-Uphoff, I., Ravela, S., Babaie, H. A., and Kumar, V.: Machine Learning for the Geosciences: Challenges and Opportunities, IEEE T. Knowl. Data Eng., 31, 1544–1554, https://doi.org/10.1109/TKDE.2018.2861006, 2019. a

Kattge, J., Díaz, S., Lavorel, S., Prentice, I. C., Leadley, P., Bönisch, G., Garnier, E., Westoby, M., Reich, P. B., Wright, I. J., Cornelissen, J. H. C., Violle, C., Harrison, S. P., Van BODEGOM, P. M., Reichstein, M., Enquist, B. J., Soudzilovskaia, N. A., Ackerly, D. D., Anand, M., Atkin, O., Bahn, M., Baker, T. R., Baldocchi, D., Bekker, R., Blanco, C. C., Blonder, B., Bond, W. J., Bradstock, R., Bunker, D. E., Casanoves, F., Cavender-Bares, J., Chambers, J. Q., Chapin Iii, F. S., Chave, J., Coomes, D., Cornwell, W. K., Craine, J. M., Dobrin, B. H., Duarte, L., Durka, W., Elser, J., Esser, G., Estiarte, M., Fagan, W. F., Fang, J., Fernández-Méndez, F., Fidelis, A., Finegan, B., Flores, O., Ford, H., Frank, D., Freschet, G. T., Fyllas, N. M., Gallagher, R. V., Green, W. A., Gutierrez, A. G., Hickler, T., Higgins, S. I., Hodgson, J. G., Jalili, A., Jansen, S., Joly, C. A., Kerkhoff, A. J., Kirkup, D., Kitajima, K., Kleyer, M., Klotz, S., Knops, J. M. H., Kramer, K., Kühn, I., Kurokawa, H., Laughlin, D., Lee, T. D., Leishman, M., Lens, F., Lenz, T., Lewis, S. L., Lloyd, J., Llusià, J., Louault, F., Ma, S., Mahecha, M. D., Manning, P., Massad, T., Medlyn, B. E., Messier, J., Moles, A. T., Müller, S. C., Nadrowski, K., Naeem, S., Niinemets, Ü., Nöllert, S., Nüske, A., Ogaya, R., Oleksyn, J., Onipchenko, V. G., Onoda, Y., Ordoñez, J., Overbeck, G., Ozinga, W. A., Patiño, S., Paula, S., Pausas, J. G., Peñuelas, J., Phillips, O. L., Pillar, V., Poorter, H., Poorter, L., Poschlod, P., Prinzing, A., Proulx, R., Rammig, A., Reinsch, S., Reu, B., Sack, L., Salgado-Negret, B., Sardans, J., Shiodera, S., Shipley, B., Siefert, A., Sosinski, E., Soussana, J.-F., Swaine, E., Swenson, N., Thompson, K., Thornton, P., Waldram, M., Weiher, E., White, M., White, S., Wright, S. J., Yguel, B., Zaehle, S., Zanne, A. E., and Wirth, C.: TRY – a Global Database of Plant Traits, Glob. Change Biol., 17, 2905–2935, https://doi.org/10.1111/j.1365-2486.2011.02451.x, 2011. a

Klemmer, K., Rolf, E., Robinson, C., Mackey, L., and Rußwurm, M.: SatCLIP: Global, General-Purpose Location Embeddings with Satellite Imagery, Proceedings from 39th Annual AAAI Conference on Artificial Intelligence, AAAI 2025, Philadelphia, USA, 25 February–4 March 2025, 4347–4355, https://doi.org/10.1609/aaai.v39i4.32457, 2025. a

Kraft, B., Jung, M., Körner, M., Requena Mesa, C., Cortés, J., and Reichstein, M.: Identifying Dynamic Memory Effects on Vegetation State Using Recurrent Neural Networks, Frontiers in Big Data, 2, 1–14, https://doi.org/10.3389/fdata.2019.00031, 2019. a, b

Kraft, B., Besnard, S., and Koirala, S.: Emulating Ecological Memory with Recurrent Neural Networks, in: Deep Learning for the Earth Sciences, Chap. 18, John Wiley & Sons, Ltd, ISBN 978-1-119-64618-1, 269–281, https://doi.org/10.1002/9781119646181.ch18, 2021. a

Kraft, B., Jung, M., Körner, M., Koirala, S., and Reichstein, M.: Towards hybrid modeling of the global hydrological cycle, Hydrol. Earth Syst. Sci., 26, 1579–1614, https://doi.org/10.5194/hess-26-1579-2022, 2022. a

Kraft, B., Schirmer, M., Aeberhard, W. H., Zappa, M., Seneviratne, S. I., and Gudmundsson, L.: CH-RUN: a deep-learning-based spatially contiguous runoff reconstruction for Switzerland, Hydrol. Earth Syst. Sci., 29, 1061–1082, https://doi.org/10.5194/hess-29-1061-2025, 2025. a

Lipton, Z. C., Berkowitz, J., and Elkan, C.: A Critical Review of Recurrent Neural Networks for Sequence Learning, arXiv [preprint], https://doi.org/10.48550/arXiv.1506.00019, 29 June 2015. a

Martens, B., Miralles, D. G., Lievens, H., van der Schalie, R., de Jeu, R. A. M., Fernández-Prieto, D., Beck, H. E., Dorigo, W. A., and Verhoest, N. E. C.: GLEAM v3: satellite-based land evaporation and root-zone soil moisture, Geosci. Model Dev., 10, 1903–1925, https://doi.org/10.5194/gmd-10-1903-2017, 2017. a, b

Mauder, M., Jung, M., Stoy, P., Nelson, J., and Wanner, L.: Energy balance closure at FLUXNET sites revisited, Agr. Forest Meteorol., 358, 110235, https://doi.org/10.1016/j.agrformet.2024.110235, 2024. a

Migliavacca, M., Sonnentag, O., Keenan, T. F., Cescatti, A., O'Keefe, J., and Richardson, A. D.: On the uncertainty of phenological responses to climate change, and implications for a terrestrial biosphere model, Biogeosciences, 9, 2063–2083, https://doi.org/10.5194/bg-9-2063-2012, 2012. a

Nakagawa, R., Chau, M., Calzaretta, J., Keenan, T., Vahabi, P., Todeschini, A., Bassiouni, M., and Kang, Y.: Upscaling Global Hourly GPP with Temporal Fusion Transformer (TFT), arXiv [preprint], https://doi.org/10.48550/arXiv.2306.13815, 23 June 2023. a, b

Nash, J. E. and Sutcliffe, J. V.: River flow forecasting through conceptual models part I – A discussion of principles, J. Hydrol., 10, 282–290, https://doi.org/10.1016/0022-1694(70)90255-6, 1970. a

Nelson, J. A., Carvalhais, N., Cuntz, M., Delpierre, N., Knauer, J., Ogée, J., Migliavacca, M., Reichstein, M., and Jung, M.: Coupling Water and Carbon Fluxes to Constrain Estimates of Transpiration: The TEA Algorithm, J. Geophys. Res.-Biogeo., 123, 3617–3632, https://doi.org/10.1029/2018JG004727, 2018. a

Nelson, J. A., Pérez-Priego, O., Zhou, S., Poyatos, R., Zhang, Y., Blanken, P. D., Gimeno, T. E., Wohlfahrt, G., Desai, A. R., Gioli, B., Limousin, J.-M., Bonal, D., Paul-Limoges, E., Scott, R. L., Varlagin, A., Fuchs, K., Montagnani, L., Wolf, S., Delpierre, N., Berveiller, D., Gharun, M., Marchesini, L. B., Gianelle, D., Šigut, L., Mammarella, I., Siebicke, L., Black, T. A., Knohl, A., Hörtnagl, L., Magliulo, V., Besnard, S., Weber, U., Carvalhais, N., Migliavacca, M., Reichstein, M., and Jung, M.: Ecosystem Transpiration and Evaporation: Insights from Three Water Flux Partitioning Methods across FLUXNET Sites, Glob. Change Biol., 26, 6916–6930, https://doi.org/10.1111/gcb.15314, 2020. a

Nelson, J. A., Walther, S., Gans, F., Kraft, B., Weber, U., Novick, K., Buchmann, N., Migliavacca, M., Wohlfahrt, G., Šigut, L., Ibrom, A., Papale, D., Göckede, M., Duveiller, G., Knohl, A., Hörtnagl, L., Scott, R. L., Dušek, J., Zhang, W., Hamdi, Z. M., Reichstein, M., Aranda-Barranco, S., Ardö, J., Op de Beeck, M., Billesbach, D., Bowling, D., Bracho, R., Brümmer, C., Camps-Valls, G., Chen, S., Cleverly, J. R., Desai, A., Dong, G., El-Madany, T. S., Euskirchen, E. S., Feigenwinter, I., Galvagno, M., Gerosa, G. A., Gielen, B., Goded, I., Goslee, S., Gough, C. M., Heinesch, B., Ichii, K., Jackowicz-Korczynski, M. A., Klosterhalfen, A., Knox, S., Kobayashi, H., Kohonen, K.-M., Korkiakoski, M., Mammarella, I., Gharun, M., Marzuoli, R., Matamala, R., Metzger, S., Montagnani, L., Nicolini, G., O'Halloran, T., Ourcival, J.-M., Peichl, M., Pendall, E., Ruiz Reverter, B., Roland, M., Sabbatini, S., Sachs, T., Schmidt, M., Schwalm, C. R., Shekhar, A., Silberstein, R., Silveira, M. L., Spano, D., Tagesson, T., Tramontana, G., Trotta, C., Turco, F., Vesala, T., Vincke, C., Vitale, D., Vivoni, E. R., Wang, Y., Woodgate, W., Yepez, E. A., Zhang, J., Zona, D., and Jung, M.: X-BASE: the first terrestrial carbon and water flux products from an extended data-driven scaling framework, FLUXCOM-X, Biogeosciences, 21, 5079–5115, https://doi.org/10.5194/bg-21-5079-2024, 2024. a, b, c, d, e, f, g, h, i

Ogle, K., Barber, J. J., Barron-Gafford, G. A., Bentley, L. P., Young, J. M., Huxman, T. E., Loik, M. E., and Tissue, D. T.: Quantifying Ecological Memory in Plant and Ecosystem Processes, Ecol. Lett., 18, 221–235, https://doi.org/10.1111/ele.12399, 2015. a

Oord, A. v. d., Dieleman, S., Zen, H., Simonyan, K., Vinyals, O., Graves, A., Kalchbrenner, N., Senior, A., and Kavukcuoglu, K.: Wavenet: A generative model for raw audio, arXiv [preprint], https://doi.org/10.48550/arXiv.1609.03499, 12 September 2016. a

Pan, S., Pan, N., Tian, H., Friedlingstein, P., Sitch, S., Shi, H., Arora, V. K., Haverd, V., Jain, A. K., Kato, E., Lienert, S., Lombardozzi, D., Nabel, J. E. M. S., Ottlé, C., Poulter, B., Zaehle, S., and Running, S. W.: Evaluation of global terrestrial evapotranspiration using state-of-the-art approaches in remote sensing, machine learning and land surface modeling, Hydrol. Earth Syst. Sci., 24, 1485–1509, https://doi.org/10.5194/hess-24-1485-2020, 2020. a

Pan, S. J. and Yang, Q.: A Survey on Transfer Learning, IEEE T. Knowl. Data Eng., 22, 1345–1359, https://doi.org/10.1109/TKDE.2009.191, 2010. a

Parker, W. S.: Reanalyses and Observations: What's the Difference?, B. Am. Meteorol. Soc., 97, 1565–1572, https://doi.org/10.1175/BAMS-D-14-00226.1, 2016. a

Pastorello, G., Trotta, C., Canfora, E., Chu, H., Christianson, D., Cheah, Y.-W., Poindexter, C., Chen, J., Elbashandy, A., Humphrey, M., Isaac, P., Polidori, D., Reichstein, M., Ribeca, A., van Ingen, C., Vuichard, N., Zhang, L., Amiro, B., Ammann, C., Arain, M. A., Ardö, J., Arkebauer, T., Arndt, S. K., Arriga, N., Aubinet, M., Aurela, M., Baldocchi, D., Barr, A., Beamesderfer, E., Marchesini, L. B., Bergeron, O., Beringer, J., Bernhofer, C., Berveiller, D., Billesbach, D., Black, T. A., Blanken, P. D., Bohrer, G., Boike, J., Bolstad, P. V., Bonal, D., Bonnefond, J.-M., Bowling, D. R., Bracho, R., Brodeur, J., Brümmer, C., Buchmann, N., Burban, B., Burns, S. P., Buysse, P., Cale, P., Cavagna, M., Cellier, P., Chen, S., Chini, I., Christensen, T. R., Cleverly, J., Collalti, A., Consalvo, C., Cook, B. D., Cook, D., Coursolle, C., Cremonese, E., Curtis, P. S., D'Andrea, E., da Rocha, H., Dai, X., Davis, K. J., Cinti, B. D., de Grandcourt, A., Ligne, A. D., De Oliveira, R. C., Delpierre, N., Desai, A. R., Di Bella, C. M., di Tommasi, P., Dolman, H., Domingo, F., Dong, G., Dore, S., Duce, P., Dufrêne, E., Dunn, A., Dušek, J., Eamus, D., Eichelmann, U., ElKhidir, H. A. M., Eugster, W., Ewenz, C. M., Ewers, B., Famulari, D., Fares, S., Feigenwinter, I., Feitz, A., Fensholt, R., Filippa, G., Fischer, M., Frank, J., Galvagno, M., Gharun, M., Gianelle, D., Gielen, B., Gioli, B., Gitelson, A., Goded, I., Goeckede, M., Goldstein, A. H., Gough, C. M., Goulden, M. L., Graf, A., Griebel, A., Gruening, C., Grünwald, T., Hammerle, A., Han, S., Han, X., Hansen, B. U., Hanson, C., Hatakka, J., He, Y., Hehn, M., Heinesch, B., Hinko-Najera, N., Hörtnagl, L., Hutley, L., Ibrom, A., Ikawa, H., Jackowicz-Korczynski, M., Janouš, D., Jans, W., Jassal, R., Jiang, S., Kato, T., Khomik, M., Klatt, J., Knohl, A., Knox, S., Kobayashi, H., Koerber, G., Kolle, O., Kosugi, Y., Kotani, A., Kowalski, A., Kruijt, B., Kurbatova, J., Kutsch, W. L., Kwon, H., Launiainen, S., Laurila, T., Law, B., Leuning, R., Li, Y., Liddell, M., Limousin, J.-M., Lion, M., Liska, A. J., Lohila, A., López-Ballesteros, A., López-Blanco, E., Loubet, B., Loustau, D., Lucas-Moffat, A., Lüers, J., Ma, S., Macfarlane, C., Magliulo, V., Maier, R., Mammarella, I., Manca, G., Marcolla, B., Margolis, H. A., Marras, S., Massman, W., Mastepanov, M., Matamala, R., Matthes, J. H., Mazzenga, F., McCaughey, H., McHugh, I., McMillan, A. M. S., Merbold, L., Meyer, W., Meyers, T., Miller, S. D., Minerbi, S., Moderow, U., Monson, R. K., Montagnani, L., Moore, C. E., Moors, E., Moreaux, V., Moureaux, C., Munger, J. W., Nakai, T., Neirynck, J., Nesic, Z., Nicolini, G., Noormets, A., Northwood, M., Nosetto, M., Nouvellon, Y., Novick, K., Oechel, W., Olesen, J. E., Ourcival, J.-M., Papuga, S. A., Parmentier, F.-J., Paul-Limoges, E., Pavelka, M., Peichl, M., Pendall, E., Phillips, R. P., Pilegaard, K., Pirk, N., Posse, G., Powell, T., Prasse, H., Prober, S. M., Rambal, S., Rannik, Ü., Raz-Yaseef, N., Rebmann, C., Reed, D., de Dios, V. R., Restrepo-Coupe, N., Reverter, B. R., Roland, M., Sabbatini, S., Sachs, T., Saleska, S. R., Sánchez-Cañete, E. P., Sanchez-Mejia, Z. M., Schmid, H. P., Schmidt, M., Schneider, K., Schrader, F., Schroder, I., Scott, R. L., Sedlák, P., Serrano-Ortíz, P., Shao, C., Shi, P., Shironya, I., Siebicke, L., Šigut, L., Silberstein, R., Sirca, C., Spano, D., Steinbrecher, R., Stevens, R. M., Sturtevant, C., Suyker, A., Tagesson, T., Takanashi, S., Tang, Y., Tapper, N., Thom, J., Tomassucci, M., Tuovinen, J.-P., Urbanski, S., Valentini, R., van der Molen, M., van Gorsel, E., van Huissteden, K., Varlagin, A., Verfaillie, J., Vesala, T., Vincke, C., Vitale, D., Vygodskaya, N., Walker, J. P., Walter-Shea, E., Wang, H., Weber, R., Westermann, S., Wille, C., Wofsy, S., Wohlfahrt, G., Wolf, S., Woodgate, W., Li, Y., Zampedri, R., Zhang, J., Zhou, G., Zona, D., Agarwal, D., Biraud, S., Torn, M., and Papale, D.: The FLUXNET2015 Dataset and the ONEFlux Processing Pipeline for Eddy Covariance Data, Scientific Data, 7, 225, https://doi.org/10.1038/s41597-020-0534-3, 2020. a, b

Paszke, A., Gross, S., Massa, F., Lerer, A., Bradbury, J., Chanan, G., Killeen, T., Lin, Z., Gimelshein, N., Antiga, L., Desmaison, A., Kopf, A., Yang, E., DeVito, Z., Raison, M., Tejani, A., Chilamkurthy, S., Steiner, B., Fang, L., Bai, J., and Chintala, S.: PyTorch: An Imperative Style, High-Performance Deep Learning Library, in: Advances in Neural Information Processing Systems, vol. 32, edited by: Wallach, H., Larochelle, H., Beygelzimer, A., d'Alché-Buc, F., Fox, E., and Garnett, R., 8024–8035, Curran Associates, Inc., http://papers.neurips.cc/paper/9015-pytorch-an-imperative-style-high-performance-deep-learning-library.pdf (last access: 7 August 2025), 2019. a

Reichstein, M., Bahn, M., Mahecha, M. D., Kattge, J., and Baldocchi, D. D.: Linking Plant and Ecosystem Functional Biogeography, P. Natl. Acad. Sci. USA, 111, 13697–13702, https://doi.org/10.1073/pnas.1216065111, 2014. a

Reichstein, M., Besnard, S., Carvalhais, N., Gans, F., Jung, M., Kraft, B., and Mahecha, M.: Modelling Landsurface Time-Series with Recurrent Neural Nets, in: IGARSS 2018–2018 IEEE International Geoscience and Remote Sensing Symposium, 22–27 July 2018, Valencia, Spain, 7640–7643, https://doi.org/10.1109/IGARSS.2018.8518007, 2018. a

Reichstein, M., Camps-Valls, G., Stevens, B., Jung, M., Denzler, J., Carvalhais, N., and Prabhat: Deep Learning and Process Understanding for Data-Driven Earth System Science, Nature, 566, 195–204, https://doi.org/10.1038/s41586-019-0912-1, 2019. a, b

Reichstein, M., Ahrens, B., Kraft, B., Camps-Valls, G., Carvalhais, N., Gans, F., Gentine, P., and Winkler, A. J.: Combining system modeling and machine learning into hybrid ecosystem modeling, in: Knowledge Guided Machine Learning, Chapman and Hall/CRC, 327–352, ISBN 9781003143376, 2022. a

Rußwurm, M. and Körner, M.: Temporal Vegetation Modelling Using Long Short-Term Memory Networks for Crop Identification from Medium-Resolution Multi-spectral Satellite Images, in: 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), 21–26 July 2017, Honolulu, HI, USA, IEEE Computer Society, ISBN 978-1-5386-0733-6, 1496–1504, https://doi.org/10.1109/CVPRW.2017.193, 2017. a

Schaaf, C. and Wang, Z.: MCD43A2 MODIS/Terra+Aqua BRDF/Albedo Quality Daily L3 Global – 500m V006, NASA EOSDIS Land Processes Distributed Active Archive Center [data set], https://doi.org/10.5067/modis/mcd43a2.006, 2015a. a

Schaaf, C. and Wang, Z.: MCD43A4 MODIS/Terra+Aqua BRDF/Albedo Nadir BRDF Adjusted Ref Daily L3 Global – 500m V006, NASA EOSDIS Land Processes Distributed Active Archive Center [data set], https://doi.org/10.5067/modis/mcd43a4.006, 2015b. a

Schaaf, C. and Wang, Z.: MCD43C4 MODIS/Terra+Aqua BRDF/Albedo Nadir BRDF-Adjusted Ref Daily L3 Global 0.05Deg CMG V006, NASA EOSDIS Land Processes Distributed Active Archive Center [data set], https://doi.org/10.5067/modis/mcd43c4.006, 2015c. 

Tramontana, G., Jung, M., Schwalm, C. R., Ichii, K., Camps-Valls, G., Ráduly, B., Reichstein, M., Arain, M. A., Cescatti, A., Kiely, G., Merbold, L., Serrano-Ortiz, P., Sickert, S., Wolf, S., and Papale, D.: Predicting carbon dioxide and energy fluxes across global FLUXNET sites with regression algorithms, Biogeosciences, 13, 4291–4313, https://doi.org/10.5194/bg-13-4291-2016, 2016. a, b

Upton, S., Reichstein, M., Gans, F., Peters, W., Kraft, B., and Bastos, A.: Constraining biospheric carbon dioxide fluxes by combined top-down and bottom-up approaches, Atmos. Chem. Phys., 24, 2555–2582, https://doi.org/10.5194/acp-24-2555-2024, 2024. a

Valmassoi, A., Keller, J. D., Kleist, D. T., English, S., Ahrens, B., Ďurán, I. B., Bauernschubert, E., Bosilovich, M. G., Fujiwara, M., Hersbach, H., Lei, L., Löhnert, U., Mamnun, N., Martin, C. R., Moore, A., Niermann, D., Ruiz, J. J., and Scheck, L.: Current Challenges and Future Directions in Data Assimilation and Reanalysis, B. Am. Meteor. Soc., 104, E756–E767, https://doi.org/10.1175/BAMS-D-21-0331.1, 2023. a

Van Houdt, G., Mosquera, C., and Nápoles, G.: A Review on the Long Short-Term Memory Model, Artif. Intell. Rev., 53, 5929–5955, https://doi.org/10.1007/s10462-020-09838-1, 2020. a

Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A. N., Kaiser, Ł., and Polosukhin, I.: Attention Is All You Need, in: Advances in Neural Information Processing Systems, Proceedings of the 31st International Conference on Neural Information Processing Systems, NIPS'17, 5–7 December 2017, Long Beach, California, USA, Curran Associates Inc., 6000–6010, ISBN 9781510860964, 2017. a

Walther, S., Besnard, S., Nelson, J. A., El-Madany, T. S., Migliavacca, M., Weber, U., Carvalhais, N., Ermida, S. L., Brümmer, C., Schrader, F., Prokushkin, A. S., Panov, A. V., and Jung, M.: Technical note: A view from space on global flux towers by MODIS and Landsat: the FluxnetEO data set, Biogeosciences, 19, 2805–2840, https://doi.org/10.5194/bg-19-2805-2022, 2022. a

Wan, Z., Hook, S., and Hulley, G.: MOD11A1 MODIS/Terra Land Surface Temperature/Emissivity Daily L3 Global 1km SIN Grid V006, NASA EOSDIS Land Processes Distributed Active Archive Center [data set], https://doi.org/10.5067/modis/mod11a1.006, 2015a.  a, b

Wan, Z., Hook, S., and Hulley, G.: MOD11C1 MODIS/Terra Land Surface Temperature/Emissivity Daily L3 Global 0.05Deg CMG V006, NASA EOSDIS Land Processes Distributed Active Archive Center [data set], https://doi.org/10.5067/modis/mod11c1.006, 2015b. a

Warm Winter 2020 Team and ICOS Ecosystem Thematic Centre: Warm Winter 2020 Ecosystem Eddy Covariance Flux Product for 73 Stations in FLUXNET-Archive Format – Release 2022-1, ICOS [data set], https://doi.org/10.18160/2G60-ZHAK, 2022. a

Xiao, J., Ollinger, S. V., Frolking, S., Hurtt, G. C., Hollinger, D. Y., Davis, K. J., Pan, Y., Zhang, X., Deng, F., Chen, J., Baldocchi, D. D., Law, B. E., Arain, M. A., Desai, A. R., Richardson, A. D., Sun, G., Amiro, B., Margolis, H., Gu, L., Scott, R. L., Blanken, P. D., and Suyker, A. E.: Data-Driven Diagnostics of Terrestrial Carbon Dynamics over North America, Agr. Forest Meteorol., 197, 142–157, https://doi.org/10.1016/j.agrformet.2014.06.013, 2014. a

Xu, T., Guo, Z., Liu, S., He, X., Meng, Y., Xu, Z., Xia, Y., Xiao, J., Zhang, Y., Ma, Y., and Song, L.: Evaluating Different Machine Learning Methods for Upscaling Evapotranspiration from Flux Towers to the Regional Scale, J. Geophys. Res.-Atmos., 123, 8674–8690, https://doi.org/10.1029/2018JD028447, 2018. a

Zhang, W., Nelson, J. A., Miralles, D. G., Mauder, M., Migliavacca, M., Poyatos, R., Reichstein, M., and Jung, M.: A New Post-Hoc Method to Reduce the Energy Imbalance in Eddy Covariance Measurements, Geophys. Res. Lett., 51, e2023GL107084, https://doi.org/10.1029/2023GL107084, 2024. a

Zhao, W. L., Gentine, P., Reichstein, M., Zhang, Y., Zhou, S., Wen, Y., Lin, C., Li, X., and Qiu, G. Y.: Physics-Constrained Machine Learning of Evapotranspiration, Geophys. Res. Lett., 46, 14496–14507, https://doi.org/10.1029/2019GL085291, 2019. a, b

Zhu, S., Quaife, T., and Hill, T.: Uniform upscaling techniques for eddy covariance FLUXes (UFLUX), Int. J. Remote Sens., 45, 1450–1476, https://doi.org/10.1080/01431161.2024.2312266, 2024. a

Download
Short summary
This study evaluates machine learning approaches for upscaling evapotranspiration from the site to the global scale. Sequential models capture temporal dynamics better, especially with precipitation data, but all models show biases in data-scarce regions. Improved upscaling requires richer training data, informed covariate selection, and physical constraints to enhance robustness and reduce extrapolation errors.
Share
Altmetrics
Final-revised paper
Preprint