Articles | Volume 30, issue 1
https://doi.org/10.5194/hess-30-141-2026
© Author(s) 2026. This work is distributed under the Creative Commons Attribution 4.0 License.
Evaluation of high-resolution meteorological data products using flux tower observations across Brazil
Download
- Final revised paper (published on 13 Jan 2026)
- Supplement to the final revised paper
- Preprint (discussion started on 28 Mar 2025)
- Supplement to the preprint
Interactive discussion
Status: closed
Comment types: AC – author | RC – referee | CC – community | EC – editor | CEC – chief editor
| : Report abuse
-
RC1: 'Comment on egusphere-2025-883', Anonymous Referee #1, 14 Jun 2025
- AC1: 'Reply on RC1', Jamie Brown, 11 Jul 2025
-
RC2: 'Comment on egusphere-2025-883', Anonymous Referee #2, 14 Jun 2025
- AC2: 'Reply on RC2', Jamie Brown, 11 Jul 2025
Peer review completion
AR – Author's response | RR – Referee report | ED – Editor decision | EF – Editorial file upload
ED: Reconsider after major revisions (further review by editor and referees) (05 Aug 2025) by Daniel Viviroli
AR by Jamie Brown on behalf of the Authors (02 Oct 2025)
Author's response
Author's tracked changes
Manuscript
ED: Referee Nomination & Report Request started (17 Oct 2025) by Daniel Viviroli
RR by Anonymous Referee #2 (27 Oct 2025)
RR by Anonymous Referee #1 (26 Nov 2025)
ED: Publish subject to technical corrections (27 Nov 2025) by Daniel Viviroli
AR by Jamie Brown on behalf of the Authors (04 Dec 2025)
Author's response
Manuscript
General Comments
The manuscript presents an evaluation of five gridded meteorological products. The analysis is based on the comparison between those products and the meteorological data from 11 flux towers in Brazil. Overall, the manuscript meets general readability standards, but would benefit from professional proofreading to enhance clarity and polish. Tables and figures are well-prepared. The scientific gap and objectives are clearly stated.
Nonetheless, I have some concerns regarding the Materials (Datasets) section. First, using only 11 sites seems insufficient for a study area as large and ecologically diverse as Brazil. Looking at Figure 1, these 11 sites can be grouped into 5 or 6 clusters. For example, as shown in Table 1, K67, K77, and K83 share the same elevation and are located within 50 km of each other. The primary difference among nearby flux tower sites appears to be land cover, which spans only four types: Tropical Rainforest, Croplands, Tropical Dry Forest, and Woodland Savanna (note that, according to Cabral et al., 2015, PDG is classified as Woodland Savanna: https://www.sciencedirect.com/science/article/pii/S2214581815000440.
My point is: why limit the analysis to these specific 11 flux tower sites? In my view, excluding conventional meteorological data would make sense only if the primary aim were to assess the ability of gridded products to represent actual ET. However, based on Section 2.1 of the manuscript, "Variables were included so that reference evapotranspiration could be calculated according to the standard FAO methodology." If that is the case, then incorporating data from the Brazilian National Meteorological Institute (INMET) could significantly increase the number of observation points and improve geographic coverage.
While it's acknowledged that South America has far fewer flux towers compared to North America or Europe, other flux towers in Brazil—available in the AmeriFlux or FLUXNET networks—are not listed in Tables 1 and 2. I strongly encourage the authors to consider including towers located in underrepresented biomes, such as the Caatinga, Pantanal, and Atlantic Forest. In summary, the authors should elaborate more clearly on why only those 11 flux towers were selected.
Methodology
The 80% data availability threshold is a sound strategy. Since flux tower data are being used to evaluate gridded products, only high-quality observed data should be considered; therefore, gap filling should be avoided. It is also crucial to ensure that the datasets share common days for comparison, which appears to be the case here.
Another important point is that performance metrics derived from larger datasets are generally more reliable, with increased statistical significance and reduced uncertainty. Thus, it would be helpful to demonstrate that, despite site differences, the results remain comparable.
Regarding temporal averaging, it is not clear whether the hourly samples retrieved in each iteration were selected randomly. While the use of two-sample K-S tests is appropriate, its efficiency may vary depending on the time of day during which the data gaps occur. For instance, under a 30-minute resolution, a dataset with evenly distributed missing values (e.g., one every hour) is likely to be much smaller than a sample with missing records only at night, for example.
As for the conversion from daily to monthly averages, is using only 50% of the days sufficient? Is there evidence that this threshold does not compromise monthly estimates? For example, if most of the missing days were cloudy, the resulting monthly average could be biased toward sunnier conditions.
Finally, it is unclear why no similar approach was applied to harmonize spatial resolution among the products. Could the authors provide justification or evidence that temporal averaging is more critical than spatial aggregation in this context? Since both observations and products were resampled to a common temporal scale, aligning spatial resolution seems equally important. A comparison between Tables 3 and 4 (particularly at the daily scale) suggests that performance improves with higher spatial resolution. Pixel-to-point comparisons may introduce bias, even for high-resolution products. Therefore, I am concerned about the fairness of comparisons involving coarser-resolution datasets.
In light of the major concerns outlined above, along with the minor issues noted below, I recommend that the manuscript undergo major revisions before it can be considered for publication.
Specific Comments
Abstract
Please clarify what you mean by “downward shortwave and longwave radiation”. Does shortwave refer only to incoming radiation, and longwave to both incoming and outgoing?
Introduction
"The validation of gridded weather products is essential to ensure a fair and reliable assessment of model performance."
"For example, a 2018 study using data from 11,427 rain gauges across Brazil revealed that the Amazon Basin—which holds 70% of the country’s freshwater—has the lowest gauge density, with only 199 located in the entire state of Amazonas." Also note that the Amazon Basin holds 70% of Brazil’s freshwater, not the state of Amazonas, which may be misleading. Clarify the distinction.
"...however, comparatively less work has been done in data-poor regions like South America."
"Secondly, [...] with observational data for each variable considered."
L87: Suggested alternatives:
(i) "Finally, how do these errors vary spatially and seasonally?"
(ii) "Finally, how do these errors vary by location and season?"
Section 2 – Datasets
In subsections 2.2.1–2.2.4, key information (e.g., time span, spatial resolution) is provided for some datasets but missing for others. Although Table 3 contains these details, consistency across subsections would improve readability.
L190: Ensure numbers are separated from units (e.g., “50 km” not “50km”). Also, define SYNOP, as it is not introduced anywhere in the manuscript.
Section 3.6
L208: Clarify what “the differences” refers to—differences between which datasets, variables, or scales?
L261–264: The observed shift in bias from daily to monthly scale is interesting. Can the authors explain why this occurs?
Section 5.3
L401: Please refer to the figure or table that supports this statement.
Section 5.4
L426: Likewise, cite the relevant figure or table.
L430–433: The claim that larger samples yield more robust correlations is valid. Has such a correlation been demonstrated in the manuscript? Please refer to the relevant figure or table.