Articles | Volume 29, issue 18
https://doi.org/10.5194/hess-29-4585-2025
© Author(s) 2025. This work is distributed under
the Creative Commons Attribution 4.0 License.Evaluation of high-intensity rainfall observations from personal weather stations in the Netherlands
Download
- Final revised paper (published on 24 Sep 2025)
- Preprint (discussion started on 08 Nov 2024)
Interactive discussion
Status: closed
Comment types: AC – author | RC – referee | CC – community | EC – editor | CEC – chief editor
| : Report abuse
-
RC1: 'Comment on egusphere-2024-3207', Anonymous Referee #1, 17 Jan 2025
- AC1: 'Reply on RC1', Nathalie Rombeek, 21 Feb 2025
-
RC2: 'Comment on egusphere-2024-3207', Anonymous Referee #2, 24 Jan 2025
- AC2: 'Reply on RC2', Nathalie Rombeek, 21 Feb 2025
Peer review completion
AR: Author's response | RR: Referee report | ED: Editor decision | EF: Editorial file upload
ED: Reconsider after major revisions (further review by editor and referees) (18 Mar 2025) by Efrat Morin

AR by Nathalie Rombeek on behalf of the Authors (29 Apr 2025)
Author's response
Author's tracked changes
Manuscript
ED: Referee Nomination & Report Request started (01 May 2025) by Efrat Morin
RR by Anonymous Referee #1 (08 May 2025)

RR by Anonymous Referee #2 (23 May 2025)
ED: Publish subject to technical corrections (14 Jun 2025) by Efrat Morin

AR by Nathalie Rombeek on behalf of the Authors (21 Jun 2025)
Author's response
Manuscript
Review report on Evaluation of high-intensity rainfall observations from personal weather stations in the Netherlands from Rombeek et al. 2024.
The authors evaluate the robustness of rainfall estimates from personal weather stations (PWSs) by comparing them to automatic weather stations in the Netherlands over six years, identifying significant underestimation of PWS rainfall, especially for extreme events. They select rain events for different aggregations and seasons and apply part of a previously published QC method. Adjustments like bias correction improved the accuracy for moderate events, but limitations persist for high-intensity rainfall, suggesting the need for dynamic calibration and additional filtering techniques. The overall quality of the manuscript and research is good and well within the scope of HESS. To enhance its depth and utility for future readers, the authors should provide a stronger justification for their choice of QC methods, or better, show some analysis in this regarding who their choice affects the results, and propose a specific bias correction factor for hourly time scales (details provided in specific comments). Otherwise I only have minor comments and would recommend the publication of this manuscript after the above issue (selected as major, but IMO a minor manjor) is addressed.
Major Comments
You already point out the importance of QC and bias correction throughout the manuscript. Therefore, my main question is, how did the choice and the parameters of the HI and FZ filter and bias correction influence the results. I miss the reasoning for the two used filters (and not using the SO filter from de Vos et al. 2019) and not the method from Bardossy et al. (2021) or other QC methods typically used for rain gauge data. I am not asking you to compare all available methods or a detailed sensitivity analysis for each parameter, but the choice of methods and parameters will have an effect over different seasons.
One example: the FZ filter discards a value if half of its neighbors are also zero and the HI filter relies on a maximum allowed factor that a station can deviate from the surrounding ones. Winter and summer precipitation might cause a different need of these parameters.
You discuss the bias correction well and it is reasonable to use a default value from a previous study. By showing the residual bias over different aggregations e.g. in Fig 8c you already indirectly give the optimal bias correction factor. You could add this as a result to the paper extending its scope a little bit. A suggestion would be to support the bias correction factor it by giving the uncertainty through a bootstrap sampling. Checking both the filtering and bias correction would allow you to further assess the robustness of the PWS rainfall rates.
I find the reasoning and structure of uncertainty factors of rain gauges in general and PWS specifically given in L 85ff to be unclear and not exhaustive. Errors for rain gauges and personal weather station are for example also undercatch due to wind, solid precipitation or evaporation, which are missing already in L39. They could be a fourth group of errors in L58ff.
Also, the phrase “in addition” in L61 after a list of three points suggests that you could add a (4) item to the list?
It would be good to have a more structured and complete list of the potential errors as they motive they choice for the quality control routine. You could even link individual QC methods to errors i.e. (3) setup and maintenance à bias correction and FZ filter
Minor comments
Abstract: You could sharpen the scope of the paper by including the content from L79 (the gap you aim to close) in it.
Introduction:
You may include following literature if you find them fitting
L29 560 ha seems very specific, could you better give a range of what is considered a small, fast reacting catchment
L52 For Netatmo I think users can decide whether data is uploaded or not? For other platforms, that is certainly the case.
L77 gives the impression that you also use the QC from Bardossy et al (2021) – which would be interesting, but might be too much here
L93 data availability was too low before, right? You could state this more specifically here
L102 You could add a statement about spatial correlation from de Beek (2012) already here to describe the area further
Fig 2. You could add IQR or min/max to the monthly barplots to give a feeling for variability
L146 Did you use a fixed window for resampling i.e. always to the full hour like XX:00 to XX:55? This could be important for the selection of events.
L189 You could add the best/worst/range for all validation metrics
L195 You mix bias and relative bias here and, in the text, please clarify
L 220 Will the
L276 with an average PWS intensity
Figure 8: Any idea why JJA and MAM seem to be very similar and DJF and SON?
L345 and L353 are a bit counter-intuitive. You want to give insight in he uncertainty, but at the same try to reduce uncertainty due to aggregation. Please clarify.
L364 to 366 Do you refer to the two highest hourly events in JJA? Maybe you could check the radar images for those two events and check the spatial distribution of rainfall during these events? Similar, looking at the 5/10 minute time series from AWS and surrounding PWS could give some insight on those two events.
Technical issues
L15 duplicated “with”
L52 “are” instead of “is”