Articles | Volume 30, issue 9
https://doi.org/10.5194/hess-30-2797-2026
© Author(s) 2026. This work is distributed under the Creative Commons Attribution 4.0 License.
Multivariate calibration can increase simulated discharge uncertainty and model equifinality
Download
- Final revised paper (published on 11 May 2026)
- Supplement to the final revised paper
- Preprint (discussion started on 09 May 2025)
- Supplement to the preprint
Interactive discussion
Status: closed
Comment types: AC – author | RC – referee | CC – community | EC – editor | CEC – chief editor
| : Report abuse
-
RC1: 'Comment on egusphere-2025-1598', Anonymous Referee #1, 02 Jun 2025
- AC2: 'Reply on RC1', Sandra Pool, 30 Aug 2025
-
RC2: 'Comment on egusphere-2025-1598', Tam Nguyen, 12 Jun 2025
- AC1: 'Reply on RC2', Sandra Pool, 30 Aug 2025
Peer review completion
AR – Author's response | RR – Referee report | ED – Editor decision | EF – Editorial file upload
ED: Publish subject to revisions (further review by editor and referees) (08 Sep 2025) by Julia Knapp
AR by Sandra Pool on behalf of the Authors (12 Mar 2026)
Author's response
Author's tracked changes
Manuscript
ED: Referee Nomination & Report Request started (24 Mar 2026) by Julia Knapp
RR by Tam Nguyen (05 Apr 2026)
RR by Anonymous Referee #1 (19 Apr 2026)
ED: Publish as is (21 Apr 2026) by Julia Knapp
AR by Sandra Pool on behalf of the Authors (23 Apr 2026)
Manuscript
I reviewed the manuscript entitled “Multivariate calibration can increase simulated discharge uncertainty and model equifinality” by Pool et al. This study employs a Monte Carlo approach (100,000 parameter sets) to compare three hydrological model calibration strategies: discharge-only, ET-only, and discharge+ET. The key finding—that incorporating ET into calibration increases flux uncertainty and discharge variability while revealing incompatible parameter spaces for discharge and ET objectives—is compelling and well-supported. However, two critical issues require resolution before publication (see below). There are also some specific comments in the annotated manuscript. Overall, I would suggest a major revision.
Major Concern: Definition of "Behavioral Parameters"
The authors defined the top 100 parameter sets as “behavioral parameters” based solely on their ranking without explicitly addressing the performance thresholds associated with these sets. For example: What NSE/KGE values correspond to the top 100 sets? Does the performance of the 100th set differ significantly from the 101st? Could expanding the threshold to the top 1,000 sets yield comparable or better performance while reducing equifinality? Since the “behavioral” classification directly shapes the conclusions (e.g., parameter space in compatibility in Section5.2), this ambiguity undermines the practical relevance of the results. Therefore, I would suggest to add performance metrics (e.g., NSE/KCE ranges) for the top 100/1,000 parameter sets in a table or figure and to justify the choice of 100 sets (e.g., via sensitivity tests or breakpoint analysis of performance vs. parameter set rank).
Moderate Concern: Parameter Distribution Analysis (Figure 4)
Figure 4 illustrates parameter distribution overlaps between calibration strategies but lacks methodological transparency: The rationale for selecting specific parameter pairs (e.g., sensitivity, correlation) is unclear. Subjective terms like “separation” need quantitative support (e.g., Kolmogorov-Smirnov tests or overlap coefficients). I would suggest to clarify how parameter pairs were chosen and to replace qualitative descriptions with metrics (e.g., heatmaps of distribution overlap or pairwise correlation matrices) to synthesize relationships across all parameters.