the Creative Commons Attribution 4.0 License.
the Creative Commons Attribution 4.0 License.

Non-asymptotic distributions of water extremes: much ado about what?
Francesco Serinaldi
Federico Lombardo
Chris G. Kilsby
Non-asymptotic (𝒩𝒜) probability distributions of block maxima (BM) have been proposed as an alternative to asymptotic distributions of BM derived by means of classic extreme-value theory (EVT). Their advantage should be the inclusion of moderate quantiles, as well as of extremes, in the inference procedures. This would increase the amount of information used and reduce the uncertainty characterizing the inference based on short samples of BM or peaks over high thresholds. In this study, we show that the 𝒩𝒜 distributions of BM suffer from two main drawbacks that make them of little usefulness for practical applications. Firstly, unlike classic EVT distributions, 𝒩𝒜 models of BM imply the preliminary definition of their conditional parent distributions, which explicitly appears in their expression. However, when such conditional parent distributions are known or estimated, the unconditional parent distribution is readily available, and the corresponding 𝒩𝒜 distribution of BM is no longer needed as it is just an approximation of the upper tail of the parent. Secondly, when declustering procedures are used to remove autocorrelation characterizing hydroclimatic records, 𝒩𝒜 distributions of BM devised for independent data are strongly biased even if the original process exhibits low or moderate autocorrelation. On the other hand, 𝒩𝒜 distributions of BM accounting for autocorrelation are less biased but still of little practical usefulness. Such conclusions are supported by theoretical arguments, Monte Carlo simulations, and re-analysis of sea level data.
- Article
(4739 KB) - Full-text XML
- BibTeX
- EndNote
In the last decades, the statistical analysis of hydroclimatic extremes has mainly relied on theoretical results and models developed by a branch of statistics called extreme-value theory (EVT) (Fisher and Tippett, 1928; Von Mises, 1936; Gnedenko, 1943; Jenkinson, 1955; Gumbel, 1958; Balkema and de Haan, 1974; Pickands III, 1975; Leadbetter, 1983; Smith, 1984; Davison and Smith, 1990; Coles, 2001; Beirlant et al., 2004; Salvadori et al., 2007). EVT describes the extremal behavior of observed phenomena by means of asymptotic probability distributions that are valid under certain assumptions about the parent process, such as large sample sizes n (i.e., n→∞ to guarantee asymptotic convergence), independence, and distributional identity. However, hydroclimatic records are commonly quite short and hardly ever behave as independent and identically distributed random variables. More often, hydroclimatic processes result from combinations of heterogeneous physical processes (e.g., Morrison and Smith, 2002; Smith et al., 2011, 2018) and exhibit autocorrelation (e.g., Kantelhardt et al., 2006; Wang et al., 2007; Serinaldi, 2010; Labat et al., 2011; Papalexiou et al., 2011; Serinaldi and Kilsby, 2016b; Lombardo et al., 2017; Iliopoulou et al., 2018; Markonis et al., 2018; Serinaldi and Kilsby, 2018; Serinaldi et al., 2018; Dimitriadis et al., 2021, and references therein), with their behavior being better described by stochastic processes incorporating such properties (e.g., Serinaldi and Kilsby, 2014a; Serinaldi and Lombardo, 2017a, b; Dimitriadis and Koutsoyiannis, 2018; Papalexiou, 2018; Koutsoyiannis, 2020; Papalexiou and Serinaldi, 2020; Koutsoyiannis and Dimitriadis, 2021; Papalexiou et al., 2021; Papalexiou, 2022; Serinaldi et al., 2022a; Koutsoyiannis, 2023, and references therein).
As a consequence, the lack of fulfillment of EVT assumptions affects the analysis of block maxima (BM) or over-threshold (OT) values as the BM and OT sample selection generally yields short sample sizes and does not remove the effects of autocorrelation and the possible heterogeneity of the generating mechanisms (see, e.g., Koutsoyiannis, 2004; Iliopoulou and Koutsoyiannis, 2019; Serinaldi et al., 2020b). Research in EVT has addressed these issues to some extent for the case of asymptotic and sub- or pre-asymptotic methods for the BM and OT processes (see Serinaldi et al., 2020b, and references therein for an overview).
On the other hand, parallel literature has focused on non-asymptotic (𝒩𝒜) approaches for BM, attempting to use as many observations as possible to infer the distribution of the largest values. 𝒩𝒜 distributions of BM include Todorovic distributions and their special cases (e.g., Todorovic, 1970; Todorovic and Zelenhasic, 1970; Lombardo et al., 2019), specifically the so-called metastatistical extreme value (MEV) distributions and their variants, such as the simplified MEV (SMEV; Marani and Ignaccolo, 2015; Zorzetto et al., 2016; De Michele and Avanzi, 2018; Marra et al., 2018; De Michele, 2019; Marra et al., 2019; Hosseini et al., 2020; Miniussi et al., 2020; Zorzetto and Marani, 2020).
Serinaldi et al. (2020b) explained the conceptual and analytical relationships among the above-mentioned 𝒩𝒜 distributions of BM in the context of compound distributions of order statistics and introduced compound beta-binomial distributions (βℬC) of BM in processes with stationary autocorrelation structure. βℬC distributions allow one to avoid declustering procedures required, for instance, by (S)MEV to obtain samples fulfilling the assumption of independence.
However, while the βℬC distributions allow a correct interpretation of the 𝒩𝒜 models of BM and their connections to their parent distributions, Serinaldi et al. (2020b) did not comprehensively explore the usefulness or lack thereof of 𝒩𝒜 models of BM in practical analysis. In this study, we further explore and discuss the extent of the redundancy of such models with respect to their parent distributions, as well as the actual lack of effectiveness in declustering procedures in the context of 𝒩𝒜-based analysis.
This paper falls into the class of so-called neutral (independent) validation–falsification studies (see, e.g., Popper, 1959; Boulesteix et al., 2018, and references therein), aiming to independently check the theoretical consistency in statistical methods applied in the analysis of hydroclimatic data (Lombardo et al., 2012, 2014, 2017, 2019; Serinaldi and Kilsby, 2016a; Serinaldi et al., 2015, 2018, 2020a, b, 2022b). We put emphasis on the common but misleading habit of seeking confirmation by iterating the application of a given method in relation to observed data whose generating processes are inherently unknown. In fact, if a method is technically flawed, its output will always be consistent across applications but systematically incorrect. In contrast, genuine neutral analysis calls into question the theory behind a method or model and checks it analytically and/or against challenging controlled conditions via suitable Monte Carlo simulations.
This study is organized as follows. In Sect. 2, we briefly review the main 𝒩𝒜 distributions of BM proposed in the literature and their relationship to the corresponding distributions of the parent process. Section 3 recalls the rationale for performing an extreme-value analysis and explains why the 𝒩𝒜 models of BM are conceptually redundant in this context. These aspects are further discussed in Sect. 4 using simple Monte Carlo simulations and by reanalyzing sea level data previously studied in the literature. Monte Carlo experiments in Sect. 5 investigate the performance of some 𝒩𝒜 models of BM under independence and serial dependence, as well as the effectiveness of declustering methods proposed to deal with autocorrelated time series and the reliability of some results previously reported in the literature. In Sect. 6, the problems concerning the use of 𝒩𝒜 models of BM for practical applications are placed within the wider context of a questionable approach to applied statistics in hydroclimatic studies. Conclusions are given in Sect. 7.
To support our discussion, we firstly recall some basic theoretical results, referring to Serinaldi et al. (2020b) and the references therein for more details about the analytical derivation of the equations reported below. Under the assumption of an identical probability distribution, BM are the largest-order statistics (David and Nagaraja, 2004, p. 1) of a sequence of m random variables with the same cumulative distribution function (cdf) FZ(z). If these variables are also independent, the cdf of BM Y in random samples of finite size m is
where Fℬ and Fβ are the binomial and beta cdf's, respectively. Under the assumption of serial dependence, the distribution of BM in finite-sized blocks is unknown as it depends on the m-dimensional joint distribution of the m variables forming a block (Todorovic, 1970; Todorovic and Zelenhasic, 1970). Closed-form solutions do exist for the case of Markovian processes, whereby the joint distribution is bivariate (Lombardo et al., 2019). For high-order dependence structures, the 𝒩𝒜 distribution of BM can be approximated by an extended beta–binomial distribution βℬ (Serinaldi et al., 2020b, Sect. 2.2):
where Fβℬ is the βℬ cdf; is the complete beta function (Arnold et al., 1992, pp. 12–13); and ρβℬ(z) is known as the “intra-class” or “intra-cluster” correlation, which depends on FZ(z) and the autocorrelation function (ACF) of the parent process , denoted as ρ. When the parent process Zi is serially uncorrelated (ρβℬ=0), Eq. (2) yields Eq. (1) as a particular case.
The process Z is named the “parent” as it is the stochastic process whose distribution FZ appears in the expression of the distribution of BM FY, and it could have no strict physical meaning. For example, the parent process used to build the distribution of BM for precipitation or streamflow sampled at a given timescale (e.g., daily) could be the process of observations over any threshold guaranteeing the selection of at least one observation per block. Therefore, inter-arrival times of the observations z are always smaller than or equal to the m time steps corresponding to the block size. As a limiting case, Z can obviously be the complete streamflow or rainfall process sampled at the finest timescale (e.g., daily).
As discussed in more depth in the next sections, every distribution of BM (asymptotic or non-asymptotic) provides just an approximation of the upper tail of the distribution of the parent process. Equations (1) and (2) indicate that two parent processes can have the exact same marginal distribution, but the expression of the corresponding 𝒩𝒜 model of BM approximating the upper tail of FZ might be different according to the presence or absence of serial dependence. In other words, serial dependence influences the patterns of the observations z within each block and, therefore, the sequences of BM and the form of their 𝒩𝒜 distribution FY. On the other hand, FZ is unaffected by serial dependence as it describes the distribution of Z, which does not imply any operation (aggregation, average, or BM selection) over a time window (block).
The assumption of intra- or inter-block distributional identity can be relaxed by resorting to the concept of mixed or compound distributions, which integrate (average) over the parameter space of the parent distribution under the assumption that these parameters can change within or between each block (Marra et al., 2019; Serinaldi et al., 2020b). For instance, such changes or fluctuations can reflect different physical generating mechanisms (e.g., convective and frontal weather systems generating storms in different seasons) or inter-block sampling uncertainty related to still unidentified physical processes, which therefore need a stochastic description. A general compact form of this class of models can be written as
where is the joint distribution of the parent process accounting for intra-block dependence, Ωθ is the state space of parameter vector θ, and 𝔼[⋅] is the expectation operator. Gl is integrated (averaged) over the number of observations L in each block of size m and the parameters Θ, which are treated as random variables with a joint probability density function (pdf) g(l,θ). Equation (3) is a generalization of Todorovic distributions, incorporating possible inter-block fluctuations of parameters of the joint distribution of parent process Z.
Since high-dimensional joint distributions Gl are difficult to handle and fit, the general model in Eq. (3) can be approximated by a compound version of the βℬ distribution in Eq. (2) for high-order dependence structures, resulting in the following compound βℬ model (βℬC) (Serinaldi et al., 2020b, Sect. 5.2):
where FβℬC is the βℬC cdf, ρ is the correlation function of the parent process Z, and Ωρ is its state space. Under the assumption of independence (ρ=0), the βℬ distribution is reduced to a binomial distribution (which can also be written in the form of a beta distribution), and Eq. (4) yields MEV models as special cases:
Analogously to Eqs. (1) and (2), Eqs. (4) and (5) approximate the upper tail of the distribution of the parent process Z:
This is, in itself, a compound distribution (averaged over the parameter space) and should not be confused with the conditional distributions FZ(z;θ), which depend on the parameters. FZ in Eq. (1) is also unaffected by serial correlation, which, in turn, changes the form of the corresponding 𝒩𝒜 distribution FY of BM. As mentioned above, we can have two parent processes with identical FZ and different FY depending on the presence or absence of serial dependence. Equations (4) and (5) are quite general and account not only for inter-block fluctuations via g(θ) but also for intra-block variability (such as different physical generating mechanisms and/or seasonal fluctuations acting at the intra-block scale) assuming that the conditional distributions FZ(z;θ) are compound or mixed – that is,
where g(ϑ;θ) describes the intra-block variability of ϑ (e.g., seasonal fluctuations or intra-annual weather system switching) conditioned on the inter-block status (e.g., El Niño or La Niña conditions spanning 1 or more years). Of course, g(ϑ;θ) is reduced to g(ϑ) if the intra-block fluctuations are assumed to be independent of inter-annual fluctuations. A typical example is the common assumption of year-to-year invariant seasonal patterns.
In the next sections, the models in Eqs. (1), (2), (4), and (5) are compared with the corresponding parent distributions. We stress that the models in Eqs. (4) and (5) must be compared with the corresponding compound parent distribution in Eq. (6), which accounts for the same intra- and/or inter-block variability. It is worth noting that the following discussion is fully general and valid for any 𝒩𝒜 model of BM requiring the preliminary knowledge or definition of FZ(z;θ) and its use in the expression of FY. Hereinafter, the terms “𝒩𝒜 model or distribution of BM” and “𝒩𝒜 model or distribution” are used interchangeably to denote the same class of models.
Asymptotic distributions provided by EVT are the limit distributions of 𝒩𝒜 models under some assumptions concerning the nature of the marginal distribution and dependence structure of the parent process Z. In particular, it is well known that the generalized extreme value (GEV) and generalized Pareto (GP) distributions are the general asymptotes of the distributions of BM and peaks over thresholds (POTs), respectively, under the assumptions of independence (or certain types of weak dependence) and distributional identity (see, e.g., Leadbetter et al., 1983; Coles, 2001). Therefore, EVT models are fairly general and relatively easy to apply, mainly because they do not require a precise knowledge of FZ (Leadbetter et al., 1983, p. 4), which, instead, explicitly appears in the expression of any 𝒩𝒜 model. This aspect has already been stressed in standard handbooks of applied statistics, such as that of Mood et al. (1974, p. 258), who stated (using our notation and setting L=m) “One might wonder why we should be interested in an asymptotic distribution of Y when the exact distribution, which is given by , where FZ is the c.d.f. [cumulative distribution function] sampled from, is known. The hope is that we will find an asymptotic distribution which does not depend on the sampled c.d.f. FZ. We recall that the central-limit theorem gave an asymptotic distribution for [sample mean] which did not depend on the sampled distribution even though the exact distribution of could be found.”
Bearing in mind that Z and Y are two different processes (Serinaldi et al., 2020b, Sect 3.2), the usefulness and widespread application of asymptotic EVT models of BM and POTs stems from the fact that such distributions approximate (converge to) the upper tail of the distribution of the parent process Z without needing to know FZ (under the above-mentioned assumptions) and only requiring a limited amount of information (i.e., BM and/or POT observations) instead of complete time series. This is paramount in practical applications as it allows the use of (i) a couple of general distributions (GEV and GP) supported by a theory that clearly identifies the range of the validity of such models and (ii) data that are more easy to collect and are more widely available worldwide compared to complete time series. For example, meteorological services provide most of the historical information on rainfall in terms of annual maximum values for specified durations to be used in the so-called intensity–duration–frequency (IDF) analysis. In these cases, we do not know FZ, and we cannot fit it either as the data representing the whole rainfall process – and, therefore, FZ – are not available. However, EVT states, for instance, that the GEV distribution asymptotically approximates the upper tail of FZ independently of the form of FZ (under certain constraints) based on theoretical results concerning the asymptotic behavior of . EVT distributions independent of the form of FZ are also useful when observations of Z are available, but defining a reliable model for FZ is too difficult due to complexity of the hydroclimatic process of interest and its generating mechanisms.
Unlike asymptotic models, 𝒩𝒜 distributions require the preliminary knowledge or fit of FZ, which explicitly appears in their expression. However, if we already know FZ (or if we have a good estimate of it), we no longer need any 𝒩𝒜 distribution of BM as the latter provides only an approximation of the upper tail of the known or fitted FZ. We do not even need any asymptotic model and/or, more generally, any model of BM or POTs as these are just processes extracted from the parent process Z, whose distribution FZ already describes the whole state space, including the extreme values. The use of extreme-value distributions makes sense if and only if we do not have enough information on FZ. Otherwise, the latter provides all the information needed to make statements about any quantile. In this context, only plays a functional or intermediate role in theoretical derivations to move from FZ to general asymptotic distributions independent of FZ, to be used when FZ is not available.
The same remarks hold true for any compound 𝒩𝒜 model such as βℬC and its special cases. In fact, these models require the preliminary inference of FZ to derive distributions (compound versions of ) that only approximate the upper tail of the previously estimated FZ. It is easy to understand that such a procedure makes little sense in practical applications: why should one search for an approximation of the upper tail of a distribution that is already known or fitted? The use of compound 𝒩𝒜 models is not even justified by their mixing nature, which allows for averaging inter-block fluctuations of parameters. In fact, as further discussed below, such a mixing procedure can be directly applied to FZ, thus obtaining a compound distribution of the parent process Z that can readily be used to make statements on any quantile, avoiding unnecessary 𝒩𝒜 approximations of the upper tail. This explains why 𝒩𝒜 models have not received much attention and why the recently proposed compound 𝒩𝒜 models are of little practical usefulness, if any. Their usefulness is mainly theoretical as they help explain the inherent differences between parent processes Z and BM processes Y, thus avoiding misconceptions and misinterpretations of different model outputs (see Serinaldi et al., 2020b).
While the concepts discussed in Sect. 3 should be well-known and self-evident, they seem to be systematically neglected in hydroclimatic literature dealing with 𝒩𝒜 models. Therefore, this section reports further discussion using some simple examples and real-world data re-analysis to highlight the relationship between 𝒩𝒜 models and the embedded distribution FZ, thus showing concretely how the former provide only a redundant approximation of the upper tail of the latter.
4.1 Estimation of 𝒯-year events: recalling basic concepts to avoid inconsistencies
The first example is freely inspired by the work of Mushtaq et al. (2022), who searched for an approach to select the most suitable distribution FZ of ordinary streamflow peaks (i.e., the parent process Z) between gamma and log-normal to be used to build MEV distributions FY for annual maxima (AM, i.e., the BM process Y). Here, we focus on the very primary logical contradiction (circular reasoning) of attempting to find a distribution FZ to build FY as a function of FZ to approximate the tail of FZ itself, which is already known exactly. In this respect, to keep the discussion as simple and focused as possible but without the loss of generality, we do not use compound models but assume that the parent process is independent and identically distributed, following a gamma distribution. Compound models and the issues related to some MEV technicalities (such as the declustering method used to obtain apparently independent ordinary events) will be discussed in the second example. Concerning the first example, we firstly discuss the above-mentioned contradiction (circular reasoning) from a conceptual perspective and then provide visual illustration by means of Monte Carlo simulations.
4.1.1 The logic behind the estimation of return levels and the role of FZ and FY
For the sake of illustration, let us suppose we have a hypothetical streamflow process sampled at a daily timescale, with us being interested in estimating a flow value that is exceeded, on average, every 𝒯 years, i.e., the so-called 𝒯-year return level corresponding to the 𝒯-year return period (see, e.g., Eichner et al., 2006; Serinaldi, 2015; Volpi et al., 2015, and references therein). Under the ideal situation where infinitely long records are available and, therefore, where FZ and FY are known exactly, one can use the distribution of the parent process FZ and determine the 𝒯-year return level as the quantile zp that is exceeded with probability , i.e., the value that is exceeded, on average, once in 𝒯 years = 365𝒯 d (leaving aside leap years). Since zp is a quantile of the distribution FZ, which describes the parent process at its finest available resolution (here, daily), it is unaffected by possible autocorrelation and clustering of 𝒯-year events (see Bunde et al., 2004, 2005; Serinaldi et al., 2020b, for an in-depth discussion). Note that this is the definition applied in the literature to compute the exact 𝒯-year return level used to assess the accuracy of 𝒩𝒜 models (see, e.g., Marani and Ignaccolo, 2015; Marra et al., 2018).
However, real-world records rarely span more than a few decades, and the data are not enough to obtain FZ (and FY) and determine directly the 𝒯-year return level for high values of 𝒯, such as 100 or 1000 years. Therefore, an alternative approach is based on the distribution of AM, i.e., BM within relatively short intervals (i.e., 365 d). Of course, a virtually infinite sequence of BM defines their exact distribution. Such a distribution allows an approximate estimation of the 𝒯-year return level as the quantile that is exceeded with probability because 1 year is the finest timescale of AM. In other words, FY cannot provide information about events occurring more often than once in m days (e.g., once per year for AM) as this is the finest sampling frequency of BM for blocks of size m. This estimation of 𝒯-year return levels based on BM involves the joint exceedance probability within each block described by the intra-block joint distribution Gl (see Sect. 2), and, therefore, it is affected by autocorrelation (see Eichner et al., 2006, for a detailed discussion).
Therefore, the distributions of AM commonly used in hydroclimatology are only approximations of the upper tail of FZ, and their estimation is justified if FZ is unknown. This can happen if (i) we have no regular records of the parent process to reliably estimate FZ or (ii) a faithful parameterization of FZ is not so easy to determine due to the difficulties in accounting for various characteristics of the underlying process, such as cyclo-stationarity, different physical generating mechanisms, and other possibly unknown factors. In these cases, EVT comes into play, stating, for instance, that, under certain assumptions, the distribution of BM within relatively short intervals (e.g., 365 d) converges to one of the three asymptotic extreme-value models summarized by the GEV distribution independently of the exact form of FZ. Of course, the approximate or partial fulfillment of EVT assumptions affects convergence. For example, autocorrelation and a lack of distributional identity slow convergence down (Koutsoyiannis, 2004; Eichner et al., 2006; Serinaldi et al., 2020b) and sometimes prevent it, resulting in degenerate models. These remarks explain why asymptotic models are such powerful tools that are widely applied in any discipline dealing with extreme values.
4.1.2 Visualizing the relationship between FZ and FY
A simple example with a graphical illustration can help better clarify the difference between FZ and FY (see Sect. 3.2 in Serinaldi et al., 2020b, for a formal discussion based on theoretical arguments). Let us assume that we have 365 × 105 observations of an independent process Z following a gamma distribution with shape and scale parameters , representing, for instance, 105 years of daily records of a hypothetical streamflow process (or a generic hydroclimatic process). These data allow one to build the empirical version of FZ and FY and the corresponding pdf's fZ and fY. In particular, Fig. 1 shows the empirical pdf's (Fig. 1a–c) and the return level plots (i.e., return level vs. return period; Fig. 1d–f) for two sub-samples of size 365 × 100 and 365 × 500 (i.e., 100 and 500 years, respectively) and for the whole data set (10 000 years). Figure 1 also displays the theoretical gamma pdf and return level curves, as well as empirical and theoretical 100-year quantiles (vertical lines). The 𝒯-year return levels are computed as the 100 % quantiles of the empirical cdf of AM and the 100 % quantiles of the theoretical and empirical cdf's of the process Z, where can be interpreted as the inter-arrival time (in years) between two records of Z.

Figure 1Probability density functions (a–c) and return level plots (return period vs. return level; d–f) of samples of varying size (365 × ) and corresponding BM (with block size m=365) drawn from a gamma distribution. The diagrams show the relationship between the parent distribution and the distribution of BM, along with the convergence of the upper tails of the empirical distribution toward the theoretical counterparts. The abscissa of dashed vertical lines indicates the value of the theoretical 100-year return level (gray lines) and its estimates from samples of the parent process Z (blue lines) and the corresponding BM process Y (red lines).
For 𝒯 greater than ≅ 20 years, the upper tail of the empirical FY (fY) matches that of the empirical FZ (fZ). This matching and convergence to the upper tail of the theoretical FZ (fZ) improve as the sample size increases. This behavior is further stressed focusing on the 100-year return levels (vertical lines in Fig. 1). It should be noted that the discrepancies between FY and FZ for 𝒯 < 20 years do not depend on the sample size. Instead, they are related to the different natures of the processes Y and Z, and their magnitude also depends on autocorrelation when data are correlated (see Sect 3.2 in Serinaldi et al., 2020b, for a theoretical discussion). Both distributions provide very close estimates of the 100-year return level for each sample size, and the accuracy obviously improves as the sample size increases. Moreover, Fig. 1 provides an intuitive (albeit very simplified) explanation of why EVT models of BM work when FZ is not available and when EVT assumptions are fulfilled.
From Fig. 1, it is evident that we do not need any model for Y if we already have a model for the parent Z. Since 𝒩𝒜 distributions require the preliminary definition or fit of a model for FZ, they have no practical usefulness as the preliminarily fitted FZ already provides all the information required to make statements on both ordinary and extreme events or quantiles. In this respect, defining 𝒩𝒜 distributions from FZ is only an unnecessary and redundant step, yielding only an approximation of the embedded FZ. These issues are further discussed in the next section in a review of a real-world data analysis previously reported in the literature.
4.2 Re-analysis of sea level data
In this section, we further illustrate the foregoing concepts by re-analyzing two sea level time series already studied by Caruso and Marani (2022). These data refer to hourly sea level records from the tide gauge of Hornbæk (Denmark) and Newlyn (United Kingdom), spanning 122 years (1891–2012) and 102 years (1915–2016), respectively. Data are freely available from the University of Hawaii Sea Level Center (UHSLC) repository (Caldwell et al., 2015, http://uhslc.soest.hawaii.edu/data/?rq#uh745a/, last access: 23 August 2022). For the sake of consistency with the original work, we removed years with less than 6 months of water level observations and days with less than 24 h of data (see Caruso and Marani, 2022). This resulted in 120 and 100 years of data for the Hornbæk and Newlyn gauges, respectively. Moreover, time series are pre-processed by filtering out the time-varying mean sea level () computed using the average of daily levels for each calendar year. Thus, the filtered time series retain the contributions from astronomical tides and storm surges.
Daily maxima are used as the basis for extreme-value analysis, which is performed by means of three different approaches: (i) GEV distribution of AM, (ii) GP distribution of POTs, and (iii) GP-based MEV of peaks over a moderate threshold (i.e., the so-called ordinary events). These extreme-value models assume that the underlying process is a collection of independent random variables. Since sea levels are a typical example of an autocorrelated process, data are preliminarily declustered by selecting peaks that are separated by at least 30 d to obtain (approximately) independent samples. In more detail, Caruso and Marani (2022) adopted “a threshold lag of 30 d, which yielded the minimum estimation error under the MEVD approach”. Therefore, declustered data are used to extract AM and POT samples over optimal statistical thresholds (Bernardara et al., 2014). Caruso and Marani (2022) selected the GP threshold for POTs by studying the stability of the GP shape parameter (Coles, 2001, p. 83), while they chose the moderate threshold of GP distributions entering MEV “by testing different threshold values and evaluating the goodness of fit of the distribution using diagnostic graphical plots”.
Before presenting results of extreme-value analysis, it is worth noting the following:
-
The extraction of independent data from correlated samples is referred to as “physical declustering” (Bernardara et al., 2014). Its algorithms rely on the physical properties of the process of interest (e.g., the lifetime of the weather systems generating a storm over an area) and/or the properties of the occurrence process (e.g., statistics of the (inter-)arrival times of rainfall storms). In this respect, a threshold selection based on “the minimum estimation error under the MEVD approach” not only requires iterative fitting of MEV components but also contrasts with the rationale of physical declustering, whose algorithms should be unrelated to the subsequent analysis and models involved. In other words, physical declustering should guarantee only the independence of the extracted sample and not the goodness of fit of a specific model (GP, MEV, or anything else).
-
Goodness of fit concerns statistical optimization, which aims to set a threshold that guarantees the convergence or fit of the POT sample to an extreme-value model. For the GP model, such a threshold should provide “the best compromise between the convergence of [the POT distribution toward] a GP distribution (bias minimization) and the necessity to keep enough data for the estimation of its parameters (variance minimization)” (Bernardara et al., 2014). In the present case, such a statistical threshold should not be required as the physical threshold was already selected to yield “the minimum estimation error under the MEVD approach” (Caruso and Marani, 2022). In fact, for Hornbæk and Newlyn data sets, the thresholds used by Caruso and Marani (2022) (i.e., 40 and 250 cm, respectively) lead us to discard approximately only 13 % of the complete declustered sample. Therefore, for the sake of comparison, we applied MEV to both the original declustered data and their over-threshold sub-samples.

Figure 2Detrended sea levels (gray line) for the gauging sites of Newlyn (UK) and Hornbæk (Denmark) and AM values for GEV analysis (a, d), POTs used for GP analysis (b, e), and over-threshold events used for fitting MEV and compound parent models (c, f). “Detrended” refers to sea level time series preliminarily filtered by removing the time-varying mean sea level.
For both data sets, Fig. 2 shows the time series of AM (Fig. 2a and d), POTs for GP (Fig. 2b and e), and POTs for MEV (Fig. 2c and f), along with the complete sample of daily maxima. Note that the sizes of POT samples are slightly different from those reported by Caruso and Marani (2022). This is likely due to the slightly different implementations of the declustering algorithm, which involves some technicalities such as the treatment of unavailable values.

Figure 3Return level diagrams (return period (in years) vs. return level) resulting from extreme-value analysis of Newlyn data (a–d) and Hornbæk data (e–h). All panels report empirical return level diagrams of AM as a common reference. Panels (a) and (e) report empirical return level diagrams of the parent sample of declustered data along with the theoretical return level diagram of the fitted GEV model. Panels (b) and (f) refer to POT sample and the corresponding GP model. Panels (c) and (g) and panels (d) and (h) show results for the MEV and compound parent distributions applied to the over-threshold data and complete declustered sample, respectively.
Figure 3 reports the results of extreme-value analysis in terms of return level plots. Figure 3a shows the empirical return level plot of the AM sample used to fit the GEV distribution and that of the corresponding declustered sample used to extract AM values. The values of the return period used to build these diagrams are estimated as , where Fn is the empirical cdf of AM or the declustered sample, and μ is the average inter-arrival time between two observations of a (discrete-time) process of interest, i.e., μ=1 for AM and for the complete declustered sample, where the random variable L denotes the varying number of events (or peaks) per year. Figure 3a and e are analogous to Fig. 1 and convey the same message but for real-world data; that is, the distribution of AM is only an approximation converging to the distribution of the parent sample for large quantiles (upper tail).
When using POT values over the threshold optimizing the GP fitting (Fig. 3b and f), we get a similar message: the distribution of AM is an approximation of the upper tail of the distribution of POTs, which plays a role similar to that of the parent sample in 𝒩𝒜 models of BM. In fact, the GP-based analysis of POTs does not require the subsequent derivation of the distribution of AM to make inferences regarding return levels as the return period (in years) of any quantile is computed as , where FGP is the GP cdf, and is the estimate of the average inter-arrival time between two POT observations. Even though this remark can seem trivial, it plays a key role in understanding the redundancy of 𝒩𝒜 distributions.
MEV models require us to preliminarily fit a model for values above a moderate threshold (or all available independent declustered data), which is our parent distribution FZ, and therefore derive the distribution of the annual maxima FY as a function of FZ. Figure 3c, d, g, and h show the empirical cdf's of both the AM and parent sample, as well as their theoretical counterpart, i.e., the GP-based MEV model and the compound GP parent. As for GEV, the MEV distribution is only an approximation of the upper tail of the fitted compound parent. However, in this case, we already have a model for the parent process, and, therefore, we do not need any distribution of AM as the fitted compound FZ already provides all the information required for inferential purposes. In other words, MEV cannot provide the correct probability of low or moderate quantiles (as with every extreme-value model of BM), and it cannot add any information compared to the corresponding fitted compound parent FZ. Once FZ is available, any other model of any sub-process (such as AM or POTs) is less informative or redundant, at most.
Figure 3c, d, g, and h also show that the claimed goodness of fit of MEV models is related to the fact that they are compound distributions rather than the fact that they are distributions of AM. In fact, MEV tails match those of the corresponding compound parent distributions. When we have a good compound model FZ integrating (i.e., averaging) seasonal fluctuations and other forcing factors (such as different generating mechanisms of rainfall, storms, flood, or storm surges), the corresponding 𝒩𝒜 model is no longer needed as it can, at most, be as accurate as the corresponding compound FZ.
The use of 𝒩𝒜 distributions is not even justified in making inferences regarding the return period and return levels. In fact, a compound FZ can be used to compute return levels in the same way as one uses GP distributions, calculating the return period as , where is the estimate of the average inter-arrival time between two observations in the sample of values above a moderate threshold (as for the case in Fig. 3c and g) or in the complete sample of independent declustered data (as for the case in Fig. 3d and h). In general, FZ does not require the derivation of the corresponding 𝒩𝒜 model for AM to make inferences with regard to the return period (expressed in years) in the same way as GP-based inference for POTs does not require the corresponding GEV model of AM.
The discussion in Sects. 3 and 4 was based on conceptual arguments, simplified numerical examples, and real-world data re-analysis with simple visual assessment. However, to be consistent with the scientific method, new models and methods should be validated or falsified against challenging and controlled conditions before being applied to real-world data coming from inherently unknown processes (Serinaldi et al., 2020a, 2022b). To this end, we set up three Monte Carlo experiments. The first experiment replicates and expands the numerical simulations reported by Marra et al. (2018) with the aim of providing independent validation and further evidence about the redundancy of 𝒩𝒜 models (here, MEV) when dealing with serially independent processes. The second experiment investigates the effect of autocorrelation on 𝒩𝒜-based analysis, evaluating the effectiveness of declustering algorithms based on threshold lags, as well as the use of βℬC models accounting for serial correlation without declustering. The third experiment replicates and expands some of the Monte Carlo simulations reported by Marani and Ignaccolo (2015) to support the introduction of MEV models. In this case, the aim is to explain the apparent discrepancies between the results in Marani and Ignaccolo (2015) and those in Marra et al. (2018).
5.1 Monte Carlo experiment 1: serially independent processes
The first experiment consists of simulating S=1000 time series of ordinary events mimicking 3, 5, 10, 20, and 50 years of records. Each year comprises l events drawn from a random variable L following a Gaussian distribution with a mean of and a standard deviation of σL = 0.3 μL. Marra et al. (2018) chose the range of μL and σL based on exploratory analysis of hourly rainfall data collected over the contiguous United States. Ordinary events are simulated from Weibull distributions with the shape parameter and the scale parameter λ=1. The κ values represent the typical range of variability of the observed rainfall data studied by Marra et al. (2018), while a constant λ is chosen for easier interpretation of the results. The simulated time series are used to estimate the 100-year return levels. The reference 100-year return level is empirically obtained from 105 years of simulated samples, and the performance of GEV, GP, and MEV is checked in terms of multiplicative bias:
where is the estimate of the target statistics (here, 100-year return level) for the kth Monte Carlo simulation (with ), and xref is the reference (true) value.
We note that the use of a Gaussian distribution with infinite support can generate a physically inconsistent negative number of events in some years. Moreover, simulating integer values from a continuous distribution requires rounding off. In these cases, more appropriate models for discrete random variables defined in [0,∞), such as binomial, beta-binomial, Poisson, or geometric, should be used. The reference 100-year return level can be computed as the 100 % quantile of the empirical cdf of AM or the 100 % quantile of the empirical cdf of the complete time series of ordinary events, where is the estimate of the average inter-arrival time (in years) between two ordinary events. For large samples, the former estimate converges to the latter for 𝒯 values greater than a few years (e.g., 3–5 years for independent data; see Fig. 1) or much more for serially dependent processes (see Sect 3.2 in Serinaldi et al., 2020b). In any case, the most accurate estimate of the 𝒯-year return level for every value of 𝒯 is given by the distribution FZ of the parent process, thus making the derivation of the distribution FY of AM redundant if the latter requires the preliminary definition of the former.
Results are reported as diagrams of the 5 %, 50 %, and 95 % quantiles of multiplicative bias versus the number of years. As expected, Fig. 4 is in perfect agreement with Fig. 7 in Marra et al. (2018) and leads to the same overall conclusions: MEV exhibits positive bias compared to GEV and GP but smaller variance. However, Fig. 4 provides an additional result concerning the performance of the compound parent distribution corresponding to MEV and shows that both models yield almost identical results apart from unavoidable sampling fluctuations in the estimation of the 5 % and 95 % quantiles based on 1000 simulated values of bias B. As discussed in Sects. 3 and 4, MEV distributions (or, more generally, 𝒩𝒜 distributions) do not add any information with respect to the parent distribution appearing in MEV formulas. Therefore, once a distribution is selected to describe the ordinary events (here, Weibull), its compound version is enough to make statements regarding any quantile, providing more information than the derived compound 𝒩𝒜 models, which approximate only the upper tail of the (embedded) parent distribution.

Figure 4Multiplicative bias for the 100-year return levels obtained from 1000 synthetic samples of varying record length (i.e., number of blocks or years) and with a varying number of ordinary events per block or year (10, 50, and 100) drawn from Weibull distribution with shape parameters κ=0.8 (a–c) and 1.25 (d–f). The reference 100-year return levels are empirically obtained from a 105-year record. Solid lines represent the median bias, while shaded areas (for GEV and MEV) and dashed lines (for GP and Compound parent) represent the 95 % Monte Carlo confidence intervals.
5.2 Monte Carlo experiment 2: serially dependent processes
This Monte Carlo experiment is designed to study the effect of autocorrelation on 𝒩𝒜-based inference. Time series of ordinary events mimicking 3, 5, 10, 20, and 50 years of daily records (i.e., 365 records per year) are simulated S = 1000 times to estimate 100-year return levels. The marginal distributions are the same as those used in the first experiment, i.e., Weibull with shape parameter and scale parameter λ=1. Autocorrelation is modeled by a first-order autoregressive (AR(1)) process with parameter , corresponding to weak, moderate, and relatively high autocorrelation. Weibull-AR(1) time series are generated by the CoSMoS framework, which enables the simulation of correlated processes with the desired marginal distribution and ACF (Papalexiou, 2018, 2022; Papalexiou and Serinaldi, 2020; Papalexiou et al., 2021, 2023).
Extreme-value analysis is performed by GEV for AM, GP for POTs of preliminarily declustered data, Weibull-based MEV for declustered data, and Weibull-based βℬC for the complete time series. Declustering is based on time lag, selecting the first lag τ0 such that the empirical ACF becomes smaller than twice the 99 % quantile of the sampling distribution of the ACF values under independence. Although this approach is slightly different from that used by Marra et al. (2018), the rationale is the same, and it yields τ0 values that guarantee sufficiently long inter-arrival times, as well as a suitable number of events per block for the considered AR(1) ACFs and sample sizes. Subsets of ordinary events used for MEV analysis are then defined as peaks separated by time intervals ≥τ0. POTs for GP analysis are extracted from these subsets, while AM for GEV analysis are selected from the original sample, assuming their inter-annual independence. Of course, βℬC analysis uses the complete data set and does not require any preliminary declustering procedure as it explicitly accounts for autocorrelation.

Figure 5Multiplicative bias for the 100-year return levels obtained from 1000 synthetic samples of varying record length (i.e., number of blocks or years) from the Weibull-AR(1) process, with Weibull shape parameters κ=0.8 (a–c) and 1.25 (d–f) and AR(1) parameter . The reference 100-year return levels are empirically obtained from a 105-year record. Solid lines represent the median bias, while shaded areas (for GEV and MEV) and dashed lines (for GP and compound parent) represent the 95 % Monte Carlo confidence intervals. MEV and compound parent distributions are fitted to preliminarily declustered data.
Figure 5 compares the results of the GEV, GP, and MEV analyses. For ρ1=0.3, values of bias B are similar to those obtained for the previous experiment in Sect. 5.1, with μL=100 (Fig. 4). This is expected as low values of ρ1 correspond to rapidly decreasing ACF and, therefore, τ0≅ 2–3 time steps, corresponding to sample sizes of ordinary events between about 120 and 180. For ρ1=0.6 and 0.9, τ0 increases to 4–6 and 15–30 time steps, respectively, corresponding to sample sizes of 60–90 and 12–24 ordinary events. The progressively reduced sample size increases MEV uncertainty, which becomes similar to that of GEV and GP models. More importantly, MEV bias dramatically increases with ρ1 and the number of years (blocks). The effect of ρ1 is easy to interpret in terms of reduced sample size resulting from declustering with larger τ0. On the other hand, the effect of the number of years could appear to be counterintuitive as one would expect more accuracy when a larger number of years is available.
Marra et al. (2018) ascribe this behavior “to uncertain estimation of the weight of the tail of the ordinary events distribution when few data points are used for the fit”. However, this would not be sufficient to explain why the smallest bias corresponds to small numbers of available years and, thus, to overall smaller samples. The actual issue is the combination of the (average) number of intra-block peaks (or intra-block sample sizes – here, l or μL), the number of blocks (here, the number of years nY), and the compounding procedure characterizing MEV.
For fixed nY, a small intra-block sample size l results in great variability in the Weibull parameters estimated in each block, which, in turn, results in heavier tails of compound distributions. As l increases, the inter-block variability of Weibull parameters decreases, and the compound distribution resulting from averaging a set of similar Weibull distributions becomes closer and closer to the theoretical Weibull used to simulate. In other words, the compounding mechanism works better in those cases in which it is less required, i.e., when the inter-block variability is small and when model averaging (of very similar models fitted on each block) is less justified and useful. On the other hand, when model averaging could be more justified, i.e., when there is substantial uncertainty in the sampling parameters, the dispersion of the sampling distribution of parameters is greater, and the tail of the resulting compound distribution is heavier, with a shape departing from that of the (true) theoretical distribution.
For given μL, when the number of years nY is small, compound 𝒩𝒜 models average a small number of components , with (e.g., we have three components for nY = 3 years). In a Monte Carlo experiment, averaging a few heterogeneous components results in a set of heterogeneous compound distributions whose differences tend to compensate for each other, on average. Therefore, the Monte Carlo ensembles of compound distributions exhibit high variability and small bias. As nY increases, the number of averaging components increases, providing a more accurate picture of the inter-block variability that is incorporated into the compound distributions. This results in Monte Carlo ensembles of compound distributions with more homogeneous and systematically heavier tails than those of the compound models resulting from small nY. Therefore, the Monte Carlo ensemble exhibits lower variance and higher bias as nY increases for a given μL.
As for Fig. 4, Fig. 5 also reports results for the compound distribution of ordinary events, which are almost indistinguishable from those of MEV analysis. Overall, Fig. 5 further confirms the redundancy of MEV models (and, more generally, 𝒩𝒜 models) once we have a compound parent distribution, which has to be estimated in any case to derive 𝒩𝒜 distributions. Moreover, uncorrelated ordinary events resulting from declustering procedures do not guarantee convergence of compound distributions (MEV or parent) to the true distribution. In fact, the bias is generally much larger than that of GEV and GP estimates, although the intra-block sample size is generally much larger than that of AM and POTs, and the compound distributions have a much larger number of parameters (from 6 to 100, resulting from the two-parameter Weibull fitted to 1-year blocks over 3 to 50 years).

Figure 6Multiplicative bias for the 100-year return levels obtained from 1000 synthetic samples of varying record length (i.e., number of blocks or years) from the Weibull-AR(1) process, with Weibull shape parameters κ=0.8 (a–c) and 1.25 (d–f) and AR(1) parameter . The reference 100-year return levels are empirically obtained from a 105-year record. Solid lines represent the median bias, while shaded areas (for GEV and MEV) and dashed lines (for GP and compound parent) represent the 95 % Monte Carlo confidence intervals. βℬC and compound parent distributions are fitted to complete autocorrelated time series.
Figure 6 compares the results of the GEV and GP analyses with those of βℬC and compound parent models. Since βℬC models (and the corresponding compound parent) use the complete time series instead of declustered data, uncertainty and bias are smaller than those of MEV models (and the corresponding compound parent). Therefore, while time lag declustering seems to yield apparently independent events, the resulting data sets do not provide a faithful description of the upper tail of the true generating process; that is, MEV models do no make suitable use of these declustered samples. Declustering has negative effects independently of the intensity of autocorrelation. Of course, larger bias and uncertainty correspond to higher ρ1 values in both MEV and βℬC analyses. In fact, MEV is affected by significant decreases in sample size due to declustering, while βℬC suffers from underestimation of ACF, which requires large sample sizes to be reliably estimated (see, e.g., Koutsoyiannis and Montanari, 2007; Serinaldi and Kilsby, 2016a). It is worth noting that the GEV and GP results are rather insensitive to autocorrelation. This is expected as the underlying joint dependence structure of AR(1) processes is a Gaussian copula, which is characterized by asymptotic tail independence and is therefore complaint with EVT assumptions. Similarly to Fig. 4 and 5, Fig. 6 shows that the βℬC model and the corresponding compound parent match (apart from discrepancies due to the issues mentioned above), confirming the redundancy of 𝒩𝒜 models.
5.3 Monte Carlo experiment 3: reviewing simulations of Marani and Ignaccolo (2015)
Figure 4 shows that MEV and its compound parent distribution yield a median multiplicative bias BM≅1.25 for 100-year return levels estimated from nY = 50 years (blocks) of data drawn from Weibull distributions with the shape parameter κ=0.8 and an average number of events per block of . On the other hand, BM≅1.0 for GEV and GP distributions. For a similar setup (i.e., nY=50, κ=0.82, and ), Marani and Ignaccolo (2015) reported probability plots (probability vs. quantiles) and relative error:
where is the estimate of the target statistics for the kth Monte Carlo simulation (with ), and xref is the reference (true) value. They found that MEV is almost unbiased, with an average relative error of , while GEV exhibits bias, with 5 % and ≅ 30 % for the 100- and 1000-year return levels, respectively. On the other hand, for the 100-year return level, simulations in Sect. 5.1 (reproducing those of Marra et al., 2018) yield 25 % for MEV and for GEV. Therefore, we re-run Monte Carlo simulations described by Marani and Ignaccolo (2015) to understand the reason for such a disagreement. We anticipate that the foregoing discrepancies depend on the misuse of methods used to summarize multi-model ensembles. Thus, before describing Monte Carlo experiments and their outcome, we need to recall some theoretical concepts that are required to correctly interpret numerical results.
5.3.1 Summarizing multi-model ensembles: some overlooked concepts
Monte Carlo simulations are usually used to study the uncertainty affecting estimates based on finite-sized samples (that provide incomplete information about the underlying process) or to approximate population distributions (or statistics) when mathematical closed-form expressions are not available. Examples of these applications are the experiments reported in Sect. 5.1 and 5.2.
In all cases, the primary output of Monte Carlo simulations is a set of parameters identifying a set of models (multi-model ensemble) that is then used to estimate the target statistics of interest. For example, simulations of S finite-sized samples in Sect. 5.1 and 5.2 are used to fit a set of S GEV distributions. These are then used to calculate a set of S 100-year return levels, which are eventually used to build confidence intervals summarizing sampling uncertainty.
However, a multi-model ensemble can be summarized in many different fashions to obtain a representative point estimate of a statistic of interest (e.g., Renard et al., 2013; Fawcett and Walshaw, 2016; Fawcett and Green, 2018). Let S be the number of Monte Carlo replications; let F(z|θk) (with ) be the kth member of the Monte Carlo multi-model ensemble (e.g., the kth Weibull distribution fitted to the kth simulated sample); let zp be a target quantile with a non-exceedance probability p; and let us define the quantile function as the inverse of the cdf, . A representative point estimate of zp can be, for instance, the mode of the S quantiles .
More popular point estimates of zp (or whatever statistics) rely on the definition of so-called predictive distributions and predictive quantile functions. The sampling predictive cdf reads as
and the corresponding quantile with a specified non-exceedance probability p is given by
The sampling predictive quantile function reads as
resulting in predictive quantile estimates of
Let us denote the empirical cdf and quantile function of the S sampled quantiles zp,k as FS and QS, respectively. Recalling that the distribution of zp can be approximated by the distribution of order statistics and that the latter is described by a generalized beta distribution (see Eq. 1, as well as Eugene et al., 2002; Tahir and Cordeiro, 2016), we can write , where . Therefore, the foregoing zp estimators can be complemented by the median estimator, defined as
Similarly, we can also define the median probability of a fixed quantile zp from an ensemble of cdf's as follows:
The foregoing formulas indicate that the three zp estimators obviously represent different quantities. Focusing on and and comparing Eqs. (11) and (13), we have it that
Equation (16) is the sampling counterpart of , which, in turn, follows from the well-known general inequality
stating that the distribution of the expected value of Z is different from the expected distribution of Z. In fact, since F is commonly a nonlinear transformation of Z (as well as of the parameters θ), it hinders the interchangeability of the (linear) expectation operator 𝔼. In passing, such an inequality also partly caused a long “querelle” (controversy) with regard to plotting position formulas (see, e.g., Makkonen, 2008; Cook, 2012; Makkonen et al., 2013).
On the other hand, zp,M is the only estimator that guarantees the identity between the zp estimates obtained from ensembles of Q or F functions. This property depends on the fact that the median (as well as every quantile) is a rank-based (central-tendency) index, and ranking is a transformation that does not depend on absolute values and therefore passes unaffected trough nonlinear monotonic functions such as Q and F. This means that the median parameters θM correspond to zp,M and pM. This property does not hold for the expectation operator 𝔼. In fact, generally, .
The foregoing concepts and properties play a key role in the correct interpretation of the results reported in the next section.
5.3.2 Numerical simulations: the consequences of overlooking theory
Marani and Ignaccolo (2015) supported the introduction of MEV by means of five Monte Carlo experiments (referred to as cases “A”, “B”, “C”, “A2”, and “B2”), comparing the accuracy of MEV to that of standard asymptotic models of BM (i.e., Gumbel and GEV distributions). For the cases B and B2, Marani and Ignaccolo (2015) did not provide enough information to enable their replication. Therefore, we focused on cases A, C, and A2, which are sufficient to support our discussion. Case A consists of simulating S=1000 samples from a Weibull distribution with a scale parameter equal to 7.3, a shape parameter of κ=0.82, a number of blocks (years) of nY=50, and a number of events per block (here, wet days per year) of l=100. Case C is similar to A, with the only difference being that the number of events per block is drawn from a uniform distribution 𝒰(21,50). The setup of case A2 is similar to that of A; however, it explores the effect of varying l from 10 to 200 by steps of 10 events per block. Therefore, Gumbel, GEV, and MEV distributions of BM are fitted to each of the S samples. For the cases A and C, the accuracy of the three models is assessed by comparing “the ensemble average distributions, ζMEV(y), ζGEV(y), ζGUM(y) as the means of the distributions of Y computed over the 1000 synthetic time series” (Marani and Ignaccolo, 2015). For the case A2, the three models are evaluated in terms of the average relative error of the estimates of the 100- and 1000-year return levels. The reference (true) return levels are empirically obtained from 106 years of simulated samples.

Figure 7Probability plots (probability vs. quantile) showing different models for AM Y resulting from the Monte Carlo experiments denoted as cases A (a, b, d, e) and C (c, f) (see main text for details about the simulation setup). Panels (a–c) reproduce results reported in Marani and Ignaccolo (2015, Fig. 3a and c), while panels (d–f) show the revised version with corrections accounting for inconsistencies in the calculation of compound quantiles and the misuse of multi-model ensemble averaging.
Figure 7a–c reproduce Fig. 3a and c in Marani and Ignaccolo (2015). For the case A, we used both κ=0.82 and 0.73 as the original parameterization cannot reproduce the results of Fig. 3a in Marani and Ignaccolo (2015). In fact, analyzing the original Fig. 3a, the reference 100- and 1000-year return levels should be close to 150 and 210, respectively, while κ=0.82 yields values close to 109 and 143, which, in turn, are consistent with case C. Therefore, we used κ=0.73 to obtain a figure that is as close as possible to the original one. Nonetheless, the exact value of κ is inconsequential in the following discussion, and we use both κ=0.82 and 0.73 for completeness.
The key aspects in Fig. 7a–c are (i) the perfect match of MEV and its compound parent, confirming the redundancy of 𝒩𝒜 models when their parents are already known, and (ii) the accuracy of MEV and its compound parent against the prominent bias of GEV, which is in contrast to results reported by Marra et al. (2018) and in the previous sections. The reason for such a discrepancy is that Fig. 7a–c (and Fig. 3a in Marani and Ignaccolo, 2015) do not show what they are supposed to do, thus making the comparison unfair and misleading. In fact, contrarily to the description in Marani and Ignaccolo (2015), the MEV curves in Fig. 7a–c do not refer to the predictive MEV obtained by averaging S MEV distributions according to Eqs. (10) and (11). Instead, recalling that MEV is itself a predictive distribution (i.e., the average of multiple components , with ; see Sect. 2), MEV curves in Fig. 7a–c refer to the predictive version (averaged over S samples) of MEV quantile functions, which are predictive quantile functions themselves resulting from averaging over nY samples.
In other words, Fig. 7a–c report the pairs instead of the claimed , and these pairs differ from each other (see Sect. 5.3.1). In more detail, , while the figure should show obtained by inverting .
On the other hand, Fig. 7a–c (and Fig. 3a and c in Marani and Ignaccolo, 2015) correctly show the predictive distributions of Gumbel and GEV. However, this hinders a fair comparison. In fact, EVT states that the asymptotic model of BM is a GEV distribution (under suitable conditions) and not the compound version of GEV resulting from averaging S GEV models. Such a compound GEV distribution always has a larger variance and heavier tails than its classical GEV counterpart (see Sect. 6). Therefore, to be consistent with EVT, the ensemble of GEV and Gumbel distributions should be summarized using a transformation, such as the median, that retains the expected GEV or Gumbel shape. Figure 7d–f show the median GEV and Gumbel distributions (resulting from Eq. 15), along with the actual predictive MEV (as it should be). Results in Fig. 7d–f are fully consistent with those reported by Marra et al. (2018) and in Sect. 5.1 and 5.2, confirming the low bias of asymptotic models and the natural tendency of compound distributions to exhibit heavier tails than their components and their generating processes. Moreover, the perfect agreement of the upper tail of MEV and that of the compound parent distributions in Fig. 7d–f further confirms (if still needed after many examples) the redundancy of 𝒩𝒜 models once their parent distributions are defined.

Figure 8Relative errors for 100- and 1000-year return levels resulting from the Monte Carlo experiment denoted as case A2 (see main text for details about the simulation setup). Panels (a) and (b) reproduce results reported in Marani and Ignaccolo (2015, Fig. 4a) for , while panels (c) and (d) show the revised version with corrections accounting for inconsistencies in the calculation of compound quantiles and the misuse of multi-model ensemble averaging.
Similar remarks hold for the case A2. Results in Fig. 8a and b are close to those reported by Marani and Ignaccolo (2015) in their Fig. 4a, with MEV showing for both the 100- and 1000-year quantiles and with GEV showing for the 100-year return level and 5 % for the 1000-year return level. The Gumbel distribution yields slightly negative for both return levels, with smaller values for higher κ, which corresponds to a generating Weibull distribution closer to exponential, thus allowing faster convergence to the first asymptotic distribution of EVT. As for the cases A and C, these results are affected by mixing predictive distributions and predictive quantile functions, as well as the improper use of the former to summarize the ensemble of GEV and Gumbel models. Figure 8c and d show the median relative errors corresponding to GEV, Gumbel, and true predictive MEV distributions. As mentioned above, the median of S the GEV and/or Gumbel distributions is still a GEV and/or Gumbel distribution. As the S relative errors of GEV return levels (Eq. 9) are just the rescaled value of GEV return levels, if the median GEV correctly describes the reference (true) distribution of AM, the median relative error (over S=1000 samples) is expected to be equal to zero. On the other hand, the mean relative error of GEV return levels is expected to be different from zero as it would correspond to the difference between a compound GEV (resulting from averaging over S samples) and the reference distribution of AM.
As expected, the GEV model correctly describes BM, while the compound structure of MEV yields heavier tails. Once again, results from MEV and compound parent are almost indistinguishable due to the redundancy of MEV (and any 𝒩𝒜 model in general).
The work by Marani and Ignaccolo (2015) also suffers from several mismatches between the text and figures. For example, concerning the case A2 and the corresponding Fig. 4a, they state that the “GEV approach systematically overestimates the 100-year extreme rainfall intensity by 5 % even for large numbers of wet days. The Gumbel approach systematically underestimates the 100-year extreme rainfall intensity by about 5 %. For the 1000-year return period intensities, the GEV approach severely overestimates actual extreme events (minimum relative error is 30 % for n = 200 events yr−1), whereas the Gumbel approach yields underestimation errors of about 10 %”. However, in contrast with the text, their Fig. 4a shows that GEV has for the 100-year return level and 10 % for the 1000-year return level, while Gumbel distributions have −15 % and ≅ −30 % for the 100- and 1000-year return levels, respectively. Concerning the case B2 and the corresponding Fig. 4b, any interpretation is impossible as Fig. 4b in Marani and Ignaccolo (2015) reports the “root mean square % error”, whereas the text refers to , and it is not even clear if Fig. 4b actually refers to the case B2.
The proposal of 𝒩𝒜 models as an alternative to classic EVT models suffers from some problems that seem to be quite widespread in the hydrological literature dealing with statistical methods (see, e.g., discussions in Serinaldi and Kilsby, 2015; Serinaldi et al., 2018, 2020a, 2022b).
-
Data analysis should be supported by preliminary scrutiny of its rationale, allowing, for instance, the recognition of the “circular reasoning” affecting the practical use of 𝒩𝒜 models of BM. Extreme-value models are powerful tools if applied in the right context according to their motivation and assumptions. Their usefulness relies on the fact that they provide an approximate description of the upper (or lower) tails of the distribution of parent processes when the latter is unknown and when there are no data (or not enough data) to reliably estimate it. 𝒩𝒜 models of BM contradict this principle. In fact, 𝒩𝒜 models require the preliminary estimation of a parent distribution FZ to build a surrogate distribution FY that approximates a tail of FZ, neglecting the fact that FZ is already known or fitted.
For example, Marra et al. (2023) studied the distribution of worldwide daily rainfall data over low and moderate thresholds showing that a Weibull model provides a good fit and reproduces L-moments of AM even when AM are excluded from calibration. Conversely, using GP tails provides the same results only over the 95 % threshold and overestimates the heaviness of the upper tail when the GP model is assumed for low or moderate thresholds (in agreement with results reported by Serinaldi and Kilsby (2014b) about the multiple-threshold method (Deidda, 2010)). The natural interpretation of these results would be that the Weibull distribution is a good model FZ for the parent process Z (positive rainfall or rainfall over low or moderate thresholds), confirming previous results reported in the literature, while the GP model works well for exceedances over high thresholds (as postulated by EVT) and does not work well (as expected) for low or moderate thresholds, that is, outside its range of validity. Recalling the theoretical link between GP and GEV, this also means that the latter is a good model for rainfall BM.
For practical applications, this should translate into the following recommendations: (i) use GEV if only BM are available (e.g., AM from hydrological yearbooks), and (ii) use FZ (e.g., (compound) Weibull) if you have information on Z, which can be either the process of all positive rainfall or the rainfall over arbitrary low or moderate thresholds if the latter is deemed easier to fit. In the latter case, calculate the 𝒯-year return levels as the 100 % quantiles of FZ, where μ is the (mean) inter-arrival time (in years) between two observations of Z (e.g., Serinaldi, 2015; Volpi et al., 2019).
Such a plain reasoning highlights that there is no need to build an additional distribution of BM (i.e., (compound) ). This is similar to the case of asymptotic models, whereby we do not need to define the GEV distribution of AM once we have already inferred a GP model of POTs. Nonetheless, Marra et al. (2023) interpreted their results as evidence in support of 𝒩𝒜 models of BM, missing the fact that the fitted Weibull distributions over zero, low, or moderate thresholds are conceptually similar to each other and can be used directly to make inferences about any desired quantile without deriving redundant models of BM (here, exponentiated Weibull).
-
New methods need to be suitably validated before being applied. Actually, applications to real-world data are often improperly used as validation. Proper validation or falsification requires the use of processes with known properties that match or contrast the model assumptions. For example, 𝒩𝒜 models, such as (S)MEV, have only been assessed for parent processes with known marginal distributions under independence (e.g., Marra et al., 2018), while the effect of dependence and the effectiveness of declustering were not checked. We encourage modelers to perform proper Monte Carlo simulations as suitable methods are readily available for such an analysis (e.g., Serinaldi and Lombardo, 2017a, b; Papalexiou, 2018; Serinaldi and Kilsby, 2018; Koutsoyiannis, 2020; Papalexiou and Serinaldi, 2020; Papalexiou et al., 2021; Papalexiou, 2022, among others). Of course, numerical experiments should be supported by the theoretical knowledge required to allow for correct implementation and interpretation and to prevent inconsistencies such as those discussed in Sect. 5.3.
On the other hand, proper validation was replaced by quite an extensive use of cross-validation exercises based on observed data (e.g., Miniussi and Marani, 2020; Mushtaq et al., 2022), which might, however, be misleading for the following reasons:
- a.
Hydroclimatic records come from processes with inherently unknown properties as only estimates of the variables of interest are available.
- b
Cross-validation is usually performed on short time series (commonly, a few years of data), and model estimates (from shorter calibration subsets) are compared with sample estimates (from shorter verification subsets), which might not be representative of the true value of the target statistics. Cross-validation relies on the assumptions that the calibration subsets are representative of the population, and out-of-sample subsets come from the same population. However, for autocorrelated processes, very long time series might be required to explore the state space of the studied process (Koutsoyiannis and Montanari, 2007; Dimitriadis and Koutsoyiannis, 2015), thus meaning that the observed series might be not representative, especially when focusing on extreme values. In hydroclimatic processes, this issue is exacerbated by the effect of long-term fluctuations characterizing the climate system at local and global spatial scales.
- c.
Standard bootstrap resampling used in cross-validation might also be misleading. In fact, it provides correct results under the assumption that the state space is explored under independence, and, therefore, relatively short samples are enough to give a reliable picture of the range of possible outcomes. If the hypothesis of independence is not valid, the observed values might cover a subset of the state space, and the standard bootstrap commonly applied in MEV literature just conceals this fact.
- a.
-
Often, inappropriate validation and iterated application to real-world data generate quite an extensive literature in which numerical artifacts are confused with physical properties (see, e.g., Serinaldi and Kilsby, 2016a; Serinaldi et al., 2020a, 2022b, for paradigmatic examples). Such a literature is often improperly used to support a given method through arguments like “there is such a strong scientific body of literature demonstrating the technical advantages of these approaches”. However, consensus is not a scientific argument. Historically, the main scientific progresses occurred when someone called into question widely accepted mainstream theories using arguments more solid than those of the superseded theories. Consensus is even more questionable when a method is iteratively applied without a necessary neutral or independent validation. The literature on 𝒩𝒜 models tends to suffer from these problems, and our discussion in Sect. 5.3 illustrates how these models have been iteratively applied without the above-mentioned independent analysis.
It is quite common to read sentences such as “these new approaches have been shown to be practically useful under real conditions, showing their practical advantage over traditional methods”. Such kinds of statements do not provide any technical information about either the relationship between the distribution of BM and POTs and their corresponding parent or the rationale and effects of compounding multiple models or the difference between the parameterizations of GEV and 𝒩𝒜 models, for instance. Moreover, if a method is biased, as shown in the previous sections, multiple applications to real-world data do not make it unbiased.
-
Often, (seemingly) new methods are not put into their broader context and are denoted by uninformative names, thus concealing their nature and hindering correct interpretation. In particular, 𝒩𝒜 distributions are only special versions of the class of compound distributions (e.g., Dubey, 1970; van Montfort and van Putten, 2002):
where is the marginal pdf of a generic variable X, f(θ) is the pdf of the parameter vector θ of the distribution f(x|θ), and Ωθ is the state space of θ when it is treated as a random variable Θ. The variance 𝕍[X] of is always greater than that of its components f(x|θ) (e.g., Karlis and Xekalaki, 2005).
Compound distributions have been presented in the literature under various names and contexts, such as “superstatistics” in physics and hydrology (Beck, 2001; Porporato et al., 2006; De Michele and Avanzi, 2018) and “predictive distributions” in theoretical and applied statistics (Benjamin and Cornell, 1970; Wood and Rodríguez-Iturbe, 1975; Stedinger, 1983; Bernardo and Smith, 1994; Kuczera, 1999; Coles, 2001; Cox et al., 2002; Gelman et al., 2004; Renard et al., 2013; Fawcett and Walshaw, 2016; Fawcett and Green, 2018), or without introducing any specific name (Koutsoyiannis, 2004; Allamano et al., 2011; Botto et al., 2014; Yadav et al., 2021). In more detail, Eq. (18) “might be referred to as the prior (Bayesian) distribution or the posterior (Bayesian) distribution on X, depending on whether a prior or posterior distribution of θ is used to determine ” (Benjamin and Cornell, 1970, pp. 632–633). f(θ) can be analytical (e.g., Skellam, 1948; Moran, 1968; Dubey, 1970; Hisakado et al., 2006) or empirical, resulting from Monte Carlo simulations, bootstrap resampling, or estimation from multiple sub-samples, such as in the case of βℬC or MEV inference.
However, using our notation, “can be interpreted as a weighted average of all possible distributions f(x|θ) which are associated with different values of θ. In this sense [Eq. (18)] can be interpreted as an application of the total probability theorem… In any event we note that the unknown parameter will not appear in as it has been “integrated out” of the equation. We also note that, as more and more data become available, the distribution of θ will be becoming more and more concentrated about the true value of the parameter. We should generally expect the distribution to be wider, e.g., to have a larger variance, than the true f(x), since the former incorporates both inherent and statistical uncertainty” (Benjamin and Cornell, 1970, pp. 632–633).
In other words, 𝒩𝒜 models, such as βℬC and MEV, are just the output of what is often referred to as multi-model ensemble averaging (e.g., Burnham and Anderson, 2002; Giorgi and Mearns, 2002, and references therein). The inherent nature of compounding or averaging procedures explains the tendency of 𝒩𝒜 models to yield with tails that are heavier than those of the true underlying distribution f(x), as well as progressive convergence of to f(x) as the (block) sample size increases and as f(θ) becomes more and more concentrated around the true value of the parameter(s). It also clarifies that the properties of 𝒩𝒜 models of BM depend on being compound models rather than extreme-value models. In fact, the same results can be obtained by directly compounding the distributions of the parent process without any additional derivation of the corresponding distributions of BM. Furthermore, recognizing the rationale of compound models allows us to understand that the BM process is different from the parent one, and the distribution of the former is useful only if the latter is not available. Finally, as shown in Sect. 5.3, understanding the nature of compounding procedures is fundamental to correctly summarize and interpret multi-model outputs.
This study presented an inquiry into non-asymptotic (𝒩𝒜) distributions FY of block maxima (BM) Y, which was motivated by their increasing use in data analysis without a necessary preliminary validation or falsification under controlled conditions, along with a deep discussion of their rationale and their relationship with the distribution FZ of the generating process Z. We discussed their redundancy in real-world analysis. This apparently bold statement relies on very basic facts: (i) the distribution FZ of a process Z provides all the information about any quantile or summary statistics (extreme or not); (ii) extreme-value distributions FY of BM corresponding to the parent process Z are only approximations of the tails of the distribution FZ, and they have a role only if FZ is unknown; and (iii) 𝒩𝒜 distributions imply the preliminary knowledge or estimation of FZ, but once FZ is known or fitted to data, 𝒩𝒜 distributions of BM are no longer needed, and their derivation is redundant as FZ already provides all the information. In this context, the use of asymptotic extreme-value models is justified by the fact that they do not require the preliminary knowledge or estimation of FZ (under suitable conditions).
While the foregoing logical arguments should be sufficient to call into question the practical use and usefulness of 𝒩𝒜 models, we further demonstrated these issues by means of simplified examples, re-analysis of real-world data, and suitable Monte Carlo simulations. The aim was to support conceptual statements with numerical experiments that are easy to reproduce and can independently be checked. In this way, debate can be based on technical counter-arguments and proper analysis of data drawn from processes with known properties, avoiding the “consensus” argument and resetting the discussion about 𝒩𝒜 models within the boundaries of the scientific method.
Of course, the questionable usefulness of 𝒩𝒜 models in practical applications does not mean that they are not useful at all. As shown in this study and by Serinaldi et al. (2020b), 𝒩𝒜 formulation clarifies the inherent relationship between the distribution of BM (FY) and that of their generating process (FZ), thus shedding light on some inferential aspects from a theoretical point of view. For example, 𝒩𝒜 formulation highlights that the difference between return periods or levels estimated from FY and FZ does not depend on sample size (as incorrectly stated in the literature) but rather on the theoretical difference in the processes Y and Z and cannot be reduced. 𝒩𝒜 expressions also allow a better understanding of the mechanism of compounding distributions to account for multiple generating processes, showing the dualism of additive and multiplicative mixing in the derivation of FY from FZ (Serinaldi et al., 2020b). In principle, 𝒩𝒜 models incorporating dependence are also the basis for the theoretical study of the corresponding asymptotic models free from the preliminary definition of FZ.
To conclude, models and methods should be thought of and used in the right context and for suitable purposes. The reliability of models must rely on a careful preliminary analysis of their consistency with the logic, theory, data, processes analyzed, and problems at hand. A cautious approach should start from the assumption that a new model is likely to be questionable in terms of novelty (it can already exist, perhaps under a different name in different disciplines), theoretical correctness, and practical usefulness. Therefore, model developers should perform a deep literature review (possibly extended to other disciplines); clearly understand the rationale, assumptions, and purpose of the model; and attempt model falsification rather than validation. New models should be tested under controlled challenging conditions. We believe that these recommendations are the cornerstones of a rigorous scientific inquiry and are too often neglected. Calling into question the practical usefulness of 𝒩𝒜 models of BM is precisely an application of that investigation method.
Data are freely available from University of Hawaii Sea Level Center (UHSLC) repository (Caldwell et al., 2015, https://doi.org/10.7289/V5V40S7W).
FS: conceptualization, methodology, software, formal analysis, writing (original draft), writing (review and editing), visualization. FL: conceptualization, methodology, writing (review and editing). CGK: conceptualization, writing (review and editing).
The contact author has declared that none of the authors has any competing interests.
Publisher's note: Copernicus Publications remains neutral with regard to jurisdictional claims made in the text, published maps, institutional affiliations, or any other geographical representation in this paper. While Copernicus Publications makes every effort to include appropriate place names, the final responsibility lies with the authors.
Federico Lombardo is grateful to Silvia Ghinaglia for her scientific advice and continuous support. The authors thank Francesco Marra (Università degli Studi di Padova), an anonymous reviewer, and the discussant Sarah Han for their critical remarks. The authors also thank Theano Iliopoulou (National Technical University of Athens), Marco Marani (Università degli Studi di Padova), and Giuseppe Mascaro (Arizona State University) for their critical comments on a previous version of the paper. The analyses were performed in R (R Development Core Team, 2023).
This paper was edited by Lelys Bravo de Guenni and reviewed by Francesco Marra and one anonymous referee.
Allamano, P., Laio, F., and Claps, P.: Effects of disregarding seasonality on the distribution of hydrological extremes, Hydrol. Earth Syst. Sci., 15, 3207–3215, https://doi.org/10.5194/hess-15-3207-2011, 2011. a
Arnold, B. C., Balakrishnan, N., and Nagaraja, H. N.: A First Course in Order Statistics, Classics in Applied Mathematics, Society for Industrial and Applied Mathematics (SIAM), Philadelphia, PA, ISBN 978-0-89871-648-1, 1992. a
Balkema, A. A. and de Haan, L.: Residual Life Time at Great Age, Ann. Probab., 2, 792–804, 1974. a
Beck, C.: Dynamical Foundations of Nonextensive Statistical Mechanics, Phys. Rev. Lett., 87, 180601, https://doi.org/10.1103/PhysRevLett.87.180601, 2001. a
Beirlant, J., Goegebeur, Y., Segers, J., Teugels, J., De Waal, D., and Ferro, C.: Statistics of Extremes: Theory and Applications, Wiley Series in Probability and Statistics, Wiley, Chichester, England, ISBN 0-471-97647-4, 2004. a
Benjamin, J. R. and Cornell, C. A.: Probability, Statistics, and Decision for Civil Engineers, McGraw-Hill, New York, ISBN 07-004549-6, 1970. a, b, c
Bernardara, P., Mazas, F., Kergadallan, X., and Hamm, L.: A two-step framework for over-threshold modelling of environmental extremes, Nat. Hazards Earth Syst. Sci., 14, 635–647, https://doi.org/10.5194/nhess-14-635-2014, 2014. a, b, c
Bernardo, J. M. and Smith, A. F. M.: Bayesian Theory, John Wiley & Sons, New York, ISBN 0-471-92416-4, 1994. a
Botto, A., Ganora, D., Laio, F., and Claps, P.: Uncertainty compliant design flood estimation, Water Resour. Res., 50, 4242–4253, 2014. a
Boulesteix, A., Binder, H., Abrahamowicz, M., Sauerbrei, W., and for the Simulation Panel of the STRATOS Initiative: On the necessity and design of studies comparing statistical methods, Biometrical J., 60, 216–218, 2018. a
Bunde, A., Eichner, J. F., Havlin, S., and Kantelhardt, J. W.: Return intervals of rare events in records with long-term persistence, Physica A, 342, 308–314, 2004. a
Bunde, A., Eichner, J. F., Kantelhardt, J. W., and Havlin, S.: Long–term memory: A natural mechanism for the clustering of extreme events and anomalous residual times in climate records, Phys. Rev. Lett., 94, 048701, https://doi.org/10.1103/PhysRevLett.94.048701, 2005. a
Burnham, K. P. and Anderson, D. R.: Formal inference from more than one model: Multimodel Inference (MMI), in: Model Selection and Multimodel Inference: A Practical Information-Theoretic Approach, edited by: Burnham, K. P. and Anderson, D. R., Springer New York, New York, NY, 149–205, ISBN 0-387-95364-7, 2002. a
Caldwell, P. C., Merrifield, M. A., and Thompson, P. R.: Sea level measured by tide gauges from global oceans as part of the Joint Archive for Sea Level (JASL) since 1846, NOAA National Centers for Environmental Information [data set], https://doi.org/10.7289/v5v40s7w, 2001. a, b
Caruso, M. F. and Marani, M.: Extreme-coastal-water-level estimation and projection: a comparison of statistical methods, Nat. Hazards Earth Syst. Sci., 22, 1109–1128, https://doi.org/10.5194/nhess-22-1109-2022, 2022. a, b, c, d, e, f, g
Coles, S.: An introduction to statistical modeling of extreme values, Springer Series in Statistics, Springer-Verlag, London, ISBN 1-85233-459-2, 2001. a, b, c, d
Cook, N. J.: Rebuttal of “Problems in the extreme value analysis”, Struct. Saf., 34, 418–423, 2012. a
Cox, D., Hunt, J., Mason, P., Wheater, H., Wolf, P., Cox, D. R., Isham, V. S., and Northrop, P. J.: Floods: some probabilistic and statistical approaches, Philos. T. Roy. Soc. A, 360, 1389–1408, 2002. a
David, H. A. and Nagaraja, H. N.: Order Statistics, Wiley, Hoboken, New Jersey, ISBN 0-471-38926-9, 2004. a
Davison, A. C. and Smith, R. L.: Models for exceedances over high thresholds, J. Roy. Stat. Soc. B Met., 52, 393–442, 1990. a
De Michele, C.: Advances in Deriving the Exact Distribution of Maximum Annual Daily Precipitation, Water, 11, 2322, https://doi.org/10.3390/w11112322, 2019. a
De Michele, C. and Avanzi, F.: Superstatistical distribution of daily precipitation extremes: A worldwide assessment, Sci. Rep., 8, 14204, https://doi.org/10.1038/s41598-018-31838-z, 2018. a, b
Deidda, R.: A multiple threshold method for fitting the generalized Pareto distribution to rainfall time series, Hydrol. Earth Syst. Sci., 14, 2559–2575, https://doi.org/10.5194/hess-14-2559-2010, 2010. a
Dimitriadis, P. and Koutsoyiannis, D.: Climacogram versus autocovariance and power spectrum in stochastic modelling for Markovian and Hurst–Kolmogorov processes, Stoch. Env. Res. Risk A., 29, 1649–1669, 2015. a
Dimitriadis, P. and Koutsoyiannis, D.: Stochastic synthesis approximating any process dependence and distribution, Stoch. Env. Res. Risk A., 32, 1493–1515, 2018. a
Dimitriadis, P., Koutsoyiannis, D., Iliopoulou, T., and Papanicolaou, P.: A Global-Scale Investigation of Stochastic Similarities in Marginal Distribution and Dependence Structure of Key Hydrological-Cycle Processes, Hydrology, 8, 59, https://doi.org/10.3390/hydrology8020059, 2021. a
Dubey, S. D.: Compound gamma, beta and F distributions, Metrika, 16, 27–31, 1970. a, b
Eichner, J. F., Kantelhardt, J. W., Bunde, A., and Havlin, S.: Extreme value statistics in records with long-term persistence, Phys. Rev. E, 73, 016130, https://doi.org/10.1103/PhysRevE.73.016130, 2006. a, b, c
Eugene, N., Lee, C., and Famoye, F.: Beta-Normal distribution and its applications, Commun. Stat. Theory, 31, 497–512, 2002. a
Fawcett, L. and Green, A. C.: Bayesian posterior predictive return levels for environmental extremes, Stoch. Env. Res. Risk A., 32, 2233–2252, 2018. a, b
Fawcett, L. and Walshaw, D.: Sea-surge and wind speed extremes: optimal estimation strategies for planners and engineers, Stoch. Env. Res. Risk A., 30, 463–480, 2016. a, b
Fisher, R. A. and Tippett, L. H. C.: Limiting forms of the frequency distribution of the largest or smallest member of a sample, Math. Proc. Cambridge, 24, 180–190, 1928. a
Gelman, A., Carlin, J. B., Stern, H. S., and Rubin, D. B.: Bayesian Data Analysis, 2nd ed., Chapman and Hall/CRC, Boca Raton, FL, ISBN 1-58488-388-X, 2004. a
Giorgi, F. and Mearns, L. O.: Calculation of average, uncertainty range, and reliability of regional climate changes from AOGCM simulations via the “Reliability Ensemble Averaging” (REA) Method, J. Climate, 15, 1141–1158, 2002. a
Gnedenko, B.: Sur La Distribution Limite Du Terme Maximum D'Une Serie Aleatoire, Ann. Math., 44, 423–453, 1943. a
Gumbel, E. J.: Statistics of Extremes, Columbia University Press, New York, USA, ISBN 9780231891318, 1958. a
Hisakado, M., Kitsukawa, K., and Mori, S.: Correlated binomial models and correlation structures, J. Phys. A Math. Gen., 39, 15365, https://doi.org/10.1088/0305-4470/39/50/005, 2006. a
Hosseini, S. R., Scaioni, M., and Marani, M.: Extreme Atlantic hurricane probability of occurrence through the Metastatistical Extreme Value distribution, Geophys. Res. Lett., 47, 2019GL086138, https://doi.org/10.1029/2019GL086138, 2020. a
Iliopoulou, T. and Koutsoyiannis, D.: Revealing hidden persistence in maximum rainfall records, Hydrolog. Sci. J., 64, 1673–1689, 2019. a
Iliopoulou, T., Papalexiou, S. M., Markonis, Y., and Koutsoyiannis, D.: Revisiting long-range dependence in annual precipitation, J. Hydrol., 556, 891–900, 2018. a
Jenkinson, A. F.: The frequency distribution of the annual maximum (or minimum) values of meteorological elements, Q. J. Roy. Meteor. Soc., 81, 158–171, 1955. a
Kantelhardt, J. W., Koscielny-Bunde, E., Rybski, D., Braun, P., Bunde, A., and Havlin, S.: Long–term persistence and multifractality of precipitation and river runoff records, J. Geophys. Res.-Atmos., 111, D01106, https://doi.org/10.1029/2005JD005881, 2006. a
Karlis, D. and Xekalaki, E.: Mixed Poisson distributions, Int. Stat. Rev., 73, 35–58, 2005. a
Koutsoyiannis, D.: Statistics of extremes and estimation of extreme rainfall: I. Theoretical investigation, Hydrolog. Sci. J., 49, 575–590, 2004. a, b, c
Koutsoyiannis, D.: Simple stochastic simulation of time irreversible and reversible processes, Hydrolog. Sci. J., 65, 536–551, 2020. a, b
Koutsoyiannis, D.: Stochastics of hydroclimatic extremes – A cool look at risk, third edn., Kallipos, Open Academic Editions, Greece, ISBN 978-618-85370-0-2, 2023. a
Koutsoyiannis, D. and Dimitriadis, P.: Towards Generic Simulation for Demanding Stochastic Processes, Sci, 34, https://doi.org/10.3390/sci3030034, 2021. a
Koutsoyiannis, D. and Montanari, A.: Statistical analysis of hydroclimatic time series: Uncertainty and insights, Water Resour. Res., 43, W05429, https://doi.org/10.1029/2006WR005592, 2007. a, b
Kuczera, G.: Comprehensive at-site flood frequency analysis using Monte Carlo Bayesian inference, Water Resour. Res., 35, 1551–1557, 1999. a
Labat, D., Masbou, J., Beaulieu, E., and Mangin, A.: Scaling behavior of the fluctuations in stream flow at the outlet of karstic watersheds, France, J. Hydrol., 410, 162–168, 2011. a
Leadbetter, M. R.: Extremes and local dependence in stationary sequences, Z. Wahrscheinlichkeitstheorie, 65, 291–306, 1983. a
Leadbetter, M. R., Lindgren, G., and Rootzén, H.: Extremes and Related Properties of Random Sequences and Processes, 1 edn., Springer-Verlag, New York, US, https://doi.org/10.1007/978-1-4612-5449-2, 1983. a, b
Lombardo, F., Volpi, E., and Koutsoyiannis, D.: Rainfall downscaling in time: theoretical and empirical comparison between multifractal and Hurst-Kolmogorov discrete random cascades, Hydrolog. Sci. J., 57, 1052–1066, 2012. a
Lombardo, F., Volpi, E., Koutsoyiannis, D., and Papalexiou, S. M.: Just two moments! A cautionary note against use of high-order moments in multifractal models in hydrology, Hydrol. Earth Syst. Sci., 18, 243–255, https://doi.org/10.5194/hess-18-243-2014, 2014. a
Lombardo, F., Volpi, E., Koutsoyiannis, D., and Serinaldi, F.: A theoretically consistent stochastic cascade for temporal disaggregation of intermittent rainfall, Water Resour. Res., 53, 4586–4605, 2017. a, b
Lombardo, F., Napolitano, F., Russo, F., and Koutsoyiannis, D.: On the Exact Distribution of Correlated Extremes in Hydrology, Water Resour. Res., 55, 10405–10423, 2019. a, b, c
Makkonen, L.: Problems in the extreme value analysis, Struct. Saf., 30, 405–419, 2008. a
Makkonen, L., Pajari, M., and Tikanmäki, M.: Closure to “Problems in the extreme value analysis” (Struct. Safety 2008:30:405–419), Struct. Saf., 40, 65–67, 2013. a
Marani, M. and Ignaccolo, M.: Advances in Water Resources, A metastatistical approach to rainfall extremes, Adv. Water Resour., 79, 121–126, 2015. a, b, c, d, e, f, g, h, i, j, k, l, m, n, o, p, q, r, s, t
Markonis, Y., Moustakis, Y., Nasika, C., Sychova, P., Dimitriadis, P., Hanel, M., Máca, P., and Papalexiou, S. M.: Global estimation of long-term persistence in annual river runoff, Adv. Water Resour., 113, 1–12, 2018. a
Marra, F., Nikolopoulos, E. I., Anagnostou, E. N., and Morin, E.: Metastatistical Extreme Value analysis of hourly rainfall from short records: Estimation of high quantiles and impact of measurement errors, Adv. Water Resour., 117, 27–39, 2018. a, b, c, d, e, f, g, h, i, j, k, l, m
Marra, F., Zoccatelli, D., Armon, M., and Morin, E.: A simplified MEV formulation to model extremes emerging from multiple nonstationary underlying processes, Adv. Water Resour., 127, 280–290, 2019. a, b
Marra, F., Amponsah, W., and Papalexiou, S. M.: Non-asymptotic Weibull tails explain the statistics of extreme daily precipitation, Adv. Water Resour., 173, 104388, https://doi.org/10.1016/j.advwatres.2023.104388, 2023. a, b
Miniussi, A. and Marani, M.: Estimation of daily rainfall extremes through the Metastatistical Extreme Value distribution: Uncertainty minimization and implications for trend detection, Water Resour. Res., 56, e2019WR026535, https://doi.org/10.1029/2019WR026535, 2020. a
Miniussi, A., Marani, M., and Villarini, G.: Metastatistical Extreme Value Distribution applied to floods across the continental United States, Adv. Water Resour., 136, 103498, https://doi.org/10.1016/j.advwatres.2019.103498, 2020. a
Mood, A. M. F., Graybill, F. A., and Boes, D. C.: Introduction to the Theory of Statistics, third edn., McGraw-Hill, New York, ISBN 0-07-042864-6, 1974. a
Moran, P. A. P.: An Introduction to Probability Theory, Oxford science publications, Oxford University Press, New York, ISBN 0-19-853242-3, 1968. a
Morrison, J. E. and Smith, J. A.: Stochastic modeling of flood peaks using the generalized extreme value distribution, Water Resour. Res., 38, 41.1–41.12, 2002. a
Mushtaq, S., Miniussi, A., Merz, R., and Basso, S.: Reliable estimation of high floods: A method to select the most suitable ordinary distribution in the Metastatistical extreme value framework, Adv. Water Resour., 161, 104127, https://doi.org/10.1016/j.advwatres.2022.104127, 2022. a, b
Papalexiou, S. M.: Unified theory for stochastic modelling of hydroclimatic processes: Preserving marginal distributions, correlation structures, and intermittency, Adv. Water Resour., 115, 234–252, 2018. a, b, c
Papalexiou, S. M.: Rainfall generation revisited: Introducing CoSMoS-2s and advancing copula-based intermittent time series modeling, Water Resour. Res., 58, e2021WR031641, https://doi.org/10.1029/2021WR031641, 2022. a, b, c
Papalexiou, S. M. and Serinaldi, F.: Random fields simplified: Preserving marginal distributions, correlations, and intermittency, with applications from rainfall to humidity, Water Resour. Res., 56, e2019WR026331, https://doi.org/10.1029/2019WR026331, 2020. a, b, c
Papalexiou, S.-M., Koutsoyiannis, D., and Montanari, A.: Can a simple stochastic model generate rich patterns of rainfall events?, J. Hydrol., 411, 279–289, 2011. a
Papalexiou, S. M., Serinaldi, F., and Porcu, E.: Advancing space-time simulation of random fields: From storms to cyclones and beyond, Water Resour. Res., 57, e2020WR029466, https://doi.org/10.1029/2020WR029466, 2021. a, b, c
Papalexiou, S. M., Serinaldi, F., and Clark, M. P.: Large-domain multisite precipitation generation: Operational blueprint and demonstration for 1,000 sites, Water Resour. Res., 59, e2022WR034094, https://doi.org/10.1029/2022WR034094, 2023. a
Pickands III, J.: Statistical Inference Using Extreme Order Statistics, Ann. Stat., 3, 119–131, 1975. a
Popper, K. R.: The logic of scientific discovery, Hutchinson & Co., Ltd., London, UK, ISBN 0-203-99462-0, 1959. a
Porporato, A., Vico, G., and Fay, P. A.: Superstatistics of hydro-climatic fluctuations and interannual ecosystem productivity, Geophys. Res. Lett., 33, L15402, https://doi.org/10.1029/2006GL026412, 2006. a
R Development Core Team: R: A Language and Environment for Statistical Computing, R Foundation for Statistical Computing, Vienna, Austria, ISBN 3-900051-07-0, http://www.R-project.org/ (last access: 31 October 2023), 2023. a
Renard, B., Sun, X., and Lang, M.: Bayesian methods for non-stationary extreme value analysis, in: Extremes in a Changing Climate: Detection, Analysis and Uncertainty, edited by AghaKouchak, A., Easterling, D., Hsu, K., Schubert, S., and Sorooshian, S., Springer Netherlands, Dordrecht, 39–95, https://doi.org/10.1007/978-94-007-4479-0, 2013. a, b
Salvadori, G., De Michele, C., Kottegoda, N. T., and Rosso, R.: Extremes in nature: An approach using copulas, Springer, Dordrecht, the Netherlands, ISBN 978-1-4020-4414-4, 2007. a
Serinaldi, F.: Use and misuse of some Hurst parameter estimators applied to stationary and non–stationary financial time series, Physica A, 389, 2770–2781, 2010. a
Serinaldi, F.: Dismissing return periods!, Stoch. Env. Res. Risk A., 29, 1179–1189, 2015. a, b
Serinaldi, F. and Kilsby, C. G.: Simulating daily rainfall fields over large areas for collective risk estimation, J. Hydrol., 512, 285–302, 2014a. a
Serinaldi, F. and Kilsby, C. G.: Rainfall extremes: Toward reconciliation after the battle of distributions, Water Resour. Res., 50, 336–352, 2014b. a
Serinaldi, F. and Kilsby, C. G.: Stationarity is undead: Uncertainty dominates the distribution of extremes, Adv. Water Resour., 77, 17–36, 2015. a
Serinaldi, F. and Kilsby, C. G.: The importance of prewhitening in change point analysis under persistence, Stoch. Env. Res. Risk A., 30, 763–777, 2016a. a, b, c
Serinaldi, F. and Kilsby, C. G.: Understanding persistence to avoid underestimation of collective flood risk, Water, 8, 152, https://doi.org/10.3390/w8040152, 2016b. a
Serinaldi, F. and Kilsby, C. G.: Unsurprising surprises: The frequency of record-breaking and overthreshold hydrological extremes under spatial and temporal dependence, Water Resour. Res., 54, 6460–6487, 2018. a, b
Serinaldi, F. and Lombardo, F.: BetaBit: A fast generator of autocorrelated binary processes for geophysical research, Europhys. Lett., 118, 30007, https://doi.org/10.1209/0295-5075/118/30007, 2017a. a, b
Serinaldi, F. and Lombardo, F.: General simulation algorithm for autocorrelated binary processes, Phys. Rev. E, 95, 023312, https://doi.org/10.1103/PhysRevE.95.023312, 2017b. a, b
Serinaldi, F., Bárdossy, A., and Kilsby, C. G.: Upper tail dependence in rainfall extremes: would we know it if we saw it?, Stoch. Env. Res. Risk A., 29, 1211–1233, 2015. a
Serinaldi, F., Kilsby, C. G., and Lombardo, F.: Untenable nonstationarity: An assessment of the fitness for purpose of trend tests in hydrology, Adv. Water Resour., 111, 132–155, 2018. a, b, c
Serinaldi, F., Chebana, F., and Kilsby, C. G.: Dissecting innovative trend analysis, Stoch. Env. Res. Risk A., 34, 733–754, 2020a. a, b, c, d
Serinaldi, F., Lombardo, F., and Kilsby, C. G.: All in order: Distribution of serially correlated order statistics with applications to hydrological extremes, Adv. Water Resour., 144, 103686, https://doi.org/10.1016/j.advwatres.2020.103686, 2020b. a, b, c, d, e, f, g, h, i, j, k, l, m, n, o, p, q, r
Serinaldi, F., Briganti, R., Kilsby, C. G., and Dodd, N.: Sailing synthetic seas: Stochastic simulation of benchmark sea state time series, Coast. Eng., 176, 104164, https://doi.org/10.1016/j.coastaleng.2022.104164, 2022a. a
Serinaldi, F., Lombardo, F., and Kilsby, C. G.: Testing tests before testing data: An untold tale of compound events and binary dependence, Stoch. Env. Res. Risk A., 36, 1373–1395, 2022b. a, b, c, d
Skellam, J. G.: A probability distribution derived from the binomial distribution by regarding the probability of success as variable between the sets of trials, J. Roy. Stat. Soc. B Met., 10, 257–261, 1948. a
Smith, J. A., Villarini, G., and Baeck, M. L.: Mixture distributions and the hydroclimatology of extreme rainfall and flooding in the Eastern United States, J. Hydrometeorol., 12, 294–309, 2011. a
Smith, J. A., Cox, A. A., Baeck, M. L., Yang, L., and Bates, P.: Strange floods: The upper tail of flood peaks in the United States, Water Resour. Res., 54, 6510–6542, 2018. a
Smith, R. L.: Threshold methods for sample extremes, in: Statistical Extremes and Applications, edited by: Tiago, O. J., Springer Netherlands, Dordrecht, Springer Netherlands, Dordrecht, the Netherlands, 621–638, https://doi.org/10.1007/978-94-017-3069-3, 1984. a
Stedinger, J. R.: Design events with specified flood risk, Water Resour. Res., 19, 511–522, 1983. a
Tahir, M. H. and Cordeiro, G. M.: Compounding of distributions: a survey and new generalized classes, Journal of Statistical Distributions and Applications, 3, 1–35, 2016. a
Todorovic, P.: On Some Problems Involving Random Number of Random Variables, Ann. Math. Stat., 41, 1059–1063, 1970. a, b
Todorovic, P. and Zelenhasic, E.: A Stochastic Model for Flood Analysis, Water Resour. Res., 6, 1641–1648, 1970. a, b
van Montfort, M. A. and van Putten, B.: A comment on modelling extremes: Links between Multi-Component Extreme Value and General Extreme Value distributions, Journal of Hydrology (New Zealand), 41, 197–202, 2002. a
Volpi, E., Fiori, A., Grimaldi, S., Lombardo, F., and Koutsoyiannis, D.: One hundred years of return period: Strengths and limitations, Water Resour. Res., 51, 8570–8585, 2015. a
Volpi, E., Fiori, A., Grimaldi, S., Lombardo, F., and Koutsoyiannis, D.: Save hydrological observations! Return period estimation without data decimation, J. Hydrol., 571, 782–792, 2019. a
Von Mises, R.: La distribution de la plus grande de n valeur, Rev. Math. Union Interbalcanique, 1, 141–160, in: Selected Papers of Richard von Mises: Volume II. Probability and Statistics, General, American Mathematical Society, Providence, Rhode Island, edited by: Frank, P., Goldstein, S., Kac, M., Prager, W., Szegö, G., and Birkhoff, G., 271–294, Library of Congress Catalog Number 63-18572, http://alexander.shen.free.fr/vonMises_64_SelectedPapersVol2OCR.pdf (last access: 25 February 2025), 1936. a
Wang, W., Van Gelder, P. H. A. J. M., Vrijling, J. K., and Chen, X.: Detecting long-memory: Monte Carlo simulations and application to daily streamflow processes, Hydrol. Earth Syst. Sci., 11, 851–862, https://doi.org/10.5194/hess-11-851-2007, 2007. a
Wood, E. F. and Rodríguez-Iturbe, I.: Bayesian inference and decision making for extreme hydrologic events, Water Resour. Res., 11, 533–542, 1975. a
Yadav, R., Huser, R., and Opitz, T.: Spatial hierarchical modeling of threshold exceedances using rate mixtures, Environmetrics, 32, e2662, https://doi.org/10.1002/env.2662, 2021. a
Zorzetto, E. and Marani, M.: Extreme value metastatistical analysis of remotely sensed rainfall in ungauged areas: Spatial downscaling and error modelling, Adv. Water Resour., 135, 103483, https://doi.org/10.1016/j.advwatres.2019.103483, 2020. a
Zorzetto, E., Botter, G., and Marani, M.: On the emergence of rainfall extremes from ordinary events, Geophys. Res. Lett., 43, 8076–8082, 2016. a
- Abstract
- Introduction
- Overview of 𝒩𝒜 distributions of BM
- Modeling extreme values: asking “why” before looking for “how”
- Investigating circular reasoning and redundancy: do we need 𝒩𝒜 distributions of BM in practical applications?
- Smoke and mirrors on the water extremes: a matter of compound distributions, neglected dependence, and misuse of multi-model ensemble averaging
- Discussion
- Conclusions
- Data availability
- Author contributions
- Competing interests
- Disclaimer
- Acknowledgements
- Review statement
- References
- Abstract
- Introduction
- Overview of 𝒩𝒜 distributions of BM
- Modeling extreme values: asking “why” before looking for “how”
- Investigating circular reasoning and redundancy: do we need 𝒩𝒜 distributions of BM in practical applications?
- Smoke and mirrors on the water extremes: a matter of compound distributions, neglected dependence, and misuse of multi-model ensemble averaging
- Discussion
- Conclusions
- Data availability
- Author contributions
- Competing interests
- Disclaimer
- Acknowledgements
- Review statement
- References