the Creative Commons Attribution 4.0 License.
the Creative Commons Attribution 4.0 License.
Multiobjective calibration by combination of stochastic and gradientlike parameter generation rules – the caRamel algorithm
Fabrice Zaoui
Nicolas Moine
Frédéric Hendrickx
Environmental modelling is complex, and models often require the calibration of several parameters that are not able to be directly evaluated from a physical quantity or field measurement.
Multiobjective calibration has many advantages such as adding constraints in a poorly constrained problem or finding a compromise between different objectives by defining a set of optimal parameters.
The caRamel optimizer has been developed to meet the requirement for an automatic calibration procedure that delivers not just one but a family of parameter sets that are optimal with regard to a multiobjective target. The idea behind caRamel is to rely on stochastic rules while also allowing more “local” mechanisms, such as the extrapolation along vectors in the parameter space.
The caRamel algorithm is a hybrid of the multiobjective evolutionary annealing simplex (MEAS) method and the nondominated sorting genetic algorithm II (εNSGAII). It was initially developed for calibrating hydrological models but can be used for any environmental model.
The caRamel algorithm is well adapted to complex modelling. The comparison with other optimizers in hydrological case studies (i.e. NSGAII and MEAS) confirms the quality of the algorithm.
An R package, caRamel
, has been designed to easily implement this multiobjective algorithm optimizer in the R environment.
Environmental modelling is complex, and models often require the calibration of many parameters that cannot be directly estimated from a physical quantity or a field measurement. Moreover, as models' outputs exhibit errors whose statistical structure may be difficult to characterize precisely, it is frequently necessary to use various objectives to evaluate the modelling performance. In other words, it is often difficult to find a rigorous likelihood function or sufficient statistics to be maximized/minimized (Fisher, 1922); for example, it is well known that errors in a simulated discharge time series are not normally distributed, and do not have constant variance or autocorrelation (Sorooshian and Dracup, 1980). In addition, Efstratiadis and Koutsoyiannis (2010) list other advantages of multiobjective calibration such as ensuring parsimony between the number of objectives and the parameters to optimize, fitting distributed responses of models on multiple measurements, recognizing the uncertainties and structural errors related to the model configuration and the parameter estimation procedure, and handling objectives that have contradictory performance.
Multiobjective calibration allows for a compromise between these different objectives to be found by defining a set of optimal parameters. Practical experience shows that singleobjective calibrations are efficient for highlighting a certain property of a system, but this might lead to increasing errors in some other characteristics (Mostafaie et al., 2018). Evolutionary algorithms have been widely used to explore the Paretooptimal front in multiobjective optimization problems that are too complex to be solved by descent methods with classical aggregation approaches. Evolutionary algorithms are advantageous not only because there are few alternatives for searching substantially large spaces for multiple Paretooptimal solutions but also due to their inherent parallelism and capability to exploit similarities of solutions by recombination that enables them to approximate the Paretooptimal front in a single optimization run (Zitzler et al., 2000).
Many studies have used the multiobjective approach in environmental modelling (Oraei Zare et al., 2012; Ercan and Goodall, 2016) or in land use models (Gong et al., 2015; Newland et al., 2018). In hydrology, Madsen (2003) implemented automatic multiobjective calibration of the MIKE SHE model (Refsgaard and Storm , 1995) on the Danish Karup catchment (440 km^{2}) using the shuffled complex evolution (SCE) algorithm (Duan et al., 1992). Yang et al. (2014) ran a multiobjective optimization of the MOBIDIC (Campo et al., 2006) distributed hydrologic model on the Davidson catchment (North Carolina, 105 km^{2}) using the nondominated sorting genetic algorithm II (NSGAII, Deb et al., 2002). More recently Smith et al. (2019) led a multiobjective ensemble approach to hydrological modelling in the UK over 303 catchments for historic drought reconstruction with the GR4J conceptual model (Coron et al., 2017) using Latin hypercube sampling (McKay et al., 1979) and a Paretooptimizing ranking approach accounting for unacceptable tradeoffs (Efstratiadis and Koutsoyiannis, 2010). Mostafaie et al. (2018) compared five different calibration techniques on the GR4J lumped hydrological model using in situ runoff and daily data from the Gravity Recovery And Climate Experiment (GRACE, Tapley et al., 2004). They came to the following conclusions: according to the diversitybased metrics, the NSGAII method is the best approach; according to the accuracy metric, multiobjective particle swarm optimization (MPSO, Reddy and Nagesh Kumar , 2007) is ranked first; and, considering the cardinality measure, the performance of all algorithms is found to be the same.
The caRamel optimizer has been developed to meet the need for an automatic calibration procedure that delivers not only one but a family of parameter sets that are optimal with regard to a multiobjective target (Le Moine, 2009). Madsen (2003) indicated that global populationevolutionbased algorithms are more effective than multistart local search procedures, which, in turn, perform better than purely local search methods. However, most of the multiobjective algorithms rely mainly on stochastic generation rules, with few deterministic aspects, as is the case in the widely used NSGAII for instance. The idea behind caRamel is not just to keep these stochastic “global” mechanisms (such as recombination or multivariate sampling using the covariance) but also to allow more “local” mechanisms, such as extrapolation along vectors in the parameter space, that are associated with an improvement in all objective functions (a “gradientlike” qualitative approach extended to the set of objective functions).
The caRamel algorithm was initially developed and used for the calibration of hydrological models by studies such as Rothfuss et al. (2012), Magand et al. (2014), Le Moine et al. (2015), Monteil et al. (2015), which were all previous to the R package release, or Rouhier et al. (2017), which utilized an R version for the calibration of a hydrologic model over the Loire Basin (35 707 km^{2}). The interesting performance of the caRamel algorithm in such studies prompted us to describe the algorithm in detail in the present paper.
Considering the increasing use of R in hydrology (Slater et al., 2019), we decided to build an R package, caRamel
, for use in any model in the R environment.
The user simply has to define a vectorvalued function (at least two objectives) for the model to calibrate as well as upper and lower bounds for the calibrated parameters.
This paper aims to describe the principles of the caRamel algorithm via an analysis of its results when used for the parametrization of hydrological models. Pieces of code are provided in the Appendix. For an analytical example and for three river case studies, a comparison with the two calibration algorithms that inspired caRamel, the nondominated sorting genetic algorithm II (NSGAII; Reed and Devireddy, 2004) and the multiobjective evolutionary annealing simplex method (MEAS; Efstratiadis and Koutsoyiannis, 2008), is also presented.
The intent of multiobjective calibration is to find sets of parameters that provide a compromise between several potentially conflicting objectives; for instance, how to achieve a good simulation of both flood and lowflow conditions in a hydrological model. Multiobjective calibration is also a means of adding some constraints to an underconstrained problem when many parameters have to be quantified. This can help to reduce the equifinality of parameter sets. Her and Seong (2018) showed that the introduction of an adequate number of objective functions could improve the quality of a calibration without requiring additional observations. The amount of equifinality and the overall output uncertainty decreased while the model performance was maintained as the number of objective functions increased sequentially until four objective functions.
To introduce our notation, Fig. 1 shows a simplified calibration problem in which there is

a model with n_{θ}=2 parameters to calibrate (θ_{1} and θ_{2}). Thus, the model structure is unequivocally represented by the vector $\mathit{\theta}=({\mathit{\theta}}_{\mathrm{1}},{\mathit{\theta}}_{\mathrm{2}})$ in a n_{θ}=2dimensional space, called “parameter space ℰ_{θ}”.

a vector y of n_{y} observed values that should be simulated by the model. For example, for daily times series of 1 year at two gauging stations, ${n}_{y}=\mathrm{2}\times \mathrm{365}=\mathrm{730}$. The simulation is represented by a vector $\widehat{\mathit{y}}\left(\mathit{\theta}\right)$ in a n_{y}dimensional space (that cannot be illustrated graphically), called “observable space ℰ_{y}”.

a vector of n_{f} objective values f(θ,y). For the example in Fig. 1, $\mathit{f}=({f}_{\mathrm{1}},{f}_{\mathrm{2}})$ in a space with n_{f} dimensions, called “objective space ℰ_{f}”.
We will use the following notations: vectors or matrices are presented using bold italic and bold roman font respectively (θ, y, f, Σ…), vector elements and scalars are presented using roman font (θ_{1}, θ_{2}, λ,…), and spaces or ensembles are presented using italic font (ℰ_{θ}, ℱ, 𝒜,…).
Figure 1 also illustrates the relevance of multiobjective calibration with regard to two kinds of equifinality:

equifinality of a structure – the two points θ and θ^{′} that are quite distant in the parameter space ℰ_{θ} become quite near in the observable space ℰ_{y}.

equifinality related to the objective – the vectors θ and θ^{′′} are equifinal regarding f_{1}, and the additional objective f_{2} helps to discriminate them. The use of additional objectives may then help to better constrain the calibration.
The purpose of a multiobjective algorithm is to approach the Pareto front, ℱ, of nondominated solution in the objective space using an ensemble of points called the approximated Pareto front $\widehat{\mathcal{F}}$. We call “archive $\widehat{\mathcal{A}}$” the ensemble of parameter sets from ℰ_{θ} for which simulation outputs are in $\widehat{\mathcal{F}}$.
The caRamel algorithm belongs to the genetic algorithm family. The idea is to start from an ensemble of parameter sets (called a “population”) and to make this population evolve following certain generation rules (Fig. 2). At each generation, new sets are evaluated regarding the objectives, and only the more “suitable” sets are kept to build the new population. The caRamel algorithm is largely inspired by

the multiobjective evolutionary annealing simplex method (MEAS; Efstratiadis and Koutsoyiannis, 2005; Efstratiadis and Koutsoyiannis, 2008), with respect to the directional search method, based on the simplexes of the objective space, and

the nondominated sorting genetic algorithm II (εNSGAII; Reed and Devireddy, 2004), for the classification of parameter vectors and the management of precision by ε dominance.
This section describes the functioning of the caRamel algorithm; this algorithm has been implemented in an R package, caRamel
, that is described in Appendix A.
3.1 Generation rules
The caRamel algorithm has five rules for producing new solutions at each generation: (1) interpolation, (2) extrapolation, (3) independent sampling with a priori parameter variance, (4) sampling with respect to a correlation structure and (5) recombination.
The first two rules (interpolation and extrapolation) are based on a n_{θ}dimensional Delaunay triangulation in the objective space ℰ_{f}. They assume that two neighbouring points in the objective space ℰ_{f} have two adjacent points in the parameter space ℰ_{θ} as antecedents; therefore, one can try to “guess” the directions of improvement in the parameter space from the improvement directions (in a Pareto sense) in the objective space, at least near the optimal zone.
The following two rules create new parameter sets by exploring the parameter space in a nondirectional and less local way – either by independent variations in each parameter or by multivariate sampling using the covariance structure of all parameter sets located near the estimated Pareto front at the current iteration.
Finally, the recombination rule consists of creating new parameter sets using two partial subsets derived from a pair of previously evaluated parameter sets (inspired by Baluja and Caruana, 1995).
3.1.1 Rule 1: interpolation
For rules 1 and 2, we use the notion of simplex which is a generalization of the notion of a triangle to higher dimensions: a 0simplex is a point, a 1simplex is a line segment, a 2simplex is a triangle and a 3simplex is a tetrahedron. A vertex is a point where two or more edges meet. The explanation of the first rule is based on Fig. 3a. First, a triangulation of the points in the objective space ℰ_{f} is established: simplexes built with these points f(θ_{i}) are a partition of the explored zone in this space (Efstratiadis and Koutsoyiannis, 2005).
Let us consider a simplex with at least one vertex on the approximated Pareto front. This simplex is the result of the function f from an ensemble of (n_{f}+1) points from the n_{θ}dimensional parameter space ℰ_{θ}. Under the hypothesis of continuity of f, a linear combination of the form $\stackrel{\mathrm{\u0303}}{\mathit{\theta}}={w}_{\mathrm{1}}{\mathit{\theta}}_{\mathrm{1}}+\mathrm{\dots}+{w}_{({n}_{f}+\mathrm{1})}{\mathit{\theta}}_{({n}_{f}+\mathrm{1})}$, with the barycentric coordinates w_{i}≥0 and ${\sum}_{i}{w}_{i}=\mathrm{1}$, might give a new Paretooptimal solution $\mathit{f}\left(\stackrel{\mathrm{\u0303}}{\mathit{\theta}}\right)$ inside this zone.
First the triangulation is established, then simplex volumes are computed. The probability of generating one new point with a simplex is proportional to its volume when it has at least one point on the Pareto front (otherwise it is zero). If the simplex is selected, then a set of barycentric coordinates are computed by randomly generating (n_{f}+1) values ε_{i} in a uniform distribution on [0,1]:
3.1.2 Rule 2: extrapolation
Extrapolation is based on the same hypothesis of continuity as interpolation. In this case, it is tested to find if an improvement may be obtained by extrapolating from certain directions. These directions are computed from the triangulation by selecting the edges that have only one vertex on to the approximated Pareto front (the second vertex is dominated by the first). These oriented edges computed from the objective space represent directions of improvement in the parameter space (Fig. 3b).
The length $L=\u2225\mathit{f}\left({\mathit{\theta}}_{\mathrm{1}}\right)\mathit{f}\left({\mathit{\theta}}_{\mathrm{2}}\right)\u2225$ of each selected edge and the mean length $\stackrel{\mathrm{\u203e}}{L}$ are computed. The probability of using an edge is proportional to its length L. In this case, the research vector in the parameter space is defined in Eq. (2), and a new parameter set is generated by $\stackrel{\mathrm{\u0303}}{\mathit{\theta}}={\mathit{\theta}}_{\mathrm{1}}+\mathit{\lambda}\mathit{U}$, where λ is a scalar from an exponential distribution with average of 1.
3.1.3 Rule 3: independent sampling with a priori parameter variance
The drawback of the first two rules is that the generation of new vectors is only based on a small number of existing vectors. To compensate for this search by gradient and to avoid convergence toward a local optimum, the third generation rule has two goals:

to make the parameters vary within a larger range than with local rules, and

to make the parameters vary independently of one another.
When considering a vector θ from the archive $\widehat{\mathcal{A}}$, the third rule is to generate n_{θ} new vectors (${\stackrel{\mathrm{\u0303}}{\mathit{\theta}}}_{k}$ with k from 1 to n_{θ}) by making each element of θ (Eq. (3) vary individually, where ${\mathit{\sigma}}_{i}^{\mathrm{2}}$ is the a priori variance of the ith parameter and ε_{i} is a value from a normal distribution (with an average of 0 and a variance of 1). The a priori variance is computed for each parameter from the bounds of variation indicated as the input of the optimizer.
The algorithm selects the n_{θ} vectors that maximize each element of the objective vector individually as well as an additional vector that represents a “central” point of the Pareto front. To select this vector, the minimum of each vector $\mathit{\theta}\in \widehat{\mathcal{A}}$ is computed, and the vector that maximizes this value is chosen.
One generation of this rule then produces $({n}_{f}+\mathrm{1})\times {n}_{\mathit{\theta}}$ new vectors. For this reason, this rule is applied every K generation, with K to be defined by the user. By default, K is computed so that each rule generates the same number of vectors on average.
3.1.4 Rule 4: sampling with respect to a correlation structure
The variance–covariance matrix Σ is computed using Eq. (4), where 𝔼[X] is the expectancy of a random variable X, θ is a vector from the archive 𝒜, $\mathit{\mu}={\mathbb{E}}_{\mathit{\theta}\in \mathcal{A}}\left[\mathit{\theta}\right]$ is the barycentre of 𝒜 and M^{T} is the transpose of the matrix M.
This matrix reflects the correlation structure between the parameter sets. For instance, in the case of a hydrological model, parameters are frequently not independent of each other. This rule intends to obtain an estimate $\widehat{\mathbf{\Sigma}}$ of Σ and $\widehat{\mathit{\mu}}$ of μ in order to generate new parameter vectors that respect this correlation structure and, therefore, limit the risk of generating “nonfunctional” parameter sets.
There are many possibilities in selecting the vector for evaluating the covariance matrix:

Vectors may be selected from a library of “historical” vectors for the calibrated model. The drawback is that this library has to be previously established, and it does not take the progression of the running calibration into account.

Vectors may be selected from the archive $\widehat{\mathcal{A}}$ that provides points on the approximated Pareto front at the running generation. The new vectors frequently improve the front, but, as the variance is low, they do not avoid convergence toward a local optimum.

All vectors of the running population may be selected. This helps to maintain diversity, but it has a high computational cost as few new vectors will make the front progress.
Finally, the algorithm uses a mix between items 2 and 3: all simplexes from the first rule triangulation that have at least one vertex in the approximated Pareto front are selected. Reference vectors for the computation of the variance–covariance matrix are defined by the ensemble 𝒢 from the objective space whose images by f are all the vertices of these simplexes. The estimates $\widehat{\mathbf{\Sigma}}$ and $\widehat{\mathit{\mu}}$ are computed in Eqs. (5)–(6):
This operation increases the number of selected points for the averages computation significantly. However, there is still the risk is of having a variance that is too low. To reduce this risk, the variance of all of the parameters is increased by the same factor (empirically doubled): $\widehat{\widehat{\mathbf{\Sigma}}}=\mathrm{2}\widehat{\mathbf{\Sigma}}$.
The new vectors are obtained from a classical procedure for multivariate generation:

computation of the upper triangular matrix T with ${\mathbf{T}}^{T}\phantom{\rule{0.125em}{0ex}}\mathbf{T}=\widehat{\widehat{\mathbf{\Sigma}}}$, by Cholesky decomposition;

generation of vectors $\stackrel{\mathrm{\u0303}}{\mathit{\theta}}=\widehat{\mathit{\mu}}+{\mathbf{T}}^{T}\cdot \mathit{\epsilon}$, where ε is a vector with n_{θ} independent and normally distributed components with an average of 0 and a variance of 1.
This fourth rule enables us to randomly explore some area of space ℰ_{θ} while implicitly reducing its dimension via the correlations between parameters. This reduces the number of evaluations of the objective function that are needed .
3.1.5 Rule 5: recombination
With respect to rule 4, recombination considers that the parameters from a model are not independent. In a hydrological model, they can frequently be grouped in functional blocks (for instance, rapid runoff, base flow, snow dynamics, transfer and so on). A new parameter vector is simply generated by combining blocks of parameters from vectors of the archive $\widehat{\mathcal{A}}$. The parameter blocks are specific to the calibrated model and are defined by the user.
3.2 Population downsizing
At the end of each generation, the population is kept under a maximum size (N_{max} sets). This limitation is set for memory reasons (no need to keep poor parameter sets) and to reduce computational time, as the triangulation computation is carried out at each generation.
The population downsizing is adapted from εNSGAII (Reed and Devireddy, 2004) and is performed in three steps (Fig. 4):

Pareto ranking – the parameter vectors are sorted according to the ranking order of the Pareto level to which they belong. Points from level 1 are nondominated, points from level 2 are dominated only by points from level 1 and so on.

Downsizing according to the chosen precision – the objective space is partitioned by an n_{f}dimensional grid with the precision δ_{i} for each of the n_{f} objective values. All of the points in the same hypercube are considered to be equifinal with regard to accuracy, and only one point is kept. The selected point is the one that belongs to the lowest Pareto level. When many points are on the lowest level, the selected point is taken at random from among them.

Keeping the population size under N_{max} – if the number of sets is still above N_{max}, only the N_{max} sets of the lower level are kept.
The aim is to assess the performance of the caRamel algorithm against two other optimizers using various case studies. Two optimizers have been selected for the comparison: NSGAII (Deb et al., 2002) and MEAS (Efstratiadis and Koutsoyiannis, 2008). The comparison focuses on different aspects: optimization evolution evaluated by specific metrics and optimization results in the objective space, parameters space and observable space. This section presents the optimizer configuration, the evaluation metrics and the four case studies.
4.1 Optimizer configurations
The caRamel algorithm is used in its general form, with a generation of five new parameters sets for each rule by iteration, involving an average of 25 parameter sets by generation.
NSGAII (Deb et al., 2002) is called by using the nsga2 function from the mco
“Multiple Criteria Optimization Algorithms and Related Functions” (Mersmann et al., 2014) R package. The arguments are the function to minimize, the input and output dimensions, the parameter bounds, the number of generations, the size of the population, and the values for the crossover, mutation probability and distribution index.
Some previous calibration experiments have been conducted to determine the best parameter configurations.
NSGAII has been used with a crossover probability set to 0.5 and mutation probability set to 0.3.
The MEAS algorithm (Efstratiadis and Koutsoyiannis, 2005) combines a performance evaluation procedure based on a Pareto approach and the concept of feasibility, an evolving pattern based on the downhill simplex method, and a simulated annealing strategy, to control randomness during evolution. The algorithm evolution is sensitive to the value of the mutation probability which has been adapted to each case study according to its complexity (5 % for Kursawe and 50 % for the other case studies).
For each optimizer, the end of one optimization is set to a maximum number of model evaluations depending on the case studies. As the algorithms use random functions, 40 optimizations of each test case have been run for each optimizer to obtain representative results. In order to focus on the evolution of the optimization, the initial population is the same for each optimizer (40 initial populations for each case study).
We chose to run an important number of model evaluations and optimizations to get representative results and assess the reproducibility of the optimization. Other benchmark methodology would be conceivable, such as that presented by Tsoukalas et al. (2016) where several test functions and two water resources applications are implemented to compare the surrogateenhanced evolutionary annealing simplex (SEEAS) algorithm to four other monoobjective optimization algorithms. In this study, two alternative computational budgets (indicated by the maximal number of model evaluations) are considered that impact the parameters of the optimizers.
4.2 Optimization metrics
To evaluate the optimizer performance, we chose metrics from the literature. Evaluating optimization techniques experimentally always involves the notion of performance. In the case of multiobjective optimization, the definition of quality is substantially more complex than for singleobjective optimization problems, because the optimization goal itself consists of multiple objectives (Zitzler et al., 2000). Riquelme et al. (2015) categorize the metrics to evaluate three main aspects:

accuracy, which is the closeness of the solutions to the theoretical Pareto front (if known) or relative closeness;

diversity, which can be described by the spread of the set (range of values covered by the solutions) and the distribution (relative distance among solutions in the set);

cardinality, which qualifies the number of Paretooptimal solutions in the set.
To quantify these aspects, we selected three different metrics that are evaluated in the objective space:

hypervolume (HV), which is a volumebased index that takes accuracy, diversity and cardinality into account (Zitzler and Thiele, 1999) and computes the volume between the vectors of the estimated Pareto front $\widehat{\mathcal{F}}$ and a reference point;

generational distance (GD), which is a distancebased accuracy performance index (Van Veldhuizen, 1999, Eq. 7) that is expressed as
$$\begin{array}{}\text{(7)}& \mathrm{GD}={\displaystyle \frac{{\left({\sum}_{i=\mathrm{1}}^{n}{d}_{i}^{\mathrm{2}}\right)}^{\mathrm{1}/\mathrm{2}}}{n}},\end{array}$$where n is the number of vectors in the approximated Pareto front $\widehat{\mathcal{F}}$, and d_{i} is the Euclidean distance between each vector and the nearest member of the reference front;

generalized spread (GS), which evaluates the diversity of the set (Zhou et al., 2006; Jiang et al., 2014).
The evaluation of the GS and GD metrics requires us to establish a reference front. For each case study, this reference front is built by evaluating the Pareto front on all of the final optimization results of all optimizers.
4.3 Case studies
Four case studies have been designed to have an increasing complexity: case study 1 is an analytical example with a Kursawe test function (Kursawe, 1991); case study (2) is on a pluvial catchment with a GR4J opensource hydrological model (Coron et al., 2017, 2019); case study (3) is on a pluvial catchment with a MORDORTS semidistributed model (Rouhier et al., 2017); and case study (4) is on a snowy catchment, also with a MORDORTS model.
4.3.1 The Kursawe test function
The objective of a test function is to evaluate some characteristics of optimization algorithms. The final Pareto front has a specific shape (nonconvex, asymmetric and discontinuous) with an isolated point that the optimizer has to accurately reproduce. The Kursawe function is a benchmark test for many researchers (Lim et al., 2015). It has three parameters (x_{1}, x_{2} and x_{3}) and two objectives (Obj_{1} and Obj_{2}) to minimize (Kursawe, 1991, Eq. 8).
The optimizations are run on 50 000 model evaluations.
The R script to run the Kursawe function optimization with caRamel is available in Appendix B or as a vignette in the caRamel
package.
4.3.2 Calibration of the GR4J model on a pluvial catchment
The GR4J hydrological model is a widely used global rainfall–runoff model (Perrin et al., 2003) that
has been implemented in an opensource R package airGR
(Coron et al., 2017, 2019).
This package contains a data sample from a catchment called “Blue River at Nourlangie Rock” (360 km^{2}, code L0123001), which has a pluvial regime (Fig. 5a).
The advantage of using this case study is in having an opensource script with open data.
GR4J has four parameters to calibrate: the production store capacity X1, the intercatchment exchange coefficient X2, the routing store capacity X3 and the unit hydrograph time constant X4.
The calibration is done on the daily time series for the period from 1990 to 1999. The Kling–Gupta efficiency (KGE, Gupta et al., 2009) is frequently used in hydrology. The KGE can be split into three components that reflect the correlation between the simulated and observed values (KGE_{r}), the bias in standard deviation (KGE_{α}) and the bias in volume (KGE_{β}). The calibration is carried out on these three components (Eq. 9).
where r is the linear correlation coefficient between simulated and observed time series, σ_{s} and σ_{o} represent their standard deviations, and μ_{s} and μ_{o} represent their mean values.
For each component, the optimal value is 1 and the optimization consists of a maximization. At the end of the optimization only the sets with KGE_{β}>0 are considered, as a KGE_{β} with a negative value indicates poor quality for hydrological results. This leads us to exclude a few sets for calibration with NSGAII and caRamel but not for calibration with MEAS.
The R script to run an optimization of the GR4J model with caRamel is available in Appendix C.
4.3.3 Calibration of the MORDORTS model on two contrasting catchments
The spatially distributed MORDORTS rainfall–runoff model (Rouhier et al., 2017) is a spatialized version of the conceptual MORDORSD model (Garavaglia et al., 2017) that has been widely used for operational applications at Électricité de France (EDF; the French electric utility company). The catchment is divided into elementary subcatchments connected according to the hydrographic network which constitutes a hydrological mesh.
This model was implemented at a daily time step for two French catchments with contrasting climates. The Tarn catchment at Millau (Fig. 6a) covers an area of 2335 km^{2} and has a moderate altitude ranging from 350 to 1600 m. The regime is pluvial, with almost no influence from snow. The Durance at La Clapière catchment (2170 km^{2}, Fig. 6b) is located in the French Alps and has elevations ranging from 800 to about 4000 m. Its hydrological regime is strongly influenced by snow, with a maximum during the melting season in June (Fig. 5c).
The hydrological meshes have been built with an average cell area of 100 km^{2}, meaning that 28 cells are needed for the Tarn catchment and 22 cells for the Durance catchment.
MORDORTS has 22 free parameters in its comprehensive formulation. For the Tarn case study, a simplified formulation is adopted with 12 free parameters to calibrate in order to describe the functioning of conceptual reservoirs, evapotranspiration correction and wave celerity (Table 1). For the Durance catchment, parametrization of the snow module of MORDORTS is more complex, and 16 parameters are to be calibrated for the hydrological model. The parameter distribution is uniform for the two case studies, which means that the same set of parameters applies to all cells. Calibration is conducted over 10 years (1 January 1991–31 December 2000) based on three objectives that have to be maximized.
For the Tarn catchment, the calibration is based on the Nash–Sutcliffe efficiencies (NSE; Nash and Sutcliffe, 1970) at three gauging stations: the catchment outlet (Tarn at Millau) and two interior points (Tarn at Montbrun and Dourbie at Gardiès). For the Durance catchment, the Kling–Gupta efficiency (KGE; Gupta et al., 2009) is computed at three gauging stations: the catchment outlet (Durance at La Clapière) and two interior points (Durance at ValdesPrés and Guil at MontDauphin). The theoretical optimum is the point (1, 1, 1) in the objective space.
Four aspects are considered with respect to the results of the case studies: the shape of the final Pareto fronts, the dynamics of the optimizations, the distribution of the calibrated parameters and the consequences of the latter on simulated discharges for the hydrological case studies. To illustrate the results on the simulated discharges, a “best compromise” set has been selected regarding the distance to point (1,1,1) in the objective space for each hydrological case study.
5.1 Final Pareto front
First of all, it is important to accurately reproduce the disconnected Pareto front for the Kursawe test function, and this is the case for all of the optimizers (Fig. 7) with no noticeable differences between the solutions. This confirms the effectiveness of three different algorithms on a lowdimension research benchmark for the multiobjective optimization.
Concerning the three hydrological case studies, the solutions of the Pareto fronts look quite similar for caRamel and NSGAII and more narrow with MEAS (Fig. 8). The number of sets for the Pareto front changes depending on the case, and there is no rank for the optimizers. For the Blue River study, there are 1172 sets with caRamel, 878 sets with NSGAII and 268 points with MEAS; there are 1457, 789 and 1882 sets for the Tarn study, and 708, 408 and 525 sets for the Durance study with caRamel, NSGAII and MEAS respectively. The differences between Pareto fronts are not a priori in favour of a single MEASbased algorithm. They are given for a limited number of cases which are not necessarily representative of a general behaviour.
5.2 Dynamics of the optimizations
Figure 9 summarizes the dynamics of the optimizations for the four case studies.
The caRamel algorithm converges more quickly for accuracy (the HV and GD metrics in most cases). The caRamel's dynamics is closer to NSGAII's dynamics than to MEAS's dynamics, as they have almost the same final values for the three metrics. This confirms the distinctive behaviour between the two classes of algorithms.
With respect to the diversity criteria, GS dynamics is different for the Kursawe test case than for the hydrological case studies. For the Kursawe test case, the optimal final front has a spread, so all optimizers give the same results. For the hydrological cases, the optimal solution is a point (1,1,1); thus, the Pareto front may get smaller with the optimization. NSGAII and caRamel look alike, as they generate more diversity than MEAS (GS final values). On average, caRamel gives better values than NSGAII for the three real cases.
Finally, the envelopes over 40 optimizations are comparable for the three optimizers, which means that reproducibility is always obtained but with different regularities depending on the case or the optimizer without any notable feature. In some cases, a smoother statistical GS convergence would have implied more optimizations.
5.3 Parameter distribution
Figure 10 displays the distribution of parameters from the three case studies.
In the parameter space, the optimizers provide very similar results that explore the equifinality of the model, meaning that different parameter sets show similar performance (Fig. 10). Some parameters (such as kr or lkn) may have optimized values on the whole range defined by the bounds, whereas other parameters are better constrained (X4 and cel). These constitute a family of sets that are optimal with regard to the chosen objectives.
The difference in the diversity of the final sets is also visible in the parameter distributions. Distributions are quite similar for caRamel and NSGAII but are much narrower for MEAS. This once again confirms the different behaviour of MEAS, with weaker general performance for the cases studied here.
5.4 Impact on model results
The consequences with respect to the simulated discharges are displayed in Fig. 11. The envelopes with NSGAII and caRamel are quite similar, whereas the envelope with MEAS is narrower, as expected. This confirms that caRamel and NSGAII generate more diversity on their Pareto front. The red line represents the simulated discharges with the best compromise set and fits quite well with the observed discharges. Multiobjective calibration allows for a range of variation of the calibrated discharges around the best compromise simulation.
Figure 11c represents a flood event on the Tarn River at Millau. The observed discharge points are in the envelope of simulation. The best compromise simulation does not accurately reproduce the flood peak. The figure also displays the simulated discharges obtained by optimizing parameters on the three gauging stations separately, and the simulation with the set that optimizes the NSE at Millau fits better with the observed points.
The caRamel function is an optimization algorithm for multiobjective calibration, and it results in a family of parameter sets that are Paretooptimal with regard to the different objectives. The algorithm is a hybrid of the MEAS algorithm (Efstratiadis and Koutsoyiannis, 2005), using the directional search method based on the simplexes of the objective space, and the εNSGAII algorithm, using the archiving management of the parameter vectors classified by ε dominance (Reed and Devireddy, 2004). The combination of stochastic and gradientlike parameter generation rules helps the convergence of optimization while preserving the diversity of the population in both the objective and parameter spaces. Four case studies with increasing complexity have been used to compare caRamel with NSGAII and MEAS. The results are quite similar between optimizers and show that optimization converges more quickly with caRamel.
An optimization algorithm might be delicate to use because of the choice of input arguments, which are specific to the algorithm and might require some “expert knowledge”. The sensitivity to caRamel internal parameters is not presented in this paper, but we have carried out some sensitivity analyses using the Morris method (Morris, 1991) in order to recommend some default values to the user. First, it is recommended that users assign the same weight to each generation rule by indicating the same number of parameter sets to generate. It is advantageous to produce a small number of sets by generation in order to reduce the number of model evaluations and obtain more rapid convergence. By default, five sets are generated for each rule. The size of the initial population should be large enough to have enough variability (at least 50 sets for a complex model). Moreover, as convergence can be sensitive to the randomly chosen initial population, it is recommended that the user run two or three optimizations to assess reproducibility.
Multiobjective optimization may require thousands of evaluations, which can be a limitation for the calibration of timeconsuming models. To cope with this issue, parallel computation is implemented in the caRamel
R package.
Better consideration of equality or inequality constraints, such as the relationship between two parameters, could be an improvement. Another perspective would be the ability of caRamel to deal with discrete parameters.
caRamel
R package The caRamel
R package has been designed as an optimization tool for any environmental model, provided that it is possible to evaluate the objective functions in R. The main function, caRamel, is called with the following syntax: caRamel(nobj, nvar, minmax, bounds, func, popsize, archsize, maxrun, prec). Arguments are detailed in Table A1.
The main argument of caRamel is the objective function that has to be defined by the user. This enables flexibility, as the user gives all of the necessary information: the number and the definition of all the objectives, the minimization or maximization goal for each objective function, the number of parameters to calibrate and their bounds, and other numerical parameters such as the maximum number of simulations allowed.
Additional optional arguments give the following possibilities:

creation of blocks/subsets of parameters that should be jointly recombined (for example, parameters of a same module);

choice of parallel or sequential computation;

continuation of optimization starting from an existing population;

saving of the population after each generation or only the final one;

indication of the number of parameter sets produced by generation.
As a result, the function returns a list of six elements:

success – a logical, “TRUE” if the optimization process ran with no errors,

parameters – a matrix of parameter sets from the Pareto front (dimension: number of sets in the front, number of calibrated parameters),

objectives – a matrix of associated objective values (dimension: number of sets in the front, number of objectives),

save_crit – a matrix that describes the evolution of the optimization process; for each generation, the first column is the number of model evaluations, and the following columns are the optimum of each objective taken separately (dimension: number of generations, number of objectives +1),

total_pop – total population (dimension: number of parameters sets, number of calibrated parameters + number of objectives).

gpp – the calling period for the third generation rule (independent sampling with a priori parameters variance). It is computed by the algorithm if the user does not fix it.
The R package contains an R vignette that gives optimization examples of the Schaffer function (Schaffer, 1984; two objectives, one parameter) and the Kursawe function (Kursawe, 1991; two objectives, three parameters).
The data analysis was performed with the opensource environment R (https://www.rproject.org; R Core Team, 2019). The algorithm is provided as an R package “caRamel
”, which is available from GitHub (https://doi.org/10.5281/zenodo.3895601; Zaoui and Monteil, 2020) or from CRAN (https://cran.rproject.org/package=caRamel; Le Moine et al., 2019).
The Blue River at Nourlangie Rocks case study was run using the airGR
package for the GR4J hydrological model and for the data set (available at https://cran.rproject.org/package=airGR; Coron et al., 2019).
NLM developed the algorithm in the Scilab platform. FH, FZ and CM adapted the algorithm to an R package and performed the various test cases. CM prepared the paper with contributions from all coauthors.
The authors declare that they have no conflict of interest.
The authors wish to thank the editor and the reviewers of this paper for their constructive suggestions.
Special thanks to Andreas Efstratiadis, who gave us MEAS source code, and to Guillaume Thirel, who provided the first script of airGR
.
This paper was edited by Elena Toth and reviewed by Andreas Efstratiadis and Guillaume Thirel.
Baluja, S. and Caruana, R.: Removing the genetics from the standard genetic algorithm, in: Machine Learning Proceedings 1995, Morgan Kaufmann, San Francisco, USA, 38–46, 1995. a
Campo, L., Caparrini, F., and Castelli, F.: Use of multiplatform, multitemporal remotesensing data for calibration of a distributed hydrological model: an application in the Arno basin, Italy, Hydrol. Process., 20, 2693–2712, 2006. a
Coron, L., Thirel, G., Delaigue, O., Perrin, C., and Andréassian, V.: The Suite of Lumped GR Hydrological Models in an R package, Environ. Model. Softw., 94, 166–171, https://doi.org/10.1016/j.envsoft.2017.05.002, 2017. a, b, c
Coron, L., Delaigue, O., Thirel, G., Perrin, C., and Michel, C.: airGR: Suite of GR Hydrological Models for PrecipitationRunoff Modelling. R package version 1.3.2.23, available at: https://cran.rproject.org/package=airGR (last access: 15 June 2020), 2019. a, b, c
Deb, K., Pratap, A., Agarwal, S. Meyarivan, T.: A fast and elitist multiobjective genetic algorithm: NSGAII, IEEE Trans. Evol. Comput., 6, 182–197, 2002. a, b, c
Duan, Q., Sorooshian, S., and Gupta, V.: Effective and efficient global optimization for conceptual rainfallrunoff models, Water Resour. Res., 28, 1015–1031, 1992. a
Efstratiadis, A. and Koutsoyiannis, D.: The multiobjective evolutionary annealingsimplex method and its application in calibration hydrological models, in: EGU General Assembly 2005, Geophys. Res. Abstr., 7, 04593, https://doi.org/10.13140/RG.2.2.32963.81446, 2005. a, b, c, d
Efstratiadis, A. and Koutsoyiannis, D.: Fitting hydrological models on multiple responses using the multiobjective evolutionary annealing simplex approach, in: Practical hydroinformatics: Computational intelligence and technological developments in water applications, edited by: Abrahart, R. J., See, L. M., and Solomatine, D. P., Springer, 259–273, https://doi.org/10.1007/9783540798811_19, 2008. a, b, c
Efstratiadis, A. and Koutsoyiannis, D.: One decade of multiobjective calibration approaches in hydrological modelling: a review. Hydrolog. Sci. J., 55, 58–78, https://doi.org/10.1080/02626660903526292, 2010. a, b
Ercan M. B. and Goodall, J. L.: Design and implementation of a general software library for using NSGAII with SWAT for multiobjective model calibration. Environ. Model. Softw., 84, 112–120, https://doi.org/10.1016/j.envsoft.2016.06.017, 2016. a
Fisher, R. A.: On the mathematical foundations of theoretical statistics, Philos. T. Roy. Soc. A, 222, 309–368, https://doi.org/10.1098/rsta.1922.0009, 1922. a
Garavaglia, F., Le Lay, M., Gottardi, F., Garçon, R., Gailhard, J., Paquet, E., and Mathevet, T.: Impact of model structure on flow simulation and hydrological realism: from a lumped to a semidistributed approach, Hydrol. Earth Syst. Sci., 21, 3937–3952, https://doi.org/10.5194/hess2139372017, 2017. a
Gong, W., Duan, Q., Li, J., Wang, C., Di, Z., Dai, Y., Ye, A., and Miao, C.: Multiobjective parameter optimization of common land model using adaptive surrogate modeling, Hydrol. Earth Syst. Sci., 19, 2409–2425, https://doi.org/10.5194/hess1924092015, 2015. a
Gupta, H. V., Kling, H., Yilmaz, K., and Martinez, G. F.: Decomposition of the Mean Squared Error & NSE Performance Criteria: Implications for Improving Hydrological Modelling, J. Hydrol., 377, 80–91, https://doi.org/10.1016/j.jhydrol.2009.08.003, 2009. a, b
Her, Y. and Seong, C.: Responses of hydrological model equifinality, uncertainty, and performance to multiobjective parameter calibration, J. Hydroinform., 20, 864–885, https://doi.org/10.2166/hydro.2018.108, 2018. a
Jiang, S., Ong, Y. S., Zhang, J., and Feng, L.: Consistencies and Contradictions of Performance Metrics in Multiobjective Optimization, IEEE Transactions on Cybernetics, 2014. a
Kursawe, F.: A variant of evolution strategies for vector optimization, in: PPSN I, vol. 496, Lect Notes in Comput Sc., SpringerVerlag, Berlin, Germany, 193–197, 1991. a, b, c
Le Moine, N.: Description d’un algorithme génétique multiobjectif pour la calibration d’un modèle pluiedébit, PostDoctoral Status Rep. 2, UPMC/EDF, 13 pp., available at: https://www.metis.upmc.fr/~lemoine/docs/CaRaMEL.pdf (last access: 15 June 2020), 2009 (in French). a
Le Moine, N., Hendrickx, F., Gailhard, J., Garçon, R., and Gottardi, F.: Hydrologically Aided Interpolation of Daily Precipitation and Temperature Fields in a Mesoscale Alpine Catchment, J. Hydrometeorol., 16, 2595–2618, https://doi.org/10.1175/JHMD140162.1, 2015. a
Le Moine, N., Monteil, C., and Zaoui, F.: caRamel: Automatic Calibration by Evolutionary Multi Objective Algorithm, R package version 1.1, available at: https://CRAN.Rproject.org/package=caRamel (last access: 15 June 2020), 2019. a
Lim, W. J., Jambek, A. B., and Neoh, S. C.: Kursawe and ZDT functions optimization using hybrid micro genetic algorithm (HMGA), Soft Comput., 19, 3571–3580, https://doi.org/10.1007/s0050001517675, 2015. a
Madsen, H.: Parameter estimation in distributed hydrological catchment modelling using automatic calibration with multiple objectives, Adv. Water Resour. 26, 205–216, https://doi.org/10.1016/S03091708(02)000921, 2003. a, b
Magand, C., Ducharne, A., Le Moine, N., and Gascoin, S.: Introducing hysteresis in snow depletion curves to improve the water budget of a land surface model in an Alpine catchment. J. Hydrometeorol., 15, 631–649, https://doi.org/10.1175/JHMD13091.1, 2014. a
McKay, M. D., Beckman, R. J., and Conover, W. J.: Comparison ofvthree methods for selecting values of input variables in the analysis of output from a computer code, Technometrics, 21, 239–245, 1979. a
Mersmann O., Trautmann, H., Steuer, D., Bischl, B., and Deb, K.: mco: Multiple Criteria Optimization Algorithms and Related Functions, version 1.015.1, available at: https://CRAN.Rproject.org/package=mco (last access: 15 June 2020), 2014. a
Monteil, C., Hendrickx, F., Samie, R., and Sauquet, E.: Modeling a complex system of multipurpose reservoirs under prospective scenarios (hydrology, water uses, water management): the case of the Durance River basin (South Eastern France, 12 800 km^{2}), Geophys. Res. Abstr., 17, EGU201541211, available at: https://www.researchgate.net/publication/323399497_Modeling_a_complex_system_of_multipurpose_reservoirs_under_prospective_scenarios_hydrology_water_uses_water_management_the_case_of_the_Durance_River_basin_South_Eastern_France_12_800_km2 (last access: 15 June 2020), 2015. a
Morris, M. D.: Factorial sampling plans for preliminary computational experiments, Technometrics, 33, 161–174,1991. a
Mostafaie, A., Forootan, E., Safari, A., and Schumacher, M.: Comparing multiobjective optimization techniques to calibrate a conceptual hydrological model using in situ runoff and daily GRACE data, Comput. Geosci., 22, 789–814, https://doi.org/10.1007/s1059601897268, 2018. a, b
Nash,J. E., and Sutcliffe, J. V.: River flow forecasting through conceptual models part I — A discussion of principles, J. Hydrol., 10, 282–290, https://doi.org/10.1016/00221694(70)902556, 1970. a
Newland, C. P., Maier, H. R., Zecchin, A. C., Newman, J. P., and van Delden, H.: Multiobjective optimisation framework for calibration of Cellular Automata landuse models, Environ. Model. Softw., 100, 175–200, https://doi.org/10.1016/j.envsoft.2017.11.012, 2018. a
Oraei Zare, S., Saghafian, B., and Shamsai, A.: Multiobjective optimization for combined quality–quantity urban runoff control, Hydrol. Earth Syst. Sci., 16, 4531–4542, https://doi.org/10.5194/hess1645312012, 2012. a
Perrin, C., Michel, C., and Andréassian, V.: Improvement of a parsimonious model for streamflow simulation, J. Hydrol., 279, 275–289, https://doi.org/10.1016/S00221694(03)002257, 2003. a
R Core Team: R: A language and environment for statistical computing, R Foundation for Statistical Computing, Vienna, Austria, available at: https://www.Rproject.org/ (last access: 15 June 2020), 2019. a
Reddy, M. J. and Nagesh Kumar, D.: Multiobjective particle swarm optimization for generating optimal tradeoffs in reservoir operation, Hydrol. Process., 21, 2897–2909, https://doi.org/10.1002/hyp.6507, 2007. a
Reed, P. and Devireddy, D.: Groundwater monitoring design: a case study combining epsilondominance archiving and automatic parametrization for the NSGAII, in: Applications of multiobjective evolutionary algorithms, Advances in natural computation series, edited by: CoelloCoello, C., World Scientific, New York, USA, vol. 1, 79–100, https://doi.org/10.1142/9789812567796_0004, 2004. a, b, c, d
Refsgaard, J. C. and Storm, B.: MIKE SHE, in: Computer models of watershed hydrology, edited by: Singh, V. P., Water Resources Publications, Colorado, USA, 809–846, 1995. a
Riquelme, N., Von Lucken, C., and Baran, B.: Performance metrics in multiobjective optimization, XLI Latin American Computing Conference (CLEI), 19–23 October 2015, Arequipa, Peru, https://doi.org/10.1109/CLEI.2015.7360024, 2015. a
Rothfuss, Y., Braud, I., Le Moine, N., Biron, P., Durand, J.L., Vauclin, M., and Bariac, T.: Factors controlling the isotopic partitioning between soil evaporation and plant transpiration: Assessment using a multiobjective calibration of SiSPATIsotope under controlled conditions, J. Hydrol., 442–443, 75–88, https://doi.org/10.1016/j.jhydrol.2012.03.041, 2012. a
Rouhier, L., Le Lay, M., Garavaglia, F., Le Moine, N., Hendrickx, F., Monteil, C., and Ribstein, P.: Impact of mesoscale spatial variability of climatic inputs and parameters on the hydrological response, J. Hydrol., 553, 13–25, https://doi.org/10.1016/j.jhydrol.2017.07.037, 2017. a, b, c
Schaffer, J. D.: Some experiments in machine learning using vector evaluated genetic algorithms (artificial intelligence, optimization, adaptation, pattern recognition), PhD, Vanderbilt University, Computer Science Dept., Nashville, USA, 1984. a
Slater, L. J., Thirel, G., Harrigan, S., Delaigue, O., Hurley, A., Khouakhi, A., Prosdocimi, I., Vitolo, C., and Smith, K.: Using R in hydrology: a review of recent developments and future directions, Hydrol. Earth Syst. Sci., 23, 2939–2963, https://doi.org/10.5194/hess2329392019, 2019. a
Smith, K. A., Barker, L. J., Tanguy, M., Parry, S., Harrigan, S., Legg, T. P., Prudhomme, C., and Hannaford, J.: A multiobjective ensemble approach to hydrological modelling in the UK: an application to historic drought reconstruction, Hydrol. Earth Syst. Sci., 23, 3247–3268, https://doi.org/10.5194/hess2332472019, 2019. a
Sorooshian, S. and Dracup, J. A.: Stochastic parameter estimation procedures for conceptual rainfallrunoff models: Correlated and heteroscedastic error case, Water Resour. Res., 16, 430–442, https://doi.org/10.1029/WR016i002p00430, 1980. a
Tapley, B. D., Bettadpur, S., Watkins, M., and Reigber, C.: The gravity recovery and climate experiment: mission overview and early results, Geophys. Res. Lett., 31, L09607, https://doi.org/10.1029/2004GL019920, 2004. a
Tsoukalas, I., Kossieris, P., Efstratiadis, A., and Makropoulos, C.: Surrogateenhanced evolutionary annealing simplex algorithm for effective and efficient optimization of water resources problems on a budget, Environ. Model. Softw., 77, 122–142, https://doi.org/10.1016/j.envsoft.2015.12.008, 2016. a
Van Veldhuizen, D. A.: Multiobjective Evolutionary Algorithms: Classifications, Analyses, and New Innovations, Faculty of the Graduate School of Engineering of the Air Force Institute of Technology, Air University, Dayton, Ohio, USA, Dissertation AFIT/DS/ENG/9901, 1999. a
Yang, J., Castelli, F., and Chen, Y.: Multiobjective sensitivity analysis and optimization of distributed hydrologic model MOBIDIC, Hydrol. Earth Syst. Sci., 18, 4101–4112, https://doi.org/10.5194/hess1841012014, 2014. a
Zaoui, F. and Monteil, C.: caRamel version 1.1, Zenodo, 2020.https://doi.org/10.5281/zenodo.3895601, 2020. a
Zhou, A., Jin, Y., Zhang, Q., Sendhoff, B., and Tsang, E.: Combining modelbased and geneticsbased offspring generation for multiobjective optimization using a convergence criterion, in: Proc. IEEE Cong. Evol. Comput., 16–21 July 2006, Vancouver, Canada, 892–899, 2006. a
Zitzler, E. and Thiele, L.: Multiobjective Evolutionary Algorithms: A Comparative Case Study and the Strength Pareto Approach, IEEE T. Evolut. Comput., 3, 257–271, 1999. a
Zitzler, E., Deb, K., and Thiele, L.: Comparison of Multiobjective Evolutionary Algorithms: Empirical Results, Evol. Comput., 8, 173–195, 2000. a, b
 Abstract
 Introduction
 Context and notations
 The caRamel algorithm description
 Optimization evaluation framework
 Results of calibration evaluations
 Conclusions

Appendix A: The
caRamel
R package  Appendix B: Example of R script for Kursawe test function optimization
 Appendix C: Example of R script for GR4J optimization
 Code availability
 Author contributions
 Competing interests
 Acknowledgements
 Review statement
 References
 Abstract
 Introduction
 Context and notations
 The caRamel algorithm description
 Optimization evaluation framework
 Results of calibration evaluations
 Conclusions

Appendix A: The
caRamel
R package  Appendix B: Example of R script for Kursawe test function optimization
 Appendix C: Example of R script for GR4J optimization
 Code availability
 Author contributions
 Competing interests
 Acknowledgements
 Review statement
 References