Technical note: A procedure to clean, decompose, and aggregate time series
 Laboratoire des Sciences du Climat et de l'Environnement, LSCE/IPSL, CEACNRSUVSQ, Université ParisSaclay, GifsurYvette 91191, France
 Laboratoire des Sciences du Climat et de l'Environnement, LSCE/IPSL, CEACNRSUVSQ, Université ParisSaclay, GifsurYvette 91191, France
Correspondence: François Ritter (ritter.francois@gmail.com)
Hide author detailsCorrespondence: François Ritter (ritter.francois@gmail.com)
Errors, gaps, and outliers complicate and sometimes invalidate the analysis of time series. While most fields have developed their own strategy to clean the raw data, no generic procedure has been promoted to standardize the preprocessing. This lack of harmonization makes the intercomparison of studies difficult, and leads to screening methods that can be arbitrary or casespecific. This study provides a generic preprocessing procedure implemented in R (ctbi for cyclic/trend decomposition using bin interpolation) dedicated to univariate time series. Ctbi is based on data binning and decomposes the time series into a longterm trend and a cyclic component (quantified by a new metric, the Stacked Cycles Index) to finally aggregate the data. Outliers are flagged with an enhanced box plot rule called Logbox that corrects biases due to the sample size and that is adapted to nonGaussian residuals. Three different Earth science datasets (contaminated with gaps and outliers) are successfully cleaned and aggregated with ctbi. This illustrates the robustness of this procedure that can be valuable to any discipline.
 Article
(3425 KB) 
Supplement
(617 KB)  BibTeX
 EndNote
In any discipline, raw data need to be evaluated during a preprocessing procedure before performing the analysis. Errors are removed, values that deviate from the rest of the population are flagged (outliers, see Aguinis et al., 2013), and in some cases gaps are filled. Because the raw data are altered, preprocessing is a delicate and timeconsuming task that can be neglected due to cognitive biases deflecting our understanding of reality (“I see what I want to see”), or due to our impatience to obtain results. The fate of extreme values is crucial, as they usually challenge scientific or economic theories (Reiss and Thomas, 2007).
Time series are particularly difficult to preprocess (Chandola et al., 2009). A value can or cannot be considered as an outlier just depending on its timestamp (e.g., a freezing temperature in summer), large data gaps are common, abrupt changes can occur, and a background noise covers the true signal. In Earth science, in situ or remote measurements routinely produce time series that first need to be visually inspected. The expert knowledge of the researcher, technician, or engineer is essential to flag suspicious periods of possible instrument failure (e.g., a rain gauge blocked by snowflakes), violation of the experimental conditions (e.g., a passing car during CO_{2} measurements in a forest), or human error (e.g., calibration of the wrong sensor). Once these suspicious periods have been flagged, a preprocessing algorithm is necessary to evaluate the quality of the remaining portion of the measurements. However, there currently is no consensus on which procedure to use even in the simple univariate case: a recent review (Ranjan et al., 2021) covered more than 37 preprocessing methods for univariate time series, and Aguinis et al. (2013) listed 14 different outlier definitions that are mutually exclusive. Despite this (overwhelming) abundance of methods and conventions, there are surprisingly few R packages that offer a preprocessing function. It is worth mentioning hampel (package pracma, Borchers, 2021) that applies a Hampel filter (Pearson, 2002) to time series and flags outliers based on the mean absolute deviation (MAD), which is a robust approximation of the standard deviation defined as MAD$\left(x\right)=\mathrm{1.4826}\times \mathrm{M}(\mathrm{}xM(x\left)\mathrm{}\right)$, with M the median operator. However, the hampel function is not robust to missing values and the scaling factor of 1.4826 is not adapted to nonGaussian residuals. Another option is the function tsoutliers (package forecast, Hyndman and Khandakar, 2008) that applies a seasonal and trend decomposition using loess (STL, Cleveland et al., 1990) to data showing a seasonal pattern, complemented by a smoothing function to estimate the trend of nonseasonal time series (Friedman's super smoother, Friedman, 1984). The residuals obtained can be transformed to follow a Gaussian distribution (Cox–Box method, Box and Cox, 1964), and then outliers are flagged using the box plot rule (Tukey, 1977). This method will be proved in this study to work well with data associated with nearlyGaussian residuals, but to show poor performance otherwise.
This study offers an alternative preprocessing procedure (implemented in R) called ctbi for cyclic/trend decomposition using bin interpolation. The time series is divided into a sequence of nonoverlapping time intervals of equal period (called bins), and outliers are flagged with an enhanced version of the box plot rule (called Logbox) that is adapted to nonGaussian data for different sample sizes. Ctbi fulfills four purposes as follows:
 i.
Data cleaning. Bins with insufficient data are discarded, and outliers are flagged in the remaining bins. If there is a cyclic pattern within each bin, missing values can be imputed as well.
 ii.
Decomposition. The time series is decomposed into a longterm trend and a cyclic component.
 iii.
Cyclicity analysis. The mean cycle of the stacked bins is calculated, and the strength of the cyclicity is quantified by a novel index, the Stacked Cycles Index.
 iv.
Aggregation. Data are averaged (or summed) within each bin.
This procedure is particularly adapted to univariate time series that are messy, with outliers, data gaps, or irregular time steps. The inputs offer a large flexibility in terms of imputation level or outlier cutoff, but also in the timestamp of the bins: a day does not necessarily start at midnight or a year the 1 January. The timeline is not limited to daily or monthly data, but can vary from milliseconds to millenaries. The outputs keep track of the changes brought to the data: contaminated bins are flagged as are outliers and imputed data points.
This paper is divided into two distinct parts. The first part describes the Logbox method, and compares its performance with five other outlier detection methods in the literature based on daily precipitation and temperature data extracted from centuryold weather stations. The second part describes the ctbi procedure, and then applies it to three datasets that have been contaminated beforehand to show the efficiency of the algorithm. A comparison with tsoutliers is performed, and, finally, limitations and good practice recommendations are discussed.
2.1 Context
This first part is dedicated to the detection of outliers present in univariate datasets (without the time component). The box plot (or Tukey's) rule is a commonly used method to flag outliers below a lower boundary l and above an upper boundary u (Tukey, 1977):
with q the sample quantile (e.g., q(0.5) is the median) and α=1.5 a constant that corresponds to 99.3 % of Gaussian data falling within [l,u]. This method is simple and robust to the presence of a maximum of 25 % of outliers in the dataset (known as the breakdown point). When a real data point falls outside the [l,u] range, it is considered as an erroneously flagged outlier (or type I error). Conversely, a type II error occurs when a real outlier is not flagged. The type I error is more common for three reasons as follows:
 i.
For small Gaussian samples (n < 30), up to 8.6 % of data (Hoaglin et al., 1986) can be cut due to the inaccuracy of the sample quantile for small n.
 ii.
For large Gaussian samples (n>10^{3}), α=1.5 is inappropriate because the number of erroneously flagged outliers increases linearly with n due to the 99.3 % of data captured by [l,u].
 iii.
For nonGaussian populations, α=1.5 is generally too restrictive. For example, ∼4.8 % of data following an exponential distribution would be cut.
Studies have corrected biases in the detection of outliers in small samples (see Carling, 2000; Schwertman et al., 2004) and large samples (Barbato et al., 2011), but these methods were adapted to Gaussian populations. For nonGaussian populations, Kimber (1990) and Hubert and Vandervieren (2008) have adjusted α to the skewness (related to the asymmetry of a distribution), but did not consider the kurtosis (related to the tail weight) that will be proven to be a key variable in this study. Therefore, there currently is no generic procedure that can be used when the population is nonGaussian.
To understand how to address this problem, two sets of common distributions with known skewness S, kurtosis excess κ_{ex}, and quantile function Q are used (Fig. 1). The first set is the Pearson family composed of lighttailed distributions that represent any theoretically possible residuals with moderate S and κ_{ex}. Pearson originally worked to create distributions that cover the entire (S, κ_{ex}) space (Pearson, 1895, 1901, 1916), but they took their modern names later on (gamma, inverse gamma, beta prime, Student, Pearson IV). The second set is the generalized extreme value (GEV) family composed of the Gumbel, Weibull, and Fréchet that are heavytailed distributions (high S and κ_{ex}) used in extreme value theory to model the behavior of extrema (Jenkinson, 1955). Based on this framework, this study finds that $\mathit{\alpha}\left(n\right)=A\mathrm{log}\left(n\right)+B+\frac{C}{n}$ reasonably addresses all previously mentioned issues, with C fixed as a constant (C=36). The two parameters A and B correspond to the nature of the distribution, and are estimated based on a predictor of the maximum tail weight and inspired by Moors (1988). A comparison between this procedure (called Logbox) and five other existing models is performed on residuals obtained from 6307 weather stations with more than 100 years of daily temperature and precipitation measurements (Fig. 2). Finally, Logbox is implemented in part II to clean the residuals obtained after fitting the univariate time series with a robust and nonparametric method.
2.2 Method
2.2.1 Distributions
Residuals with moderate κ_{ex} and S are represented in this study with 4999 lighttailed distributions from the Pearson family (Pearson, 1895, 1901, and 1916) composed of the Gaussian, gamma (196 distributions, including the exponential), inverse gamma (170), beta prime (1135), Pearson IV (3377), and Student (120) distributions (Fig. 1a). These distributions cover the entire (κ_{ex},S^{2}) space without overlap, except for the beta distribution that has been discarded due to a bounded support (unrealistic residuals). The shape parameters of each distribution have been chosen to produce regularly spaced points with a mean distance of 0.05 in the (κ_{ex},S^{2}) space, and with a range between the Gaussian and the exponential: ${\mathit{\kappa}}_{\mathrm{ex}}\in [\mathrm{0},\mathrm{6}]$ and $S\in [\mathrm{0},\mathrm{2}]$. Heavytailed residuals are represented with 368 distributions from the generalized extreme value (GEV) family (Fig. 1d) composed of the Gumbel, Weibull (244 distributions), and Fréchet (123). Their shape parameters cover a larger range: ${\mathit{\kappa}}_{\mathrm{ex}}\in [\mathrm{0},\mathrm{500}]$ and $S\in [\mathrm{0},\mathrm{15}]$.
2.2.2 The Logbox model
Based on the box plot rule, α can be defined as
with n the sample size, Q the population quantile function, and f a function that gives the number of erroneously flagged outliers. In the original box plot rule, $Q={\mathrm{\Phi}}^{\mathrm{1}}$ (with Φ the cumulative distribution function of the Gaussian) and f(n)=0.007n, which leads to α=1.5. As explained in the introduction, this choice of f is not valid for large sample sizes due to the linear dependence on n. A flat number of erroneously flagged outliers (f(n)=b) or a logarithmic relationship (f(n)=b log(n)) would not be appropriate either, because α(n) could take arbitrary large values as $\mathrm{1}\frac{f\left(n\right)}{\mathrm{2}n}$ would approach 1 too rapidly (Q(1)=∞). This study suggests instead $f\left(n\right)=\mathrm{0.001}\sqrt{n}$ as a compromise. For example, for a sample of size n=10^{2}, 10^{4} or 10^{6}, respectively 0.01, 0.1, or 1 point would be erroneously flagged as outlier (instead of 0.7, 70, or 7000 points with the original box plot rule). To characterize the relationship α(n) versus n, α is derived with high accuracy (Q implemented in R) for each distribution of the Pearson and GEV family for five sample sizes (n_{i}=10^{i} with $i\in [\mathrm{2},\mathrm{6}]$). It appears that $\mathit{\alpha}\left(n\right)=A\mathrm{log}\left(n\right)+B$ is an accurate model for both the Pearson family (mean of ${r}^{\mathrm{2}}=\mathrm{0.994}\pm \mathrm{0.005}$) and the GEV family (${r}^{\mathrm{2}}=\mathrm{0.99}\pm \mathrm{0.01}$). Barbato et al. (2011) found the same law for the Gaussian distribution based on empirical considerations only, with reported values of A=0.15 and B=1.15. For comparison, this study finds A=0.08 and B=2 for the Gaussian distribution (r^{2}=0.999).
The relationship $\mathit{\alpha}\left(n\right)=A\mathrm{log}\left(n\right)+B$ now needs to be extended to small or nonGaussian samples. To account for biases emerging at small sample size, an additional term is added following Carling (2000): $\mathit{\alpha}\left(n\right)=A\mathrm{log}\left(n\right)+B+\frac{C}{n}$. The parameter C=36 has been numerically determined with a Monte Carlo simulation on the distributions of the Pearson family to ensure that the percentage of erroneously flagged outliers corresponds to ∼0.1 % for n=9 (Supplement). To account for nonGaussian populations, A and B will be estimated with a new robust predictor sensitive to the tail weight. Let (m_{−}, m_{+}) be two functions defined as ${m}_{}=({E}_{\mathrm{3}}{E}_{\mathrm{1}})/({E}_{\mathrm{6}}{E}_{\mathrm{2}}$) and ${m}_{+}=({E}_{\mathrm{7}}{E}_{\mathrm{5}})/({E}_{\mathrm{6}}{E}_{\mathrm{2}})$, with ${E}_{i}=q(i/\mathrm{8})$ the sample octile. The centered Moors $m={m}_{}+{m}_{+}\mathrm{1.23}$ is a known robust predictor of the kurtosis excess with a breakdown point of 12.5 % (Moors, 1988; Kim and White, 2004). However, this study introduces a modified version defined as ${m}_{*}=\mathrm{max}({m}_{},{m}_{+})\mathrm{0.6165}$. The parameter m_{*} is more appropriate than m to determine if a sample is lighttailed or heavytailed. For example, a Gaussian distribution (${m}_{}={m}_{+}\approx \mathrm{0.6165}$) and a rightskewed distribution with one heavy tail (${m}_{}=\mathrm{0.1}$ and ${m}_{+}=\mathrm{1.13}$) will share identical m but different m_{*}. The relationships shown in Fig. 1 are $A={g}_{A}\left({m}_{*}\right)$ and $B={g}_{B}\left({m}_{*}\right)$, with ${g}_{A}\left(x\right)=\mathrm{0.2294}{e}^{\mathrm{2.9416}x\mathrm{0.0512}{x}^{\mathrm{2}}\mathrm{0.0684}{x}^{\mathrm{3}}}$ (r^{2}=0.999) and ${g}_{B}\left(x\right)=\mathrm{1.0585}+\mathrm{15.6960}x\mathrm{17.3618}{x}^{\mathrm{2}}+\mathrm{28.3511}{x}^{\mathrm{3}}\mathrm{11.4726}{x}^{\mathrm{4}}$ (r^{2}=0.999) for $x\in [\mathrm{0},\mathrm{2}]$. Each function has been parameterized based on the Pearson and GEV family together (Fréchet has been excluded due to a different behavior). The coefficients have been determined with a Monte Carlo simulation that minimizes the root mean square error (N∼10^{8}).
For an unknown sample of size n≥9, the Logbox procedure is finally the following: m_{*} is computed (bounded by [0,2]) and the box plot rule is used, with $\mathit{\alpha}\left(n\right)={g}_{A}\left({m}_{*}\right)\mathrm{log}\left(n\right)+{g}_{B}\left({m}_{*}\right)+\frac{\mathrm{36}}{n}$.
2.2.3 Former models
Logbox is compared to five other models (Kimber, 1990; Hubert and Vandervieren, 2008; Schwertman et al., 2004; Leys et al., 2013; Barbato et al., 2011). The first two models (Kim. and Hub.) adjust the box plot method with respect to the skewness as follows:
and
with the function h defined as h(MC)=e^{4MC} for MC < 0 and h(MC)= e^{3MC} for MC ≥ 0. The Medcouple $\mathrm{MC}\in [\mathrm{1},\mathrm{1}]$ is a robust estimator of S, with an algorithm complexity of O(n log n) and a breakdown point of 25 % (Brys et al., 2004). The third model (Sch.) constructs the lower and upper boundary around the median,
with k_{n} a function of the sample size n to adjust for small samples (given as a table in Schwertman et al., 2004) and Z a constant related to the percentage of data captured by $[{l}_{\mathrm{Sch}.},{u}_{\mathrm{Sch}.}]$, here picked as Z=3 (Gaussian case for the ±3σ window). The fourth model (Ley.) uses the MAD around the median,
Finally, the last model (Bar.) is similar to the Logbox procedure but parameterized on the Gaussian distribution only,
2.2.4 Comparison between models
The comparison between models is performed on two sets of residuals obtained from weather stations part of the Global Historical Climatology Network (GHCNdaily) with at least 100 years of daily temperature (2693 stations, 9.4×10^{7} d) or daily precipitation (6277 stations, 5.8×10^{7} wet days, dry days are excluded). Because this network is used to calibrate products that are remotesensingbased and because suspicious values are routinely flagged (Menne et al., 2012a), the risk of errors in these centuryold stations can be considered small. The residuals are extracted with the robust method described in part II based on nonoverlapping bins (bins with less than 80 % of data are discarded). To reduce the impact of the extraction method on the analysis, three bin intervals (5, 10, and 20 d) are used to obtain three replicas for each station. The sensitivity of each outlier detection method to the sample size has also been estimated. For each station and for each sample size l_{i}=10i (i varying from 1 to 10), ${N}_{i}=\phantom{\rule{0.125em}{0ex}}\frac{\mathrm{100}}{i}$ samples are randomly selected, and the number of flagged outliers is summed over all the N_{i} samples (the total number of points is constant, ${N}_{i}\times {l}_{i}=\mathrm{1000})$.
For the five models (Ley., Hub., Kim., Sch., Bar.), the percentage of flagged outliers is computed for each station, and then the mean (±1 SD) is calculated over all stations. For the Logbox model, this method is not appropriate because the expected number of erroneously flagged outlier per station is less than 1 ($\mathrm{0.001}\sqrt{\sim {\mathrm{10}}^{\mathrm{4}}}\sim \phantom{\rule{0.125em}{0ex}}\mathrm{0.1})$. Instead, the percentage of flagged outliers is calculated over the total number of points: $\mathit{\rho}=\left(\mathrm{\Sigma}{n}_{j}^{\mathrm{flagged}}\right)\times \mathrm{100}/\left(\mathrm{\Sigma}{n}_{j}\right)$, with j a station. The variability is estimated by subsampling the total number of stations N_{s}: $\sqrt{{N}_{\mathrm{s}}}$ sets of $\sqrt{{N}_{\mathrm{s}}}$ random stations are selected without replacement. The parameter ρ is computed for each set, and the associated variability is calculated on all ρ values (±1 SD in Fig. 2f and quantiles in Fig. 2c).
2.3 Results and discussion
The parameter α=1.5 used in the box plot rule is sensitive to the sample size n, and the relationship $\mathit{\alpha}\left(n\right)=A\mathrm{log}\left(n\right)+B+\frac{C}{n}$ corrects for this effect for both lighttailed distributions (Pearson family, Fig. 1a) and heavytailed distributions (GEV family, Fig. 1d). The value of A,B, and C depends on the outlier threshold level and the nature of the distribution. The convention in this study is to set the expected number of erroneously flagged outliers to $f\left(n\right)=\mathrm{0.001}\sqrt{n}$, which corresponds to a percentage of type I error of $\mathrm{0.1}/\sqrt{n}$ %. This leads to homogeneous A and B values among the Pearson family ($A=\mathrm{0.8}\pm \mathrm{0.3}$, $B=\mathrm{3}\pm \mathrm{1}$, Fig. 1b, c) used to numerically determine C=36 (Supplement). Because the value of A and B rapidly diverges for heavytailed distributions, a model adapted to the shape of the residuals is required (Fig. 1e, f). To keep this model simple, the asymmetry of a distribution (i.e., the skewness) is ignored in this study in order to only focus on the weight of the heavier tail. Possible outliers might not be flagged on the light tail of an asymmetric distribution (risk of type II error), but residuals with strong asymmetry are usually produced when the range of possible values is semibounded (e.g., precipitation in $[\mathrm{0},+\mathrm{\infty}[)$, which makes the detection of errors trivial (negative precipitation). For this purpose, the parameter m_{*} is a robust predictor of the maximum tail weight with a breakdown point of 12.5 %. Finally, $\mathit{\alpha}\left(n\right)={g}_{A}\left({m}_{*}\right)\mathrm{log}\left(n\right)+{g}_{B}\left({m}_{*}\right)+\frac{\mathrm{36}}{n}$ for n≥9 and ${m}_{*}\in \left[\mathrm{0},\mathrm{2}\right]$, with the functions g_{A} and g_{B} parameterized on both families (Fig. 1e, f). The Fréchet distribution has been excluded because its tails are decaying too rapidly (the A and B coefficients are bounded despite an extreme kurtosis).
The Logbox procedure is tested and compared with five other models on daily precipitation and temperature residuals from centuryold weather stations (Fig. 2). It is firstly visually striking that the outlier threshold from the five traditional methods cut too many data points not only for the precipitation but also for the temperature residuals (Fig. 2a, b). The percentage of flagged data points per station varies around 1.7±1 % for the temperature (Fig. 2d, median of $\mathrm{36}\phantom{\rule{0.125em}{0ex}}\mathrm{634}\approx \mathrm{4}\times {\mathrm{10}}^{\mathrm{4}}$ d per station), and from 4.1 % (Bar.) to 10.5 % (Ley.) for the precipitation (Fig. 2e, median of 8352≈10^{4} wet days per station).
The reason for the large discrepancy between observed and expected percentage of flagged outliers (∼0.7 % based on the box plot rule) is that these methods have been designed for nearlyGaussian residuals. Even daily temperatures are diverging from normality because the fitting model used to extract residuals from the time series minimizes the root mean square error. The anomalies are therefore more concentrated around 0 than those produced by a Gaussian, but with larger extremes (Fig. 2a, leptokurtic distribution). Only the Bar. model correctly captures outliers present in the temperature residuals (0.17 % data points flagged), as it accounts for large sample size effects (logarithmic law in α similar to Logbox). However, Bar. fails at capturing outliers in the precipitation residuals because this method has been parameterized on the Gaussian only. For small samples, the type I error is even higher in all traditional methods due to the inaccuracy of the quantiles: from 1.4 % (Bar.) to 4.2 % (Hub.) of temperature residuals are cut for n=100 (Fig. 2d). This analysis proves that none of the former methods are suitable to detect outliers in nonGaussian residuals.
In comparison, the Logbox procedure shows a percentage of flagged outliers close to the expected values for large sample sizes (Fig. 2c), with 0.0006±0.0003 % for the temperature (expected value of 0.0005 %) and 0.0017±0.0009 % for the precipitation (expected value of 0.001 %). These results are surprisingly accurate knowing that 12.5 % of the extreme values are disregarded for robustness reasons (m_{*}), and also knowing that Logbox has only been parametrized on theoretical distributions (Pearson and GEV family). For smaller sample sizes (n < 30 in Fig. 2f), the precipitation residuals are cut too frequently (∼0.25 %) compared to the expected threshold (∼0.03%), but the temperature residuals are not cut enough. The constant parameter used to correct for a sample size effect (C=36) is only adapted to nearlyGaussian residuals, and it cannot be better estimated because any predictor (such as m_{*}) becomes inaccurate at smaller sample sizes. However, the percentage of flagged outliers remains within 1 order of magnitude of the expected threshold, which is a reasonable compromise between type I errors (precipitation) and type II errors (temperature).
To summarize, Logbox enhances the box plot rule by considering the sample size effect and by adapting the cutting thresholds to the data. This method has been implemented in the function ctbi.outlier (in the R package ctbi) that will be used to flag potential outliers in the residuals obtained by the aggregation procedure described in part II.
3.1 Context
This second part is dedicated to the preprocessing, partial imputation, and aggregation of univariate time series. In order to flag outliers, one first needs to produce residuals that represent the variability around the signal. In its simplest form, the time series y_{t} is represented with the following additive decomposition (Hyndman and Athanasopoulos, 2021): ${y}_{\mathrm{t}}=\phantom{\rule{0.125em}{0ex}}{T}_{\mathrm{t}}+{S}_{\mathrm{t}}+{\mathit{\u03f5}}_{\mathrm{t}}$, with T_{t} a longterm trend, S_{t} a cyclic component (originally seasonal component, but the term cyclic is preferred here as it is more generic), with period τ (∀t, ${S}_{\mathrm{t}}={S}_{\mathrm{t}+\mathit{\tau}})$ and ε_{t} the residuals that are considered to be stationary. A popular algorithm that performs this decomposition is the seasonal and trend decomposition using loess (or STL, Cleveland et al., 1990), that is robust to the presence of outliers. The enhanced version of the algorithm, STLplus (Hafen, 2016), is also robust to the presence of missing values and data gaps. Unfortunately, there are three major drawbacks to using STLplus in the general case: (i) this algorithm has specifically been designed for signals showing seasonal patterns, which makes it less relevant for other types of data; (ii) the longterm trend based on loess needs several input parameters (s.window, s.degree, etc.), and the decomposition is therefore not unique; and (iii) the algorithm has a complexity of O(n^{2}) due to the loess, which is resource intensive and not adapted to long time series (n > 10^{7}). In particular, the first point explains why the function tsoutliers needs to use a smoothing function (Friedman, 1984) to complement the STL procedure.
A new robust and nonparametric procedure (ctbi) is proposed instead to calculate T_{t} and S_{t} using nonoverlapping bins. Outliers are flagged in the residuals ϵ_{t} with the Logbox method described in part I, and imputation is performed using T_{t}+S_{t} if the cyclic pattern is strong enough, which is quantified by a new index introduced in this study (the Stacked Cycles Index or SCI). Bins with sufficient data can finally be aggregated, while other bins are discarded. The procedure is simple (entirely described in Fig. 3), the longterm trend T_{t} is unique and nonparameterized (based on linear interpolations crossing each bin), and the cyclic component S_{t} is simply the mean stack of bins using detrended data (equivalent to STL for periodic time series). The algorithm complexity is of the order of O(n log (n)) because the loess is not necessary anymore. In the following, the procedure is first described more in details and then applied to three case studies (a temperature, precipitation, and methane dataset) that have been contaminated with outliers, missing values, and data gaps. Comparison with the raw data demonstrates the reliability of the ctbi procedure, whose performance is compared to tsoutliers.
3.2 Method
3.2.1 Definitions
Bin
A bin is a time window characterized by a left side (inclusive), a right side (exclusive), a center, and a period (e.g., 1 year in Fig. 3a). Any univariate time series can be decomposed in a sequence of nonoverlapping bins, with the first and last data point contained in the first and last bin, respectively (Fig. 3a). The bin size n_{bin} is the rounded median of the number of points (including “Not Available”, or “NA”, values) present in each nonempty bin. A bin is accepted when its number of nonNA data points is above n_{bin}(1−f_{NA}) with ${f}_{\mathrm{NA}}\in [\mathrm{0},\mathrm{1}]$ the maximum fraction of NA values per bin (input left to the user). Otherwise, the bin is rejected and all its data points are set to NA (Fig. 3a, bin 4).
Longterm trend
The longterm trend (medianbased) is a linear interpolation of the median values associated with each side (calculated between two consecutive centers, see Fig. 3a). A side value is set as missing if the number of nonNA data points (between the two nearest consecutive centers) is below n_{bin}(1−f_{NA}). To solve for boundaries issues and missing sides values, the interpolation is extended using the median value associated with each center (bin 1, 3, and 5 in Fig. 3a). Once the outliers have been quarantined, the longterm trend (meanbased) will be calculated following the same method, but using the mean instead of the median (Fig. 3c).
Cycle
The cycle (medianbased) is composed of n_{bin} points that are the medians of the stack of all accepted bins with the longterm trend (median based) removed (Fig. 3b1). Once the outliers have been quarantined, the cycle (meanbased) will be the mean stack of accepted bins with the longterm trend (meanbased) removed (bin 2, 3, and 5 in Fig. 4a). The cyclic component S_{t} is the sequence of consecutive cycles.
Stacked Cycles Index
SCI ≤1 is an adimensional parameter quantifying the strength of a cycle based on the variability around the mean stack (Fig. 4). Its structure is similar to another index developed in a former study (Wang et al., 2006), however a factor of ${N}_{\mathrm{bin}}^{\mathrm{1}}$ has been added to correct for a bias emerging at a small number of bins (N_{bin} is the number of accepted bins). This correcting factor has been calculated based on stationary time series of Gaussian noise (with therefore a null cyclicity per definition, see Supplement).
3.2.2 Ctbi procedure
Inputs

The univariate time series (first and second column: time and raw data, respectively).

One bin center or one bin side (e.g., 1 June 2020).

The period of the bin (e.g., 1 year).

The aggregation operator (mean, median, or sum).

The range of possible values (default value ${y}_{\mathrm{lim}}\in \phantom{\rule{0.125em}{0ex}}]\mathrm{\infty},+\mathrm{\infty}[)$.

The maximum fraction of NA values per bin (default value f_{NA}=0.2).

The $A,B,C$ coefficients used in the Logbox method (automatically calculated by default, coeffoutlier='auto').

The minimum SCI for imputation (default value SCI_{min}=0.6).
Outputs

The original dataset with nine columns: (i) time; (ii) outlierfree and imputed data; (iii) index of the bins associated with each data points (the index is negative if the bin is rejected); (iv) longterm trend; (v) cyclic component; (vi) residuals (including the outliers); (vii) quarantined outliers; (viii) value of the imputed data points; and (ix) relative position of the data points in their bins between 0 (the point falls on the left side) and 1 (the point falls on the right side).

The aggregated dataset with 10 columns: (i) aggregated time (center of the bins); (ii) aggregated data; (iii) index of the bin (negative value if the bin is rejected); (iv) start of the bin; (v) end of the bin; (vi) number of points per bin (including NA values); (vii) number of NA values per bin, originally; (viii) number of outliers per bin; (ix) number of imputed points per bin; and (x) variability associated with the aggregation (standard deviation for the mean, MAD for the median, and nothing for the sum).

The mean cycle with three columns: (i) time boundary of the first bin with n_{bin} points equally spaced; (ii) the mean value associated with each point; and (iii) the standard deviation associated with the mean value.

A summary of the bins: the Stacked Cycles Index (SCI), the representative number of data points per bin (n_{bin}), and the minimum number of data points for a bin to be accepted (n_{bin min}).

A summary of the Logbox output: the coefficients A, B, and C, m_{*}, the number of points used, and the lower/upper outlier threshold.
Step 1: data screening
The bin size n_{bin} is calculated; values above or below y_{lim} are set to NA; the number of accepted bins N_{bin} is assessed; all data points within rejected bins are set to NA; and the longterm trend and cycle (both median based) are calculated (Fig. 3a, b1).
Step 2: outliers
Outliers are flagged in the residuals (detrended and deseasonalized data) using Logbox (Fig. 3b2); outliers are quarantined and their values are set to NA; the number of accepted bins N_{bin} is updated; and all data points within newly rejected bins are set to NA (bin 1 in Fig. 3c).
Step 3: longterm trend and cycle (mean based)
The longterm trend and the cycle are calculated using the mean instead of the median (Fig. 3c); and SCI is calculated (Fig. 4).
Step 4: imputation
If SCI>SCI_{min}, all NA values in accepted bins are imputed with the longterm trend + the mean cycle (imputation bounded by y_{lim}). Repeat step 3 and step 4 three times to reach convergence.
Step 5: aggregation
Accepted bins are aggregated around their center.
3.2.3 Case studies
Three univariate datasets are chosen to illustrate the potential of the aggregation procedure (Fig. 5, first column). The first dataset is an in situ temperature (in ^{∘}C) measured during summer in the canopy of an Oak woodland of California (month of August, temporal resolution of 5 min) and provided by the National Ecological Observatory Network (NEON 2021, site SJER). The second dataset is an in situ daily precipitation record (in mm) measured at the station of Cape Leeuwin (South westerly coast of Australia) from 1990 to 2020 and available on the Global Historical Climatology Network (Menne et al., 2012a). The last dataset is a methane proxy record (in ppbv) published in Loulergue et al. (2008) that covers 800 000 years with irregular time steps (varying from 1 to 3461 years, with a median of 311 years). None of the datasets contain obvious outliers or a large data gap.
3.2.4 Contamination of the datasets
To test for the robustness of the aggregation procedure, the three raw datasets are contaminated by 30 % (Fig. 5, second column) with the use of three data gaps (20 % of the dataset), random NA values (9.5 % of the dataset), and outliers (0.5 % of the dataset). The three data gaps are picked with random length and position. The position of the outliers and the NA values follows a Poisson law. The value of the outliers is picked equal to ${y}_{\mathrm{min}}\frac{\mathrm{1}}{\mathrm{2}}(\mathit{\mu}{y}_{\mathrm{min}})$ or ${y}_{\mathrm{max}}+\frac{\mathrm{1}}{\mathrm{2}}({y}_{\mathrm{max}}\mathit{\mu})$, with y_{min}, y_{max} and μ respectively the minimum, maximum and mean of the dataset (temperature and methane datasets). The precipitation is supposed to follow a heavytail distribution (extremes are more frequent), and negative values are impossible, which is why outlier values are set to 1.6×y_{max} instead (Supplement).
3.2.5 Aggregation of the datasets
Each dataset (raw and contaminated version) is consecutively aggregated twice (Fig. 6). The temperature dataset is aggregated (using the mean) every hour (n_{bin}=12) and then every day (n_{bin}=24). The precipitation dataset is aggregated (using the sum) every month (n_{bin}=31) and then every year (n_{bin}=12). The methane dataset is aggregated (using the mean) every 2000 years (n_{bin}=4) and then every 20 000 years (n_{bin}=10). For each dataset, the mean cycle of the second level of aggregation is shown in Fig. 6 (second column). The aggregation inputs are chosen as default values. The only exceptions are coeffoutlier=NA and SCI_{min}=NA for the raw data (outliers are not checked, data are not imputed), f_{NA}=1 for the methane dataset (bins with at least one nonNA data point are accepted due to the high irregularity in the sampling frequency), and ${y}_{\mathrm{lim}}=[\mathrm{0},+\mathrm{\infty}[$ for the precipitation dataset (negative precipitation is impossible). The number of false positive (real data points flagged as outliers) and false negative (outliers that have not been flagged) are counted during the first level of aggregation (Table 1), and compared with the tsoutliers function with λ= “auto”, which means that the residuals have been transformed to follow a Gaussian with the Cox–Box method (Box and Cox, 1964), or λ= NULL, which means the original residuals are not transformed. The box plot rule in tsoutliers uses α=3, and the longterm trend or cyclic component are not available for comparison.
3.3 Results and discussion
The three univariate time series have been chosen as case studies due to their various statistical characteristics that are commonly seen in the scientific or economic field (Fig. 5, 1st column). The longterm trend follows smooth or moderate variations in the temperature and precipitation datasets, but shows a much higher volatility in the methane dataset. The cyclic pattern varies from strong diurnal cycles (temperature) and moderate seasonal cycles (precipitation) to no apparent cyclicity over a period of 20 000 years (methane). The detrended and deseasonalized residuals follow distributions from Gaussian (temperature) or seemingly exponential (methane) to heavytailed (precipitation). Finally, the sampling frequency goes from subhourly (temperature) or daily (precipitation) to highly variable (1 to 3461 years, methane). To test the limits of the aggregation procedure, these three datasets are severely contaminated by data gaps, outliers, and missing values (Fig. 5, 2nd column).
The first level of aggregation recovers most of the destroyed signal, with ∼80 % of the bins being accepted for all three datasets (Fig. 6). In these accepted bins, all outliers have been correctly flagged (Table 1). The mean percentage of difference between the contaminated and raw aggregates (level 1) is virtually zero for the temperature (0±0.1 %), the methane ($\mathrm{0.1}\pm \mathrm{2}\phantom{\rule{0.125em}{0ex}}\mathit{\%}$), and the precipitation (0±17 %). For the methane dataset, the only false positive (Table 1) is located at the beginning of the time series (modern time), because the anthropogenic change in CH_{4} is unprecedented when compared to the geological history (the longterm fit does not capture the abrupt increase due to climate change). In comparison, the function tsoutliers successfully flags the outliers in the contaminated temperature and methane datasets (with the Cox–Box method), however it fails with the contaminated daily precipitation dataset (Table 1). This comes from the inability of tsoutliers to handle heavytailed distributions, creating 55 false negatives (all outliers have been missed) with the Cox–Box method and 1125 false positives without it, due to the limitation of the box plot rule using a constant α=3 (see part I).
The second level of aggregation has been performed to test for the cyclicity in the signal (Fig. 6, 2nd column) using the mean cycles and their associated Stacked Cycles Index (Fig. 4). The raw and contaminated mean cycles share similar magnitude within 1 standard deviation on the mean, and their SCI are the same: −0.02 for the methane (no apparent cycles of 20 000 years period), 0.65 for the precipitation (moderate seasonality), and 0.88 for the temperature (strong diurnal cycles). The SCI reveals itself being useful when comparing signals of different nature or periodicities, which is not possible for seasonal indices that only focus on one field (e.g., hydrology) or data format. (e.g., monthly), such as the seasonality index of Feng et al. (2013). The cyclicity seen in the temperature and precipitation is strong enough to impute the missing data in all accepted bins, which further improves the reconstruction of the signal. Because SCI has a similar structure than a coefficient of determination, imputations based on high SCI (> 0.6) are respecting the original signal, which is sometimes not the case with a linear interpolation. These three case studies demonstrate that ctbi is capable of aggregating signals of poor quality that have a stationary variance in the residuals. The next section explains how to handle more complex time series.
3.4 Limits and recommendations
The ctbi procedure complements the expert knowledge related to a dataset, but it does not replace it. In particular, this procedure is not capable of detecting long periods of instrument failure or human error, and it is essential to flag them manually and/or visually before running ctbi. This procedure also presents difficulties to preprocess signals with a complex seasonality associated with residuals of nonstationary variance. A typical example is a daily precipitation record with a pronounced monsoon: several months of droughts (low variability in the signal) are followed by few weeks of severe floods (high variability). These two periods do not have the same statistical characteristics and need to be treated separately. In this situation, two pools of bins can be created using the MAD as a robust indicator of variability within each bin. The procedure is the following: (i) apply ctbi with the median operator (do not flag outliers or impute data, coeffoutlier=NA, and SCI_{min}=NA) so that each bin will be associated with a specific MAD; (ii) flag bins with a low MAD (“dry” season) and a high MAD (“wet” season); (iii) split the raw data into two datasets of bins with a low and high MAD, respectively; (iv) apply ctbi separately to each dataset to flag outliers and/or impute data; and (v) merge the two datasets. This procedure is successfully applied to a soil respiration dataset (Supplement).
Other issues can usually be addressed by varying the inputs: period of the bin, maximum ratio of missing values per bin (f_{NA}), and cyclic imputation level (SCI_{min}). It is recommended to pick the period of a bin so that it contains on average between 4 and ∼50 data points. Below 4 would decrease the breakdown point to unsafe levels (one outlier would be enough to contaminate the bin) and above 50 would produce a longterm trend that might not properly capture the variability in the signal. A maximum of 20 % of the bin can be missing by default (f_{NA}=0.2), but when data are sparse and irregularly distributed, a value of f_{NA}=1 is possible (example with the methane dataset: bins with only one data point were accepted). Finally, the imputation level (default of SCI_{min}=0.6) can vary between 0 (forced imputation even without cyclic pattern) and 1 (no imputation).
Although univariate time series are the simplest type of temporal data, this study reveals a lack of consensus in the literature on how to objectively flag outliers, especially in raw data of poor quality. In part I, a comparison between outlier detection methods is performed on daily residuals from centuryold weather stations (precipitation and temperature data). All traditional outlier detection methods flag extreme events as outliers too frequently (type I error). The alternative procedure developed in this study (Logbox) improves the box plot rule by replacing the original α=1.5 with $\mathit{\alpha}=A\mathrm{log}\left(n\right)+B+\frac{\mathrm{36}}{n}$, with A and B determined with a predictor of the maximum tail weight (m_{*}). Logbox is parameterized on two families of distributions (Pearson and generalized extreme value), and the theoretical percentage of type I error decreases with the sample size (${p}_{\mathrm{theo}}=\frac{\mathrm{0.1}}{\sqrt{n}}\phantom{\rule{0.125em}{0ex}}\mathit{\%}$). Logbox therefore produces cutting thresholds that are tailored to the shape and size of the data, with a good match between observed and expected type I errors in the precipitation and temperature residuals.
In part II, a preprocessing procedure (ctbi for cyclic/trend decomposition using bin interpolation) cleans, decomposes, imputes, and aggregates time series based on data binning. The strength of the cyclic pattern within each bin is assessed with a novel and adimensional index (the Stacked Cycles Index) inspired by the coefficient of determination. The ctbi procedure is able to filter contaminated data by selecting bins with sufficient data points (input: f_{NA}), which are then cleaned from outliers (input: coeffoutlier). The cyclic pattern within each bin is evaluated (SCI) and missing data are imputed in accepted bins if the cyclicity is strong enough (input: SCI_{min}). Most of the signal can be retrieved from univariate time series with diverse statistical characteristics, illustrated in this study with a temperature, precipitation, and methane datasets that have been contaminated with gaps and outliers. Limits in the use of ctbi are acknowledged for signals with a long period of instrument failure, but also for signals presenting a complex seasonality. The last situation can be handled by splitting the raw data into two (or more) datasets containing bins with similar variability quantified by the mean absolute deviation (MAD). The preprocessing procedure is then separately applied to each dataset to correctly identify outliers. It is strongly recommended to examine the data before and after using ctbi to ensure that rejected bins and flagged outliers seem reasonable, and to be transparent about the inputs used in your future study.
The ctbi package is available on the Comprehensive R Archive Network (CRAN). The code used in this study and the Supplement is my own and is available from https://doi.org/10.5281/zenodo.7529126 (Ritter, 2023).
The GHCN dataset is available at https://doi.org/10.7289/V5D21VHZ (Menne et al., 2012b) and detailed in Menne et al. (2012a). The methane dataset is available in Loulergue et al. (2008). The temperature dataset is available at https://doi.org/10.48443/2nt3wj42 (NEON, 2021).
The supplement related to this article is available online at: https://doi.org/10.5194/hess273492023supplement.
The author has declared that there are no competing interests.
Publisher’s note: Copernicus Publications remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
The author would like to warmly thank Rob Hyndman for his advice, as well as the reviewers of this study. Jens Schumacher has particularly stimulated the improvement of the Logbox method.
This paper was edited by Anke Hildebrandt and reviewed by Thomas Wutzler, Jens Schumacher, and one anonymous referee.
Aguinis, H., Gottfredson, R. K., and Joo, H.: Bestpractice recommendations for defining, identifying, and handling outliers, Organizational Research Methods, 16, 270–301, https://doi.org/10.1177/1094428112470848, 2013.
Barbato, G., Barini, E. M., Genta, G., and Levi, R.: Features and performance of some outlier detection methods, J. Appl. Stat., 38, 2133–2149, https://doi.org/10.1080/02664763.2010.545119, 2011.
Borchers, H.: Package “pracma”, https://CRAN.Rproject.org/package=pracma (last access: 1 July 2022), R package version 2.4.2, 2021.
Box, G. E. P. and Cox, D. R.: An analysis of transformations, J. Roy. Stat. Soc. B, 26, 211–243, https://doi.org/10.1111/j.25176161.1964.tb00553.x, 1964.
Brys, G., Hubert, M., and Struyf, A.: A robust measure of skewness, J. Comput. Graph. Stat., 13, 996–1017, https://doi.org/10.1198/106186004X12632, 2004.
Carling, K.: Resistant outlier rules and the nongaussian case, Computational Statistics and Data Analysis, 33, 249–258, https://doi.org/10.1016/S01679473(99)000572, 2000.
Chandola, V., Banerjee, A., and Kumar, V.: Anomaly detection: A survey, ACM Computing Surveys, 41, 1–58, https://doi.org/10.1145/1541880.1541882, 2009.
Cleveland, R. B., Cleveland, W. S., McRae, J. E., and Terpenning, I.: STL: A seasonaltrend decomposition procedure based on loess (with discussion), J. Off. Stat., 6, 3–73, http://bit.ly/stl1990 (last access: 1 December 2021), 1990.
Feng, X., Porporato, A., and RodriguezIturbe, I.: Changes in rainfall seasonality in the tropics, Nat. Clim. Change, 3, 811–815, https://doi.org/10.1038/nclimate1907, 2013.
Friedman, J. H.: A variable span smoother, October, https://doi.org/10.2172/1447470, 1984.
Hafen, R.: Package “stlplus”, https://CRAN.Rproject.org/package=stlplus (last access: 1 July 2022), R package version 0.5.1, 2016.
Hoaglin, D. C., Iglewicz, B., and Tukey, J. W.: Performance of some resistant rules for outlier labeling, J. Am. Stat. Assoc., 81, 991–999, https://doi.org/10.1080/01621459.1986.10478363, 1986.
Hubert, M. and Vandervieren, E.: An adjusted boxplot for skewed distributions, Comput. Stat. Data An., 52, 5186–5201, https://doi.org/10.1016/j.csda.2007.11.008, 2008.
Hyndman, R. J. and Athanasopoulos, G.: (OTexts): Forecasting: principles and practice, 3rd edition, Melbourne, Australia, https://otexts.com/fpp3/ (last access: 21 December 2022), 2021.
Hyndman, R. J. and Khandakar, Y.: Automatic time series forecasting: The forecast package for r, J. Stat. Softw., 27, 1–22, https://doi.org/10.18637/jss.v027.i03, 2008.
Jenkinson, A. F.: The frequency distribution of the annual maximum (or minimum) values of meteorological elements, Q. J. Roy. Meteor. Soc., 81, 158–171, https://doi.org/10.1002/qj.49708134804, 1955.
Kim, T. H. and White, H.: On more robust estimation of skewness and kurtosis, Financ. Res. Lett., 1, 56–73, https://doi.org/10.1016/S15446123(03)000035, 2004.
Kimber, A. C.: Exploratory data analysis for possibly censored data from skewed distributions, Appl. Stat., 39, 56–73, https://doi.org/10.2307/2347808, 1990.
Leys, C., Ley, C., Klein, O., Bernard, P., and Licata, L.: Detecting outliers: Do not use standard deviation around the mean, use absolute deviation around the median, J. Exp. Soc. Psychol., 49, 764–766, https://doi.org/10.1016/j.jesp.2013.03.013, 2013.
Loulergue, L., Schilt, A., Spahni, R., MassonDelmotte, V., Blunier, T., Lemieux, B., Barnola, J. M., Raynaud, D., Stocker, T. F., and Chappellaz, J.: Orbital and millennialscale features of atmospheric CH_{4} over the past 800,000 years, Nature, 453, 383–386, https://doi.org/10.1038/nature06950, 2008.
Menne, M. J., Durre, I., Vose, R. S., Gleason, B. E., and Houston, T. G.: An overview of the global historical climatology networkdaily database, J. Atmos. Ocean. Tech., 29, 897–910, https://doi.org/10.1175/JTECHD1100103.1, 2012a.
Menne, M. J., Durre, I., Korzeniewski, B., McNeill, S., Thomas, K., Yin, X., Anthony, S., Ray, R., Vose, R. S., Gleason, B. E., and Houston, T. G.: Global Historical Climatology Network – Daily (GHCNDaily), version 3.0, NOAA National Climatic Data Center [data set], https://doi.org/10.7289/V5D21VHZ, 2012b.
Moors, J. J. A.: A quantile alternative for kurtosis, The Statistician, 37, 25–32, https://doi.org/10.2307/2348376, 1988.
NEON (National Ecological Observatory Network): Single aspirated air temperature, RELEASE2021 (DP1.00002.001), NEON [data set], https://doi.org/10.48443/2nt3wj42, 2021.
Pearson, K.: X. Contributions to the mathematical theory of evolution. – II. Skew variation in homogeneous material, Philos. T. R. Soc. A, 186, 343–414, https://doi.org/10.1098/rsta.1895.0010, 1895.
Pearson, K.: XI. Mathematical contributions to the theory of evolution. – x. Supplement to a memoir on skew variation, Philos. T. R. Soc. A, 197, 287–299, https://doi.org/10.1098/rsta.1901.0023, 1901.
Pearson, K.: IX. Mathematical contributions to the theory of evolution. – XIX. Second supplement to a memoir on skew variation, Philos. T. R. Soc. A, 216, 538–548, https://doi.org/10.1098/rsta.1916.0009, 1916.
Pearson, R. K.: Outliers in process modeling and identification, IEEE T. Contr. Syst. T., 10, 55–63, https://doi.org/10.1109/87.974338, 2002.
Ranjan, K. G., Prusty, B. R., and Jena, D.: Review of preprocessing methods for univariate volatile timeseries in power system applications, Electr. Pow. Syst. Res., 191, 106885, https://doi.org/10.1016/j.epsr.2020.106885, 2021.
Reiss, R. D. and Thomas, M.: Statistical analysis of extreme values: With applications to insurance, finance, hydrology and other fields: Third edition, Springer, https://doi.org/10.1007/9783764373993, 2007.
Ritter, F.: fritte2/ctbi_article: ctbi article (v1.0.0), Zenodo [code], https://doi.org/10.5281/zenodo.7529126, 2023.
Schwertman, N. C., Owens, M. A., and Adnan, R.: A simple more general boxplot method for identifying outliers, Computational Statistics and Data Analysis, 47, 165–174, https://doi.org/10.1016/j.csda.2003.10.012, 2004.
Tukey, J. W.: Exploratory data analysis by john w. tukey, Biometrics, 33, 131–160, 1977.
Wang, X., Smith, K., and Hyndman, R.: Characteristicbased clustering for time series data, Data Min. Knowl. Disc., 13, 335–364, https://doi.org/10.1007/s106180050039x, 2006.
time series– data recorded at specific time intervals (hours, months, etc.). It cuts time series into small pieces (called bins) and rejects bins without enough data. Errors in each bin are then flagged with a popular method called the
box plot rule, which has been improved in this study. Finally, each bin can be averaged to produce a new time series with less noise, fewer gaps, and fewer errors. This procedure can be generalized to any discipline.
time series– data recorded at specific time intervals...