Evaluation of ﬁltering methods for use on high-frequency measurements of landslide displacements

. Displacement monitoring is a critical control for risks associated with potentially sudden slope failures. In-strument measurements are, however, obscured by the presence of scatter. Data ﬁltering methods aim to reduce the scatter and therefore enhance the performance of early warning systems (EWSs). The effectiveness of EWSs depends on the lag time between the onset of acceleration and its detection by the monitoring system such that a timely warning is issued for the implementation of consequence mitigation strategies. This paper evaluates the performance of three ﬁltering meth-ods (simple moving average, Gaussian-weighted moving average, and Savitzky–Golay) and considers their comparative advantages and disadvantages. The evaluation utilized six levels of randomly generated scatter on synthetic data, as well as high-frequency global navigation satellite system (GNSS) displacement measurements at the Ten-mile land-slide in British Columbia, Canada. The simple moving average method exhibited signiﬁcant disadvantages compared to the Gaussian-weighted moving average and Savitzky–Golay approaches. This paper presents a framework to evaluate the adequacy of different algorithms for minimizing monitoring data scatter.


Introduction
Landslides are associated with significant losses in terms of mortality and financial consequences in countries all over the world.In Canada, landslides have cost Canadians approximately USD 10 billion since 1841 (Guthrie, 2013) and more than USD 200 million annually (Clague and Bobrowsky, 2010).Essential infrastructure, such as railways and roads that play vital roles in the Canadian economy, can be exposed to damage if it transverses landslide-prone areas.Attempting to completely prevent landslides is typically infeasible, as stabilizing options and realignment may be cost-prohibitive or lead to environmental damage.This accentuates the significance of adopting strategies that require constant monitoring to mitigate the consequences of sudden landslide collapses (Vaziri et al., 2010;Macciotta and Hendry, 2021).
In recent years, detailed studies have addressed the use of early warning systems (EWSs) as a robust approach to landslide risk management (Intrieri et al., 2012;Thiebes et al., 2014;Atzeni et al., 2015;Hongtao, 2020).The United Nations defines an EWS as "a chain of capacities to provide adequate warning of imminent failure, such that the community and authorities can act accordingly to minimize the consequences associated with failure" (UNISDR, 2009).Although an EWS comprises various components acting interactively, the core of its performance relies on its ability to detect the magnitude and rate of landslide displacement (Intrieri et al., 2012).Given that the timely response of an EWS determines its effectiveness, an accurate sense of landslide velocity and acceleration is necessary.Monitoring instruments able to provide real-time or near-real-time readings such as global navigation satellite systems (GNSSs) and some remote sensing techniques are satisfactory for this purpose (Yin et al., 2010;Tofani et al., 2013;Benoit et al., 2015;Macciotta et al., 2016;Casagli et al., 2017;Chae et al., 2017;Rodriguez et al., 2017Rodriguez et al., , 2018Rodriguez et al., , 2020;;Huntley et al., 2017;Intrieri et al., 2018;Journault et al., 2018;Carlà et al., 2019;Deane, 2020;Woods et al., 2020Woods et al., , 2021)).These instruments can record the displacement of locations at the surface of the landslide with a high temporal resolution, which allows the monitoring sys-tem to track movements on the order of a few millimeters per year.In practice, the results are usually obscured by the presence of scatter, also known as noise, and outliers that affect the quality of observations.These unfavourable interferences do not reflect the true behaviour of the ground motion and stem from sources such as the external environment and the quality of the communication signals and wave propagation in the case of remote sensing techniques (Wang, 2011;Carlà et al., 2017b).
Scatter can be defined as measurement data that are distributed around the "true" displacement trend such that the average difference between the scatter and the displacement trend is zero and has a finite standard deviation.Scatter in displacement measurements can significantly impact the evaluation of slope movements performed on unfiltered data and decrease the reliability of an EWS.This can lead to false warnings of slope acceleration or unacceptable time lags between the onset of slope failure and its identification and therefore a loss of credibility for an EWS (Lacasse and Nadim, 2009).As a result, scatter should be reduced as much as possible without removing the true slope displacement trends.The application of algorithms that work as filters aims to minimize the amplitude of measured scatter around the displacement trend.
Several approaches have been proposed to filter displacement measurements based on either the frequency or time domain.Fourier and wavelet transformations aim to find the frequency characteristics of the data and then attenuate or amplify certain frequencies.These approaches are discussed in Karl (1989), who suggests they are generally unsuitable for non-stationary data such as monitoring data time series.Filters that work on the time domain can be classified as recursive, kernel, or regression filters.Recursive filters, such as the exponential filtering function, calculate the filtered value at a given time based on the previous filtered value.Kernel filters, which include simple moving average (SMA) and Gaussianweighted moving average (GWMA), calculate the filtered values as the weighted average of neighbouring measurements.Of these two kernel filters, SMA is frequently used in the literature largely due to its simplicity (Dick et al., 2015;Macciotta et al., 2016Macciotta et al., , 2017b;;Carlà et al., 2017aCarlà et al., , b, 2018Carlà et al., , 2019;;Bozzano et al., 2018;Intrieri et al., 2018;Kothari and Momayez, 2018;Chen and Jiang, 2020;Zhou et al., 2020;Desrues et al., 2022;Grebby et al., 2021;Y. H. Zhang et al., 2021;Y. G. Zhang et al., 2021).Regression filters calculate the filtered values by means of regression analysis on unfiltered values (e.g., Savitzky-Golay, or S-G) (Savitzky and Golay, 1964;William, 1979;Cleveland, 1981;Cleveland and Devlin, 1988;Reid et al., 2021).Carlà et al. (2017b) studied both SMA and exponential filtering on multiple failed landslide cases and concluded the latter is inferior in terms of accuracy of failure time prediction.On the other hand, Carri et al. (2021) cautioned the designers and users of EWSs against the use of SMA when rapid movements are expected.However, published applications of filters other than SMA for landslide monitoring are scarce, and studies dedicated to comparing the functionality of other filters to that of SMA are limited.
This paper presents an approach to detect and remove outliers, evaluates the performance of three filters (SMA, GWMA, and S-G), and assesses their suitability to be utilized in an EWS.We evaluated three filters against the following criteria: (1) scatter is minimized, (2) true underlying displacement trends are kept with as little modification as possible, and (3) filtered displacement trends detect acceleration episodes in a timely manner.Moreover, the paper investigates the significance of the time lag between a landslide acceleration event and its identification by a monitoring system for the three filters evaluated.

Synthetic data generation
A numerical analysis on a synthetic dataset approach was adopted, which consists of synthetic dataset scenarios generated to resemble typical landslide displacement measurements, including acceleration and deceleration periods.These scenarios are idealizations based on observations of typical landslide displacements published in the literature (Leroueil, 2001;Intrieri et al., 2012;Macciotta et al., 2016;Schafer, 2016;Carlà et al., 2017a;Scoppettuolo et al., 2020).A total of 12 dimensionless scenarios were built, with all data between the coordinates x = 0, y = 0 and x = 1, y = 1.The x value represents time, and normalization between 0 and 1 allows for extrapolation of the findings for variable displacement measurement frequencies (e.g., the full range of x could represent a week, a month, a year).The analysis of synthetic data focuses on the ability of different algorithms to minimize scatter and identify changes in measured trends; therefore, y represents any of the displacement measurement metrics of interest (displacement, cumulative displacement, velocity, inverse velocity, etc.).Mathematical equations and graphical illustrations of the 12 scenarios are shown in Fig. 1.
The first nine scenarios are referred to as harmonic scenarios, which are characterized by gradual changes in the trend of parameter y.The remaining three scenarios show sudden variations at or near x = 0.5 and are referred to as instantaneous scenarios.Considering the discrete nature of instrument measurements, and to account for different ranges in measurement frequencies, each scenario was generated several times, each time with a different number of points (Table 1).
The next step was adding random scatter to the scenarios to represent unfiltered displacement measurements.Macciotta et al. (2016) show the scatter in displacement monitoring for a GNSS used in their analyses fitted a Gaussian distribution.We validated that the scatter distribution fit approximates a Gaussian distribution for the displacement data  5.8 months 500 000 0.9 years 750 000 1.4 years 1.00 × 10 6 1.9 years scatter of the case study in this paper.This assumption, however, has an underpinning theoretical base established by the central limit theorem in probability theory.It states that the mathematical summation of independent variables (such as scatter) goes toward a Gaussian distribution (Smith, 2013).As a result, the scatter was randomly produced from a normal distribution centered at 0, with extreme values truncated between −1 and 1 and a standard deviation of 0.20.Random generation of the scatter followed the techniques outlined in Clifford (1994) known as the acceptance-rejection method, which generates scatter values through a series of iterations until the algorithm generates the initial normal distribution.
The amplitude of the scatter around the trend in parameter y was defined for each scenario by scaling the randomly generated scatter.This allowed for the investigation of the effect of different scatter magnitudes on the performance of the filters.Scaling was done by defining the ratio n/t, which is the ratio of scatter amplitude (maximum deviation around the trend, termed n) to the range of values of the trend (t) in each scenario.Six levels of n/t (0.001, 0.005, 0.010, 0.050, 0.  and 0.150) were considered when performing the analysis to cover a range of possible levels of scatter in unfiltered measurements.Figure 2 shows two samples of synthetic unfiltered scenarios that are the result of superimposing scatter with n/t values of 0.05 and 0.10 on scenario no. 7.

Simple moving average
SMA is a well-known method for scatter reduction that attempts to reduce scatter by calculating the arithmetic mean of neighbouring points' values.A constant-length interval (window or bandwidth) is used for the calculation for each point; this is also termed a "running" average.Equation ( 1) is the formulation of this method, which was used by Macciotta et al. (2016) to analyze GNSS data scatter: where ŷi is the filtered value, y j is the unfiltered value, and p is the window length.The window length is constant across the dataset except for regions near the boundaries where fewer points are available.Accordingly, p will be adjusted to the number of available points that are indeed less than the value set by the user.This will cause variation in the effectiveness of the method at the extremes, which needs to be considered when evaluating the results of this approach.

Gaussian-weighted moving average
Varying the weights of the measurements within the calculation window in SMA can be used to develop different filtering methods.The largest weight can be given to the measurement at the time for which the calculation is being done, with weights decreasing for measurements farther away in time.
One simple weighting function that can be adopted is the Gaussian (normal) distribution.Equation ( 2) is the formulation of the Gaussian-weighted moving average (GWMA): where w j is the weight coefficient based on the Gaussian distribution, and the other terms follow the same definition as per SMA.

Savitkzy-Golay
S-G fits a low-degree polynomial equation to the unfiltered measurements within a window and defines the filtered measurements using the fitted curve (Schafer, 2011).Although this procedure seems dissimilar to the weighted averaging as discussed for GWMA, its function can be transformed into a kernel concept using the least-squares method if the data points are evenly spaced.The detailed procedure is presented in Appendix A. Figure 3 shows the weight kernel over a window of seven points attained by fitting a quadratic polynomial.An immediate observation is that some points are given negative weights.If points are not evenly spaced, the weighting kernel cannot be used, and local regression analysis should be periodically conducted for each point.Such filtering is known as locally estimated scatterplot smoothing (LOESS).This decreases the computational efficiency of filter performance and exponentially increases the execution time.

Evaluation of processing algorithms
The synthetic monitoring data and data from the case studies were filtered using SMA, GWMA, and S-G techniques.
The filters were applied with different lengths of moving windows, from 0.01 (1 %) to 0.1 (10 %) of all monitoring points, referred to as the bandwidth ratio.These limits for the bandwidth ratio were selected based on literature reports for SMA.In the filtration process, we only used the points prior to the time for which the calculation is being made (point of interest, Fig. 4).This is to reflect the reality of displacement monitoring information as applied to EWSs.To this end, filters used the first half of their kernels, but the weights were multiplied by 2 in comparison to a symmetric window in order to keep the sum of weights equal to 1.All of these filters require the definition of the bandwidth.A roughness factor was defined to aid in the evaluation of the effect of bandwidth in reducing scatter.This factor is defined as follows: where J 2 is the roughness factor, ŷ is the second derivative of filtered measurements, R a is the absolute roughness computed by Eq. ( 4), and y is the second derivative of unfiltered measurements.The second derivative measures how much the slope of the line connecting two consecutive points changes, which itself is an indication of fluctuation.
The greater this second derivative, the greater the variation.J 2 was normalized to the overall curvature of the unfiltered scenario to determine the relative scatter reduction after the application of a filter, eliminating any roughness associated with the real trend in the scenario.In limit states, a value of 1 means that fluctuations are similar to the unfiltered dataset, and therefore no improvement has been achieved; a value of 0 suggests the slope of a scenario remains unchanged and indicates a linear trend.Because all the scenarios, except the first, include trends showing concavity or convexity, a residual value for the roughness factor would be expected in the lowest limit state, meaning that a value of 0 is not necessarily a goal.J 2 was used to infer the minimum value of bandwidth ratio after which no significant change in the fluctuation of results is achieved.Considering the second power in the formulation of J 2 , all observations are valid if the scenarios are mirrored (when they vary from 1 to 0, instead of 0 to 1).The filters are not expected to remove all scatter, and the error attributed to the residual scatter can be calculated using the root mean square error (RMSE).Given that velocity values are usually used as thresholds in an EWS, one concern is whether the filter should be applied to displacement values or velocity values derived from unfiltered displacements.To address this issue, two different approaches to filtering were investigated: direct and indirect.As a result, two different approaches using the RMSE were also utilized here.

Direct scatter filtration
Direct filtration means the filter is applied to the diagram of interest.If the filtered displacement values are the goal, and the filter is applied to unfiltered displacement values, then the filtering process is called direct filtration.The same concept applies when velocity values are derived using unfiltered displacements, and the filters are then directly applied to the velocity values.In this approach, the RMSE follows Eq. ( 5): where RMSEd is the measurement of error in direct filtration, y i is the value of the true trend (for the synthetic scenario), ŷi is the filtered value, and m is the total number of points.This approach is often used in the literature (e.g., Macciotta et al., 2016;Carlà et al., 2017aCarlà et al., ,b, 2018Carlà et al., , 2019;;Intrieri et al., 2018).

Indirect scatter filtration
Some EWSs can apply the filter to the displacements but use velocity trends as the metric for evaluation.In this case, the filtered velocity values will be computed using the filtered displacements.Indirect filtration indicates the diagram of interest is the first derivative of the diagram to which the filter is applied.The RMSE, in this case, is defined as follows: where RMSEi is the measurement of error in indirect filtration, y i is the first derivative of the true trend, ŷ i is the first derivative of filtered data (derived velocity after the filter is applied to the displacements), and m is the total number of points.Similar to J 2 , all observations are valid for the mirrored scenarios of those presented in Fig. 1.This is a consequence of using the second power in the definition of RMSEi and RMSEd.

Lag quantification
Only antecedent measurements are fed into the filters, which is expected to result in a lag between the true trend and its identification by the filters.This lag means the calculated value of velocity or displacement occurred sometime in the past.Consequently, reducing this lag means less time is lost with respect to providing an early warning.To quantify the induced lag, the filtered diagrams of all scenarios at all n/t ratios and bandwidth ratio values were shifted backwards a number of points equivalent to 0.001 (0.1 %) to 0.1 (10 %) of all generated points.We refer to this as the shift ratio in the rest of this paper.This shift of filtered diagrams is expected to increase their similarity with the true trend until the best correlation is achieved.The R 2 test was used to determine how well the shifted and filtered results replicate the underlying trend.

Geocube differential GNSS system
A Geocube system is a network of differential global navigation satellite system (GNSS) units that work with a single frequency (1572.42MHz), making it cost-effective (Dorberstein, 2011; Benoit et al., 2014;Rodriguez et al., 2018).Geocubes communicate with each other through radio frequency, and a reference unit outside the boundaries of the landslide is assumed as static for differential correction to increase the poor accuracy associated with single frequency GNSSs (Benoit et al., 2014;Rodriguez et al., 2018).The ability of this system to achieve real-time positioning, remote data collection, and processing makes it a suitable candidate for incorporation into an EWS.As a result, Geocube data are used in this study to evaluate the performance of the three mentioned filters.

Outlier detection
Outliers are defined herein as abnormal inconsistencies (e.g., displacement directions, magnitudes) when compared to the majority of observations in a random sampling of data (Zimek and Filzmoser, 2018).Techniques for outlier detection have been proposed based on the statistical characteristics of datasets.One common example is the Z score method, which calculates the mean and standard deviation of data within a defined interval and identifies outlier data as those beyond 3 standard deviations from the mean (Rousseeuw and Hubert, 2011).A limitation of this kind of approach is the sensitivity of the mean and standard deviation to the outlier data points, which has led to the development of other methods that use other indices such as the median (Salgado et al., 2016).One such technique that was adopted in this study is the Hampel filter (Hampel, 1971).In this method, the median of the displacement measurements within a running bandwidth is calculated, and data outside a defined threshold from the median are identified as outliers.The threshold is defined as a constant (threshold factor) multiplied by the median absolute deviation.An asymmetric window with a bandwidth ratio of 0.004 (0.4 %) and a threshold factor of 3 was adopted following previous studies (Davies and Gather, 1993;Pearson, 2002;Liu et al., 2004;Yao et al., 2019).The data identified as outliers were then removed from the dataset.
3 Study site -Ten-mile landslide The Ten-mile landslide is located in southwestern British Columbia (BC), in the Fraser River Valley north of Lillooet (Fig. 5a).It is a reactivated portion of a post-glacial earthflow (Bovis, 1985) that was first recognized in the 1970s.The landslide velocity has increased from an average of 1 mm d −1 in 2006 to 6 mm d −1 in 2016, with a maximum measured velocity of 10 mm d −1 (Gaib et al., 2012;BGC Engineering Inc., 2016).The movement of this landslide impacts the integrity of BC Highway 99 and a section of railway operated by Canadian National Railway (CN) (Carlà et al., 2018), with most movement limited to the volume downslope from the railway due to the installation of a retaining wall (Macciotta et al., 2017a).Despite the stabilization work done to date, the uppermost tension crack has retrogressed approximately 200 m in 45 years and is now situated 60 m upslope of the railway track (Macciotta et al., 2017b).The landslide lateral extents have not expanded since 1981 according to the aerial photographs (Macciotta et al., 2017b).The Ten-mile landslide is currently approximately 200 m wide, 140 m high, and has a volume of 0.75 to 1 million m 3 , moving towards the Fraser River on a continuous rupture surface with a dip of about 22 to 24 • , which is sub-parallel to the ground surface (Rodriguez et al., 2017;Donati et al., 2020).The elevation of the shear surface and mechanism of the landslide have been inferred from the readings of multiple slope inclinometers installed in 2015 (BGC Engineering Inc., 2015).The bedrock in this region consists of volcanic rocks, such as andesite, dacite, and basalt, and is overlain by Quaternary deposits (Donati et al., 2020;Carlà et al., 2018;Macciotta et al., 2017a).The thickness of the landslide varies between 20 and 40 m, and the ground profile from the surface to depth comprises medium to high plastic clays and silts overlying colluvium material and glacial deposits, overlying bedrock (BGC Engineering Inc., 2015).The stratigraphy of the sedimented soils in the landslide area notably varies from one borehole to another and reflects the complex stratigraphy of the earthflow.(Rodriguez et al., 2018;Macciotta et al., 2017b).
A total of 11 Geocubes were installed at the Ten-mile landslide in 2016.Figure 5b is a front view of the landslide showing the locations of the Geocube units.Units 44 and 50 are installed near the uppermost tension crack identified as the current landslide backscarp, unit 69 is 30 m above the backscarp, and unit 39 is used as the reference point.Please note that unit 69 is used as the fixed Geocube and is not shown in Fig. 5b.The other units are located within the boundaries of the landslide, with a maximum distance between units of 310 m (Rodriguez et al., 2018).The time step between every two consecutive measurements is 60 s. Figure 6 shows the displacements of units 46 and 47, which were the largest in comparison to other Geocubes.

Synthetic analysis
Figure 7 shows the roughness value (J 2 ) of scenario 6 for SMA, GWMA, and S-G on a semi-logarithmic scale.This figure illustrates how, regardless of the n/t ratio, J 2 substantially decreases as the bandwidth ratio increases to 0.01 and then asymptotically approaches a final value.This means that increasing the bandwidth ratio drastically reduces scatter; however, its effectiveness is restricted as the bandwidth ratio increases above 0.01.This observation was consistent for other scenarios.J 2 values (including scenario 6 in Fig. 7) indicate that J 2 approaches its minimum at bandwidth ratio values of 0.03 to 0.04, regardless of the filter selected.

Effect of filters on trend distortion
Scenarios 11 and 12 were first analyzed to evaluate the degree to which the trend was preserved by these filters as peaks made it easier for visualization.Figure 8a shows the true trend of scenario 11 along with two SMA-filtered scenarios at bandwidth ratios of 0.04 and 0.10, respectively.This figure shows that, as the SMA filter bandwidth increases, the https://doi.org/10.5194/nhess-22-411-2022 Nat. Hazards Earth Syst.Sci., 22, 411-430, 2022  peak in measurements is identified at a later time than the true trend (x = 0.5) and the magnitude of the peak is reduced (more than 70 % reduction at a bandwidth ratio of 0.10).Furthermore, as the bandwidth ratio increases, the "instantaneous" nature of the peak is lost to a more transitional variation.This highlights a disadvantage of SMA when handling sudden changes in data trends.The calculated x value of the peak in scenario 11 is plotted for different bandwidth ratios and for all three filters in Fig. 8b.This figure shows the time at which the peak is identified lags as the bandwidth ratio increases for all filters; however, GWMA and S-G identify the peak with a much smaller lag, independent of the n/t ratio.
As an example, for a year of monitoring data at a frequency of 30 s and bandwidth ratio of 0.10, SMA, GWMA, and S-G predict the peak point approximately 17, 3.5, and 2.7 d after the real peak, respectively.This lag can be attributed to the utilization of an asymmetric window, which leads to a lagged response of the filter.As more points are included in the filtering procedure (increasing bandwidth ratio), this lag increases because the averaging process is sensitive to window type.The degree of sensitivity, however, depends on the filter.Figure 8c shows the variation in the peak magnitude with respect to the bandwidth ratio for all three filters.SMA and GWMA both underestimate the peak value, and the difference between the calculated peak and real peak increases as the bandwidth ratio increases.SMA calculations underestimate the peak more than twice as much as GWMA.On the contrary, S-G intensifies the peak up to a bandwidth ratio of 0.04, with the impact tending to diminish at larger bandwidth ratios; it predicts the true value at a bandwidth ratio value of almost 0.09.Scenario 12 was used for a detailed evaluation of the ability of these filters to conserve the underlying original trend.Figure 9 shows scenario 12 and the filtered results for all three filters and an n/t ratio of 0.15.This scenario and these specific parameters were selected for illustration purposes as they allow visual identification of differences for discussion.The SMA filter considerably underestimates the magnitude of the peak at a bandwidth ratio of 0.04, which should be the minimum bandwidth ratio according to Fig. 7.At a bandwidth ratio of 0.10, the filtered diagram is distorted in comparison to the true trend, and the initial peak is not identified.GWMA at a bandwidth ratio of 0.04 shows less underestimation of the peak magnitude, and a slight lag is visually observed at a bandwidth ratio of 0.10.This indicates the significantly better performance of GWMA over SMA.S-G results for both bandwidth ratios closely identify the time and magnitude of both peaks, indicating yet better performance.However, the peak is artificially intensified at a bandwidth ratio of 0.04, and a significant drop occurs well beyond the true trend immediately after the second peak for both bandwidth ratios (pulsating effect), which was also observed in scenario 11.Increasing the degree of the polynomial fitted as part of the S-G methodology was not completely effective at eliminating this effect.The pulsating effect was also observed when a symmetrical window was utilized and is attributed to the negative weights in the S-G kernel.

Results of direct scatter filtration
Figure 10 shows the RMSEd of all three filters for all the harmonic synthetic scenarios.This figure shows that, for these numerical analyses on synthetic scenarios, the error depends linearly on the bandwidth ratio for all of the filters and does not depend on the scenario or n/t ratio.SMA shows the greatest difference from the true trend, followed by GWMA (approximately 60 % less difference than SMA).S-G, on the other hand, almost lies on the horizontal axis for all the band- width ratios, which means the filtered results yield near-zero error.Figure 10 also shows how the error increases as the bandwidth ratio increases.This can be attributed to the utilization of an asymmetric window, which leads to a lagged response of the filter.As more points are included in the filtering procedure (increasing bandwidth ratio), this lag increases and, consequently, causes a larger error.The RMSEd values of filters for the instantaneous synthetic scenarios are shown in Fig. 11.In scenario 10, the same behaviour as noted for the harmonic scenarios can be seen for SMA and GWMA, whereas S-G is not as accurate.This is more noticeable in scenarios 11 and 12 in which S-G becomes less accurate than GWMA at larger bandwidth ratios.This result shows that S-G cannot handle the instantaneous scenarios https://doi.org/10.5194/nhess-22-411-2022 Nat. Hazards Earth Syst.Sci., 22, 411-430, 2022 as satisfactorily as the harmonic ones.The errors related to SMA and GWMA for the instantaneous synthetic scenarios show non-linear behaviour and are greater when compared to the harmonic scenarios.Figure 11 clearly shows all filters are challenged by the instantaneous variations when compared to gradual ones in direct filtration.

Results of indirect scatter filtration
Figure 12 shows the RMSEi results for the harmonic scenarios (when performing indirect filtration) on a semilogarithmic scale.We observed that the error considerably decreases as the bandwidth ratio increases to 0.02; however, to highlight the variation of error in the range of interest for the bandwidth ratio, only RMSEi values corresponding to bandwidth ratios greater than 0.04 are plotted in Figs. 12 and 13.In Fig. 12, the error for the GWMA is either equal to or slightly less than the error for the SMA, and S-G shows the least error for the harmonic scenarios.The RMSEi results for the instantaneous scenarios (Fig. 13) are similar to those for the harmonic scenarios for large n/t ratios (0.05, 0.10, and 0.15).For small n/t ratios, the GWMA is superior at bandwidth ratios above 0.06, and S-G has the worst performance.

Lag quantification
The non-symmetric inclusion of points causes the identification of a lag in the trend of filtered data.Figure 14 shows Scenario 10 with respect to the original trend, with scatter added (at an n/t value of 0.15), and the results after filtering with each of the three methods at a bandwidth ratio of 0.04.This figure clearly shows the lag between the results filtered by SMA and GWMA and the true trend.S-G results do not have as severe a lag as that resulting from the other filters; we attribute this to the negative weights in its kernel that anchor the filtered values and prevent a lagged response.A minor pulsating effect can be observed in the S-G filtered data, decreasing the calculated values at a much earlier time than the true trend.This suggests that S-G is robust with respect to identifying initial changes in monitoring trends but overcor-rects subsequent changes; SMA grossly lags with respect to the identification of any change, and GWMA has a reduced lag when compared to SMA. Figure 15a shows an example of the R 2 correlation for scenario 7, comparing the original trend and the results filtered by SMA at an n/t value of 0.01 and bandwidth ratio of 0.04.The shift ratio is the shift of filtered trends (on the horizontal axis -parameter x) relative to the range of x values.R 2 calculations are shown for the filtered data (shift ratio of 0) and as the filtered trends are shifted backwards in time (negative shift ratio values).In this analysis, the peak R 2 value (largest correlation between the shifted filtered results and original trend) indicates the shift required to minimize the lag in identifying the original trend changes, therefore providing a quantitative approach to calculating the lag in parameter x.
In the example in Fig. 15a, the lag corresponded to 0.018 (1.8 %) of the total points.
Peak R 2 values for all scenarios and n/t values are closely correlated with the bandwidth ratio.The lag, quantified by the shift ratio, is larger when the trend change is more pronounced; therefore, the correlation between the shift ratio and bandwidth ratio is different for different scenarios.Figure 15b shows the mean correlation between the shift ratio and bandwidth ratio, for all scenarios and n/t values, bounded by 1 standard deviation, for GWMA and SMA.Table 2 shows linear and quadratic regressions of this correlation and the strength of the correlation in terms of R 2 and RMSE. Figure 15b quantitatively shows that GWMA lags less than SMA with respect to identifying changes in measurement trends.Moreover, the uncertainty associated with lag for SMA is greater than for GWMA because of the larger standard deviation.Figure 15b quantifies how increasing the bandwidth ratio increases the lag with respect to identifying true measurement trends, and, although large bandwidth ratios decrease the scatter in data, the bandwidth ratio should carefully balance minimizing both scatter (J 2 ) and lag (shift ratio).S-G is not included in this analysis as the method resulted in no significant lag in identifying changes in measurement trends; however, it had the disadvantages previ-  ously noted including pulsating effects and overestimating peak values.

Results on the Ten-mile landslide
Unfiltered results reported by Geocubes 46 and 47 installed on the Ten-mile landslide were processed by all three filters.To illustrate to the reader through visual inspection the difference between the performance of SMA, GWMA, and S-G, only a 200 d window of displacement data from Geocube 46 and filtered points produced by direct filtration are shown in Fig. 16. Figure 16a also features an inset showing scaled scenario 4, which resembles the general trend of Geocube 46 data for the period from day 200 to 400. Figure 16 shows that increasing the bandwidth ratio reduces the scatter but increases the lag in the filtered results, consistent with observations on the synthetic datasets.For bandwidth ratios larger than 0.04, SMA becomes insensitive to some short-scale (20 https://doi.org/10.5194/nhess-22-411-2022 Nat. Hazards Earth Syst.Sci., 22, 411-430, 2022 Table 2. Regression correlations between shift ratio (SR) and bandwidth ratio (BR) with the strength of the correlation in terms of R 2 and RMSE. to 30 d) trends in the data (qualitative visual inspection).As an example, at a bandwidth ratio of 0.10, SMA suggests the displacement of Geocube 46 follows a bi-linear trend with an inflection point at day 240, while unfiltered points and other filters suggest other periods of acceleration and deceleration.Importantly, S-G is sensitive to even subtle variation and does not show significant lag.

Linear regression
Figure 17 shows the filtered velocity values obtained by directly filtering the calculated velocities and by indirectly filtering the displacement values before calculating the velocity from Geocube 46 data.The direct and indirect filtering approaches demonstrated similar performance in terms of scatter reduction for Geocube 46 data.As the bandwidth ratio increases, SMA tends to significantly attenuate the local maximum and minimum points in comparison to results at smaller bandwidth ratios, indicating a probable loss of information about the landslide behaviour and sensitivity of this filter to the bandwidth ratio, as also noted in Fig. 16 (curvature loss in SMA results).Indirect filtration by SMA seems to be limited near the boundary at time zero, resulting in a subdued replica of direct filtration.The length of this region is found to be governed by the bandwidth ratio, as the necessary number of points for filtering in this portion has not been provided to the filter.This is also observed in S-G results.This problem was not found in GWMA results as direct and indirect filtration both follow the same pattern.GWMA and S-G are both able to preserve the velocity variation even at the most intense filtration (bandwidth ratio of 0.10); however, variations between local maxima and minima are more extreme in S-G than GWMA results.This is attributed to peak overestimation (Figs. 8 and 9) or a pulsating effect superimposing on the peaks/troughs.Moreover, the S-G results still demonstrate relatively large fluctuations even at the largest bandwidth ratio.This means that the application of S-G might still trigger false alarms in an EWS if the landslide is moving at a faster rate or experiencing different episodes of acceleration and deceleration.To avoid this, a larger bandwidth ratio should be used, but this can be problematic due to the higher computational effort required and issues that might follow, such as the pulsating effect.
Results for Geocube 47 confirm the same observations made for Geocube 46 but also allow for an evaluation of the significance of outliers on the filtered results.Figure 18a displays the outliers detected in the displacement diagram of Geocube 47 data along with the threshold established by the Hampel algorithm using an asymmetric window, a bandwidth of 0.4 %, and a threshold factor of 3. Figure 18b-d show a magnified portion of the displacement measurements for Geocube 47 filtered by each of the three filters at three different bandwidth ratios before the elimination of outliers.This highlights the necessity of outlier elimination before the application of any scatter filter.These plots show that detecting and removing outliers significantly impacts the performance of S-G as the presence of the outlier generates a peak that follows the outlier measurement and is followed by a sudden decrease that drops well beyond the data trend.SMA tends to widen the time range affected by the outlier more than GWMA, but, for the most part, the SMA-filtered results are almost parallel to the underlying trend.All filters appear to be significantly impacted by the outlier value, suggesting a pre-processing filter is required to remove outliers regardless of the use of SMA, GWMA, or S-G to reduce scatter.The outliers were successfully identified and removed after the application of the Hampel algorithm, and the abovementioned effects were no longer observed in the filtered results.

Lag minimization in filtered Geocube results
The lag between unfiltered and filtered data for Geocube 46 (Fig. 16) is consistent with the synthetic database results.The lag quantification results (Fig. 15b) were used to provide a correction value for the filtered Geocube results.The shift ratios used for this purpose with respect to each filter and bandwidth ratio are tabulated in Table 3.To determine whether the results of lag correction using the mean correlations derived from the synthetic scenarios (Table 2) were acceptable, the filtered diagrams were shifted (using the mean line for GWMA and values between the mean and lower boundary for SMA), and different portions of the displacement diagrams for Geocubes 46 and 47 were examined.Some examples are shown in Fig. 19.The mean and standard deviation of the scatter around the trend (error distribution) were calculated by assuming a linear trend within the short periods of analysis (considered an approximation of the true displacement trend for the short time interval).These were also calculated for the filtered and shifted diagrams.The closer the mean and standard deviation of the filtered and shifted data are to those obtained from the linear trend, the better the performance is of the lag correction based on the results from the synthetic scenarios.As an example, for the period from day 250 to 260, the GWMA resulted in a standard deviation of 0.001 to 0.0015 for bandwidth ratios from 0.04 to 0.10, respectively; corresponding values for SMA were 0.0018 to 0.0021.This illustrates that shifted GWMA results are closer to the true (scatter-free) displacements because the standard deviations of scatter inferred by this filter are closer to the true scatter, although both have good agreement with the true scatter.The means of inferred scatter by both filters are also close enough to the mean of the true scatter (almost zero).
The results show the statistical indices of scatter inferred from the filtered shifted displacement measurements closely agree with that considered to be true scatter, and therefore the filtered displacement measurements are corrected for lag.This suggests the correlations stated in Fig. 15b and Table 2 based on the synthetic scenarios are applicable to minimize the lag for the Geocube system at the Ten-mile landslide.

Discussion
Previous studies dedicated to landslide monitoring consistently adopt SMA for scatter minimization in displacement data.However, the adequacy of this filter and the effect of bandwidth selection were not well understood.Analyses conducted on synthetic databases in this study using a roughness factor (J 2 ) demonstrate that at least 4 % of the total observations should be fed into the filter to ensure fluctuations are sufficiently reduced.
The results of this study show that SMA tends to considerably distort the underlying trend at a bandwidth ratio of 0.10 (Figs. 8 and 9), and its lagged response with respect to realtime monitoring is almost 3 times that of GWMA results.As a result, a bandwidth ratio between 0.04 and 0.07 is suggested.However, we caution that the bandwidth should be selected with complete awareness that SMA is highly sensitive to bandwidth, and sensitivity analyses on bandwidth are recommended when defining an EWS.Corresponding observations were made during the analysis of displacement data from Geocubes installed on the Ten-mile landslide.
Error calculations show that GWMA and S-G outperform SMA in both direct and indirect filtration and are more suchttps://doi.org/10.5194/nhess-22-411-2022 Nat. Hazards Earth Syst.Sci., 22, 411-430, 2022 cessful in preserving the true displacement trend.The nearzero lagged response of S-G makes it a notable candidate for developing an EWS.Nonetheless, its intrinsic shortcoming in handling peaks, leading to a pulsating effect, will pose challenges for its utilization.The bandwidth range used for SMA is also suggested to be applied with the S-G filter.
GWMA results suggest a proper trade-off can be achieved between minimizing the lag time and scatter and avoiding the pulsating effect.Compared to SMA and S-G, GWMA is less sensitive to changes in the bandwidth.Analyses focused on the Geocube data also confirm that GWMA is capable of constraining the fluctuations in the velocity diagram while not attenuating variations in the displacement rate diagram.
Moreover, the lag quantification chart proposed could reliably capture the required shift with a greater degree of confidence in comparison to SMA even at the largest bandwidth ratio studied here (0.10).The bandwidth for GWMA can therefore range from 0.04 to 0.10.Moreover, we observed consistency between direct and indirect filtration results using GWMA but greater differences when using SMA or S-G results.This was especially the case in the early parts of the datasets and at some locations where outlier elimination was likely ineffective.
Filter and bandwidth selections should not be arbitrarily or purely empirical, as differences in outcomes can be substantial.An automated surveillance system for landslides de- mands stability in filter performance for a variety of circumstances, considering the ground can experience irregular sequences of acceleration and deceleration.The results here suggest practice moves away from the adoption of SMA due to the limitations discussed.S-G demonstrates some inconsistent or erratic performance for certain displacement trends, which is detrimental, although overall the error is smaller than for SMA.On the balance of its strengths and limitations as evaluated in this study, GWMA appears to be the more robust approach.

Conclusions
This study evaluated the suitability of SMA, GWMA, and S-G filters for scatter reduction of datasets targeted for use in an EWS.A total of 12 different scenarios with harmonic and instantaneous changes were synthetically generated, and random variations with Gaussian distribution were then added to produce unfiltered results.The three filters considered were then each applied with different bandwidths, and the error was computed.These filters were also successfully applied to the records from two Geocubes installed on the Ten-mile landslide.The results led to the following conclusions: -When used for direct filtration of harmonic scenarios, the error resulting from the GWMA approach is approximately one-third that of the SMA approach.The S-G approach results in near-zero error regardless of the values of the bandwidth ratio and n/t.When used for direct filtration of instantaneous scenarios, the superiority of S-G is no longer unconditional and depends on the bandwidth ratio; this reflects the fact that S-G cannot appropriately handle peaks in the velocity diagram.
-When used for indirect filtration of harmonic scenarios, S-G again outperforms the other methods.The error associated with GWMA is marginally less than for SMA.These observations are not valid when the filters are applied to instantaneous scenarios as GWMA results in less error than S-G at bandwidth ratios above 0.03.-Detailed investigations with scenarios 11 and 12 demonstrate that SMA distorts the underlying trend by displacing and sometimes neglecting peak(s), while GWMA and S-G tend to preserve them somewhat similarly.
-Due to the presence of negative weights in the S-G kernel, some artificial smaller troughs and peaks are created after major peaks.This phenomenon, referred to herein as a pulsating effect, results in an unfavourable performance of S-G on the velocity and displacement diagrams, especially in the presence of outliers.
-Investigations on the roughness factor reveal the bandwidth ratio should be at least 0.04.Taking this into account, GWMA seems to be the most reasonable option as the related uncertainties are much smaller than for S-G and the error is acceptable and less than for SMA.
-A consequence of using asymmetric windows in the filtering process is a lag in the SMA and GWMA results that increases with increasing bandwidth ratio.Lag quantification suggests a correlation between the needed shift and bandwidth ratio that can be used to eliminate the lag.SMA requires approximately 3 times the shift of GWMA on average.
-Application of these filters to displacement data reported by Geocubes shows SMA and S-G are unable to properly handle data points at the beginning of the dataset (i.e., near the boundary) in indirect filtration of the velocity diagram.Moreover, SMA and S-G are inclined to, respectively, underestimate and overestimate  Review statement.This paper was edited by Filippo Catani and reviewed by Ugur Ozturk and one anonymous referee.

Figure 1 .
Figure 1.Configuration of all synthetically generated scenarios.

Figure 2 .
Figure 2. The procedure of generating a scenario with scatter: (a) generated scenario trend, (b) randomly generated scatter, and two scenarios with scatter based on n/t values of (c) 0.05 and (d) 0.10.

Figure 3 .
Figure 3.The weighting kernel of the Savitzky-Golay filter for seven points.

Figure 4 .
Figure 4. Concept of symmetric and non-symmetric window types in the filtration process.

Figure 5 .
Figure 5. (a) Location of the Ten-mile landslide (source: Earthstar Geographics LLC) and (b) front view of the Ten-mile landslide and distribution of Geocubes on its surface(Rodriguez et al., 2018;Macciotta et al., 2017b).

Figure 7 .
Figure 7. Variation in roughness factor for scenario 6 with respect to the applied filter on a semi-log scale.

Figure 8 .
Figure 8.(a) An example of peak displacement by applying SMA, as well as variation in (b) peak position and (c) peak value with respect to the filter and bandwidth ratio used (original peak at 0.5).

Figure 12 .
Figure 12.RMSEi for the harmonic scenarios on a semi-logarithmic scale.

Figure 14 .
Figure 14.Scenario 10 with and without scatter, as well as with scattered results filtered by SMA, GWMA, and S-G for an n/t value of 0.15 and a bandwidth ratio of 0.04.

Figure 15 .
Figure 15.(a) R 2 values for scenario 7 with filtered and shifted results at an n/t value of 0.01 and bandwidth ratio of 0.04 and (b) shift ratio at peak R 2 for all scenarios and n/t ratios, with the mean (solid line) bounded by 1 standard deviation (dashed lines)

Figure 19 .
Figure 19.Mean and standard deviation of scatter inferred by SMA and GWMA in comparison with true scatter in the displacement of Geocube 46.

Table 1 .
Number of points used to generate scenarios and examples of their corresponding time spans represented by the range of x from 0 to 1 if the measurement frequency is known (1 h and 1 m readings for illustrative purposes).

Table 3 .
Shift ratios used for lag minimization of Geocube 46 displacements.
ing the Geocube units.This research was conducted through the (Canadian) Railway Ground Hazard Research Program, which is funded by the Natural Sciences and Engineering Research Council of Canada (NSERC ALLRP 549684-19), Canadian Pacific Railway, CN, and Transport Canada.Financial support.This research has been supported by the Natural Sciences and Engineering Research Council of Canada (grant no.NSERC ALLRP 549684-19).