Novel anomaly detection approach for telecommunication network proactive performance monitoring

Yanhua YU , Jun WANG , Xiaosu ZHAN , Junde SONG

Front. Electr. Electron. Eng. ›› 2009, Vol. 4 ›› Issue (3) : 307 -312.

PDF (125KB)
Front. Electr. Electron. Eng. ›› 2009, Vol. 4 ›› Issue (3) : 307 -312. DOI: 10.1007/s11460-009-0051-9
RESEARCH ARTICLE
RESEARCH ARTICLE

Novel anomaly detection approach for telecommunication network proactive performance monitoring

Author information +
History +
PDF (125KB)

Abstract

The mode of telecommunications network management is changing from “network oriented” to “subscriber oriented”. Aimed at enhancing subscribers’ feeling, proactive performance monitoring (PPM) can enable a fast fault correction by detecting anomalies designating performance degradation. In this paper, a novel anomaly detection approach is the proposed time series prediction and the associated confidence interval based on multiplicative autoregressive integrated moving average (ARIMA). Furthermore, under the assumption that the training residual is a white noise process following a normal distribution, the associated confidence interval of prediction can be figured out under any given confidence degree 1-α by constructing random variables satisfying t distribution. Experimental results verify the method’s effectiveness.

Keywords

proactive performance monitoring (PPM) / anomaly detection / time series prediction / autoregressive integrated moving average (ARIMA) / white noise / confidence interval

Cite this article

Download citation ▾
Yanhua YU, Jun WANG, Xiaosu ZHAN, Junde SONG. Novel anomaly detection approach for telecommunication network proactive performance monitoring. Front. Electr. Electron. Eng., 2009, 4(3): 307-312 DOI:10.1007/s11460-009-0051-9

登录浏览全文

4963

注册一个新账户 忘记密码

Introduction

The mode of telecommunication network management is changing from “network oriented” to “subscriber oriented”. Accordingly, the mode of performance management of network, an important aspect of network management, is transforming from passiveness to initiativeness. The proactive performance monitoring (PPM) is simply one of the initiative management ways.

PPM [1,2] is concerned with performance analysis and fault detection, which are capable of detecting “soft” network and service faults automatically and adaptively in the midst of networks’ performance fluctuations and evolutions. By detecting anomalies designating performance degradation, which are the symptoms of network faults and preludes to services failure, PPM can enable a fast fault containment and correction, through which serious network failures can be avoided and duration time can be shortened. “Anomaly” is defined as “statistically unusual”. It can be detected by using two terms: baseline and thresholds (upper and lower thresholds). Thresholds fluctuate around predictive baseline. If the actual value of key performance indicator (KPI) exceeds thresholds, performance warning would be triggered to show the deteriorating performance quality.

For anomaly detection, the prediction of the baseline and thresholds of KPI is a crucial point. If the threshold is set over strictly, a fake warning may be launched. However, if the threshold is too high or too low, the mechanism will make little sense. There were some researches on the derivations of baseline and thresholds [2,3]. In Ref. [2], the threshold that is called tolerance limits or envelope is based on standard deviation of training set, while in Ref. [3], the baseline is based on a simplistic algorithm, and they are all too coarse. In this paper, time series prediction is deeply studied to settle baseline, and confidence interval of prediction error is used for thresholds. Under the assumption that a white noise process follows normal distribution, the associated confidence interval of prediction value can be worked out under any dedicated confidence degree 1-α by constructing random variables following t distribution.

This paper is organized as follows. In Sect. 2, the principle and model identification method of autoregressive integrated moving average (ARIMA) is introduced. In Sect. 3, the proposed time series prediction approach with the associated confidence interval calculation using multiplicative ARIMA is presented in particular. In Section 4, an experiment with the proposed approach is carried out. The conclusion is drawn in Sect. 5.

Overview of ARIMA

Let time series x1,x2,…,xi,…,xN denote the observations made at equidistant time interval τ0+h, τ0+2h, …, τ0+ih, …, τ0+Nh, where xi represents observation at time τ0+ih if we consider τ0 as the origin and h as the unit of time. Basically, the time series prediction can be considered as a modeling issue; a model is built between input and output. Then, the model is used to predict the future values based on the previous values.

Among the algorithms deployed in time series prediction, both autoregressive moving average (ARMA) [4] and ARIMA [5] are most commonly used. ARMA can only be used for linear and stationary time series, while ARIMA can be used for nonstationary ones exhibiting homogeneity. When the time series show heterogeneous nonstationarity, other algorithms such as artificial neural network (ANN) [6] and support vector machine (SVM) will be adopted [7-9].

In this paper, we use multiplicative ARIMA, ARIMA with seasonal part, to model KPI time series with homogeneous nonstationarity.

Since ARIMA is a model evolving from ARMA, which can be regarded as an integration of autoregressive (AR) model and moving average (MA) model, AR, MA, and ARMA models are introduced first.

AR process

Autoregressive model has the form as follows:
xt=i=1pϕixt-i+ϵt,
where ϕp0, and ϵt is a white noise process satisfying WN(0,σ2).

Accordingly, the process defined by Eq. (1) is called an autoregressive process of order p, or more succinctly, an AR(p) process.

MA process

Moving average model has the form as follows:
Yt=ϵt-i=1qθiϵt-i,
where θq0, and ϵt is a white noise process satisfying WN(0,σ2).

The process defined by Eq. (2) is called moving average process of order q, which is abbreviated to MA(q).

Mixed ARMA process

Sometimes, a series can be modeled as autoregressive process, while some others can be modeled as moving average process. However, there are time series that cannot be modeled as pure AR or MA due to the too many parameters required. In these cases, mixed ARMA model should be used where both autoregressive and moving average terms are in the model, which can be represented as
xt=i=1pϕixt-i-j=0qθjϵt-j,
or
ϕ(B)xt=θ(B)ϵt.
In Eq. (4), B stands for backward shilt operator, which is defined by Bxt=xt-1,

A process defined by Eq. (3) or (4) is called mixed autoregressive moving average process of order (p, q), which is abbreviated to ARMA(p,q).

ARIMA process

AR, MA, and ARMA models described above are all appropriate for modeling stationary process. However, in reality, many time series are nonstationary. For those nonstationary time series, nevertheless exhibiting homogeneity, we can stationalize them by differencing operation. Let =1-B denote the differencing operator so that ARIMA model can be written as
ϕ(B)dxt=θ(B)ϵt,
where d is an integer representing number of differencing operation, ϕ(B) is a stationary autoregressive operator, and θ(B) is a moving average operator. Thus, we can get the conclusion that the model represented by Eq. (5) corresponds to the assumption that the dth difference of the series can be represented by a stationary ARMA process. The process satisfying Eq. (5) is called an ARIMA process.

Similarly, if the nonstationary time series exhibit periodical homogeneity with period s, they can be stationalized by seasonal differencing. The seasonal operator is of the form s=1-Bs. If a time series is only of seasonal nonstationary, the model can be represented as
ΦP(Bs)sDxt=ΘQ(Bs)ωt,
where
ΦP(Bs)=1-Φ1Bs-Φ2B2s-Φ3B3s--ΦPBPs,
and
ΘQ(Bs)=1-Θ1Bs-Θ2B2s-Θ3B3s--ΘQBQs.

If time series have both trendy and seasonal part, integrating Eqs. (5) and (6), we can get the following model:
ϕ(B)ΦP(Bs)dsDYt=θ(B)ΘQ(Bs)ϵt.

The resulting multiplicative ARIMA process will be of order (p,d,q)×(P,D,Q)s.

Time series prediction based on ARIMA with associated confidence interval

For some KPIs such as traffic volume or call attempts, the baseline can be obtained according to time series prediction using ARIMA. The threshold then can be achieved based on the confidence interval under a dedicated confidence degree 1-α by using the hypothesis that training residual satisfying white noise follows normal distribution.

Take the hourly traffic of mobile service switching center (MSC) for example. The traffics in different hours in a day are usually different. Because PPM takes hours as a unit, we predict hourly traffic using traffics at the same time in the past days. For example, if we want to predict the traffic at 9:00 tomorrow, then the history values are traffic data at 9:00 of today and yesterday and so on. Here, we use traffic values at 9:00 from 2007-7-4 to 2007-8-17, which is totally 45 consecutive items, as training data.

There are three steps to make the time series prediction using ARIMA. First, the ARIMA model should be identified. Second, the parameters should be estimated for the identified class of ARIMA models. Third, the prediction for future values should be made by using history values. In this paper, the associated confidence interval of prediction value should also be computed under given confidence degree.

The three steps above will be discussed in detail as follows.

Model identification

To get the ARIMA model fit for a time series, the order of the model (p,d,q)×(P,D,Q)s should be decided at the first stage. Thus, an appropriate subclass of models can be identified from the general seasonal multiplicative ARIMA family denoted by Eq. (7). This stage is usually called model identification. By calculating the autocorrelation function, we can justify if the series exhibit trend or if the series shows seasonal pattern with period s. Then, we can make the series stationary by using a differencing operator or seasonal differencing operator. For stationary time series, there are several model identification criteria including AIC, BIC, F-test, etc.

In an actual application, especially when the series length of available data is not very large, the orders P, Q, p, and q would rarely need to be greater than 1. Therefore, at this stage of model identification, we should select the best model from all the combinations of (p,d,q)×(P,D,Q)s where all the orders are between zero and one according to AIC.

Parameter estimation

After the appropriate model (p,d,q)×(P,D,Q) has been identified, the parameters would be estimated. There are several estimation algorithms, such as Yule-Walker equation, minimum mean squared error (MMSE), and maximum likelihood estimation (MLE). MLE is used in this paper.

Prediction with associated confidence interval

After the model is built up, it can be used to forecast future values based on history data. There are two types of forecasting, one-step ahead forecasting, and multisteps ahead forecasting. To guarantee the precision of forecasting, one-step ahead forecasting is adopted.

Actual value at time τ0+ih is denoted by xi, and prediction value for the same time is denoted by x^i, the equation
xi=x^i+ϵi
holds, where ϵi is the prediction error and is white noise process following normal distribution, ϵiN(0,σ2). Therefore, the actual value at the monitoring time point x^n+1 and prediction value x^n+1 satisfy xn+1N(x^n+1,σ2), which is equivalent to xn-1-x^n+1σN(0,1). Using a normal distribution table, we can compute the confidence interval of actual value 1 under a given confidence degree 1-α, and the following equation is established:
P{-zπ/2xn+1-x^n+1σzπ/2}=1-α.
If the number of training items is more than 50, the training sample standard deviation
S=1k-1i=1k(ϵi-ϵ ¯i)2
can be regarded as equal to the population standard deviation σ. Let a confidence degree be 1-α=0.95, so that we can get the confidence interval of xn+1 as
(x^n+1-2σ,x^n+1+2σ)=(x^n+1-2S,x^n+1+2S).
If the confidence degree is 1-α=0.97, the confidence interval will be (x^n+1-3σ,x^n+1+3σ), which can be substituted with (x^n+1-3S,x^n+1+3S).

However, if the number of training items is less than 50, the value of both σ and S may vary significantly so that they can not be regarded as being equivalent. In this case, we propose the following algorithm to get the confidence interval.

Since training residual {ϵi} satisfies normal distribution, according to the linear characteristics of operation of random variables obeying normal distribution, its mean value denoted by ϵ ¯ also satisfies normal distribution:
ϵ ¯N(0,σ2n).

The random variable ϵn+1-ϵ ¯ will also satisfies the following normal distribution:
ϵn+1-ϵ ¯N(0,[1+1n2]σ2),
and Eq. (8) can be transformed to
U=ϵn+1-ϵ ¯n2+1n2σN(0,1).
Considering the fact that the standard deviation S of training residual satisfying χ2 distribution, a random variable V can be derived as follows:
V=(n-1)S2σ2~χ2(n-1).
Integrating Eqs. (9) and (10), another random variable named Z satisfying t distribution can be obtained by
Z=UVn-1=n2n2+1(ϵn+1-ϵ ¯)S~t(n-1).
In Eq. (11), only ϵn+1 is unknown, it can be transformed to
P{ϵ ¯-n2+1n2tπ/2(n-1)Sϵn+1ϵ ¯+n2+1n2tπ/2(n-1)S} =1-α.
Therefore, the confidence interval of xn+1 under confidence degree 1-α will be
[ϵ ¯-n2+1n2tπ/2(n-1)S+x^n+1,ϵ ¯+n2+1n2tπ/2(n-1)S +x^n+1].
Thus, the former value is taken as the lower threshold, and the latter is the upper threshold with the confidence degree 1-α.

Anomaly detection for traffic series using ARIMA

Now, we have an hourly traffic load series at 9:00 from 2007-7-4 to 2007-8-17 with length of 45, as shown in Table 1. From the first item, we always take 37 consecutive items as training data to predict the next item with the associated confidence interval.

We stationalize the original traffic time series by differencing operation. The autocorrelation function of the original traffic series is depicted in Fig. 1. This figure reveals that there is no trend part but a marked seasonal pattern with periodicity of 7, so a seasonal difference should be carried out and denotes the resulting series by {yi}, where yi=7xi=(1-B7)xi. Then, the stationary of {yi} is explored by calculating autocorrelation function. The result is shown in Fig. 2, where the seasonally differenced series { yi } is stationary.

We model the training data using ARIMA (p,0,q)×(P,1,Q7) model and identify the model using AIC. Then, all the combinations of (p,0,q)×(P,1,Q) where each order is 0 or 1 were tested and ensured that they are not zeros at the same time and select the combination that makes AIC the least. The model (1,0,1)×(1,1,0) is selected, which means that the original time series {xi} (1,0,1)×(1,1,0)7 is the best model.

We make a one-step ahead prediction as the baseline and calculate the associated confidence interval as the threshold. The one-step ahead prediction value is 2249.6, which was taken as the hourly traffic for 9:00, 2007-8-10. The sample standard deviation is S=42.03. With the confidence degree 1-α=0.95, we have
tα/2(n-1)=2.05,
and
n2+1n2tα/2(n-1)S=86.212.
With the confidence degree 1-α=0.95, the confidence interval is
[2249.6+2.77-86.212,2249.6+2.77+86.212]=[2166.158,2338.582].

Therefore, we reach conclusion that the actual traffic value for 2007-8-10 9:00 is 2238.39, which is among the confidence interval [2166.158, 2338.582]. This means that during this period, the network is in normal state.

Using the method above, the prediction with associated confidence interval can be made from 2007-8-10 until 2007-8-17. The result is shown in Table 2 and plotted in Fig. 3.

We come to the following conclusions based on Table 2 and Fig. 3:

1) The actual traffics at 9:00 from 2007-8-10 till 2007-8-16 are all within the tolerance limit and all the values of absolute percent errors (APEs) are smaller than 3%. Figure 3 shows that in normal state, the proposed prediction approach is of high precision.

2) However, at 9:00 on 2007-8-17, the actual traffic, 2101.47, falls outside tolerance limit [2156.72, 2364.56]. In this status, a performance alarm should be launched. This abnormal status has been confirmed by the fact that during this period, there was something wrong with IC2 module in BSU unit, which caused some call setup failure. That is why the traffic was so low then. 2007-8-17 is Friday when traffic is usually the heaviest in the week. However, Fig. 2 shows us an opposite case that the traffic on 2007-8-17 is the lightest in the week, which is obviously abnormal.

Conclusion

The prediction of baseline and threshold for KPIs is a crucial issue for proactive performance monitoring of telecommunication network. In this paper, the approach of time series prediction with the associated confidence interval using ARIMA was adopted to solve this problem. Furthermore, under the assumption that a white noise process follows normal distribution, the associated confidence interval of prediction can be computed under any given confidence degree 1-α by constructing random variable satisfying t distribution. Experimental results show that using the proposed modeling method, a quite precise prediction and fluctuation range can be achieved.

References

[1]

Hellerstein J L, Zhang F, Shahabuddin P. An approach to predictive detection for service management. In: Proceedings of the 6th IFIP/IEEE International Symposium on Integrated Network Management, 1999, Sloman M, Mazumdar S, Lupu E, Eds. New York: IEEE Publishing, 1999, 309-322

[2]

Feather F, Siewiorek D, Maxion R. Fault detection in an Ethernet network using anomaly signature matching. ACM SIGCOMM Computer Communication Review, 1993, 23(4): 279-288

[3]

Ho L L, Cavuto D J, Papavassiliou S, Zawadki A G. Adaptive and automated detection of service anomalies in transaction-oriented WANs: network analysis, algorithms, implementation, and deployment. IEEE Journal on Selected Areas in Communications, 2000, 18 (5): 744-757

[4]

Li J, Liu X X, Han Z J. Research on the ARMA-based traffic prediction algorithm for wireless sensor network. Journal of Electronics and Information Technology, 2007, 29(5): 1224-1227 (in Chinese)

[5]

Cadzow J A. ARMA time series modeling: an effective method. IEEE Transactions on Aerospace and Electronic Systems, 1983, AES-19(1): 49-58

[6]

Versace M, Bhatt R, Hinds O, Shiffer M. Predicting the exchange traded fund DIA with a combination of genetic algorithms and neural networks. Expert Systems with Applications, 2004, 27(3): 417-425

[7]

Mukherjee S, Osuna E, Girosi F. Nonlinear prediction of chaotic time series using support vector machines. In: Proceedings of the 1997 IEEE Workshop on Neural Networks for Signal Processing, 1997, 511-520

[8]

Shi Z W, Han M. Support vector echo-state machine for chaotic time-series prediction. IEEE Transactions on Neural Networks, 2007, 18(2): 359-372

[9]

Cao L J, Tay F E H. Support vector machine with adaptive parameters in financial time series forecasting. IEEE Transactions on Neural Networks, 2003, 14(6): 1506-1518

RIGHTS & PERMISSIONS

Higher Education Press and Springer-Verlag Berlin Heidelberg

AI Summary AI Mindmap
PDF (125KB)

921

Accesses

0

Citation

Detail

Sections
Recommended

AI思维导图

/