Self-similarity is a special feature of network data. When operating network data, the tranditional time series models, for example a autoregressive moving average model (ARMA(p,q)), are not appropriate. Since it only provides a finite parameters model, but the network data usually has long-range dependent structure. Thus, a self-similar process for describing the network data structrue is applied. In the following paragraphs, we mention the definitions of the self-similar processes and some properties of them. At the same time, we describe some important methods for checking and applying the self-similarity of network data, such as estimating the Hurst's parameter and proposing a proper modeling method.
Definition
Suppose
be a weakly-stationary (2nd-order stationary) process
with mean
, variance
, and autocorrelation function
.
Assume that the autocorrelation function
has the form
as
, where
and
is a slowly varying function at infinity, that is
for all
.
For example,
and
are slowly varying functions.
Let
,
where
, denote an aggregated point series over non-overlapping blocks of size
, for each
is a positive interger.
Exactly self-similar process
is called an exactly self-similar process if there exists a self-similar parameter
such that
has the same distribution as
. An example of exactly self-similar process with
is Fractional Gaussian Noise (FGN) with
.
Def:Fractional Gaussion Noise (FGN)
is called the Fractional Gaussion Noise, where
is a Fraction Brownian Motion.
exactly second order self-similar process
is called an exactly second order self-similar process if there exists a self-similar parameter
such that
has the same variance and autocorrelation as
.
asymptotic second order self-similar process
is called an asymptotic second order self-similar process with self-similar parameter
if
as
,
Properties of Self-Similar Processes
Long-Range-Dependence(LRD)
Suppose
be a weakly-stationary (2nd-order stationary) process with mean
and variance
. The Autocorrelation Function (ACF) of lag
is given by ![{\displaystyle \gamma (t)={\mathrm {cov} (X(h),X(h+t)) \over \sigma ^{2}}={E[(X(h)-\mu )(X(h+t)-\mu )] \over \sigma ^{2}}}](/media/api/rest_v1/media/math/render/svg/20e83e74582dbb3dfaeaac64f833ff4a333460ce)
Definition:
A weakly stationary process is said to be "Long-Range-Dependence" if 
A process which satisfies
as
is said to have long-range dependence. The spectral density function of long-range dependence follows a power law near the origin. Equivalently to
,
has long-range dependence if the spectral density function of autocorrelation function,
, has the form of
as
where
,
is slowly varying at 0.
also see
Slowly decaying variances
When an autocorrelation function of a self-similar process satisfies
as
, that means it also satisfies
as
, where
is a finite positive constant independent of m, and 0<β<1.
Testing for self-similarity
R/S analysis
Assume that the underlying process
is Fraction Gaussion Noise. Consider the series
, and let
.
The sample variance of
is 
Definition:R/S statistic
![{\displaystyle {\frac {R}{S}}(n)={\frac {1}{S(n)}}[\max _{0\leq t\leq n}(Y_{t}-{\frac {t}{n}}Y_{n})-\min _{0\leq t\leq n}(Y_{t}-{\frac {t}{n}}Y_{n})]}](/media/api/rest_v1/media/math/render/svg/e1f742593cafcddd727721274a337af8f241432c)
If
is FGN, then 
Consider fitting a regression model :
, where

In particular for a time series of length
divide the time series data into
groups each of size
, compute
for each group.
Thus for each n we have
pairs of data (
).There are
points for each
, so we can fit a regression model to estimate
more accurately. If the solpe of the regression line is between 0.5~1, it is a self-similar process.
File:Self similar pox plot.jpg
Variance-time plot
Variance of the sample mean is given by
.
For estimating H, calculate sample means
for
sub-series of length
.
Overall mean can be given by
, sample variance
.
The variance-time plots are obtained by plotting
against
and we can fit a simple least square line through the resulting points in the plane ignoring the small values of k.
For large values of
, the points in the plot are expected to be scattered around a straight line with a negative slope
.For short-range dependence or independence among the observations, the slope of the straight line is equal to -1.
Self-similarity can be inferred from the values of the estimated slope which is asymptotically between –1 and 0, and an estimate for the degree of self-similarity is given by 
File:Self similar variancetime plot.jpg
Periodogram-based analysis
Whittle's approximate maximum likelihood estimator (MLE) is applied to solve the Hurst's parameter via the spectral density of
. It is not only a tool for visualizing the Hurst's parameter, but also a method to do some statistical inference about the parameters via the asymptotic properties of the MLE. In particular,
follows a Gaussian process. Let the spectral density of
,
, where
, and
construct a short-range time series autoregression (AR) model, that is
,
with
.
Thus, the Whittle's estimator
of
minimizes
the function
, where I(w) denotes the periodogram of X as
and
. These integrations can be assessed by Riemann sum.
Then
asymptotically follows a normal distribution if
can be expressed as a form of a infinite moving average model.
To estimate
, first, one has to calculate this periodogram. Since
is an estimator of the spectral density, a series with long-range dependence should have a periodogram, which is proportional to
close to the origin. The periodogram plot is obtained by ploting
against
.
Then fitting a regression model of the
on the
should give a slop of
. The slope of the fitted straight line is also the estimation of
. Thus, the estimation
is obtained.
File:Self similar periodogram plot.jpg
Note:
There are two common problems when we apply the periodogram method. First, if the data does not follow a Gaussian distribution, transformation of the data can solve this kind of problems. Second, the sample spectrum which deviates from the assumed spectral density is another one. An aggregation method is suggested to solve this problem. If
is a Gaussian process and the spectral density function of
satisfies
as
, the function,
, converges in distribution to FGN as
.
FARIMA modeling
The time series methodology is helpful in network data analysis. Especially when data is non-stationary, a fractional autoregressive integrated moving average (FARIMA) method is applied. Like the traditional ARIMA(p, d, q) model, the only difference is the value of the parameter
must be between -0.5 and 0.5. When
, the data has long-range dependence, and the Hurst's parameter is equal to
. The advantage of this method is useful to capture the short-range dependence via ARMA (p, q) and capture the LRD property at the same time.
Modeling procedures of FARIMA:
- Step1: Consider
. Let
such that
.
- Step2: Use R/S (or other estimator) to estimate Hurst's parameter
, and then the estimation of
is obtained by 
- Step3: Take
, where
.
'B' is called 'Back shift operator', that is
- Step4: Fit an ARMA(p,q) model using the transformation data
.
Thus, a FARIMA(p,d,q) model is obtained as the form

References
- P. Whittle, "Estimation and information in stationary time series", Art. Mat. 2, 423-434, 1953.
- K. PARK,W. WILLINGER, Self-Similar Network Traffic and Performance Evaluation,WILEY,2000.
- W. E. Leland,W. Willinger,M. S. Taqqu,D. V. Wilson, "On the self-similar nature of Ethernet traffic",ACM SIGCOMM Computer Communication Review 25,202-213,1995.
- W. Willinger,M. S. Taqqu,W. E. Leland,D. V. Wilson, "Self-Similarity in High-Speed Packet Traffic: Analysis and Modeling of Ethernet Traffic Measurements",Statistical Science 10,67-85,1995.