Beruflich Dokumente
Kultur Dokumente
for many irregular phenomena observed in financial markets. Still, many of the dictates of behavioral
finance cannot be easily accommodated by the current econometric models which makes it hard to
empirically validate not only the statistical but also the economic significance of behavioral hypotheses.
In this paper, the authors discuss new trends in econometric research, namely cointegration and
equilibrium-correction models, and show how these can effectively incorporate many ideas of the
behavioral finance. This, in particular, applies to the interaction of the price and the fundamental value.
The authors provide an up-to-date review of econometric and computational intelligent models that have
been proposed for cointegration analysis and equilibrium-correction modeling, and show exactly which
behavioral elements these models are able to capture. The authors in particular, favor the use of
intelligent learning algorithms both in the detection of complicated cointegration relations and the
representation of the equilibrium restoration dynamics.
Most financial and economic theories suggest long-run equilibrium relationships between
financial or economic variables, which take the form of a linear or non-linear functional
relationship between variables of interest. Much of the theoretical work, however, is
exclusively focused on long-run relationships in ideal markets, by simplistically assuming
that equilibrium is almost instantly restored and various market irregularities are
insignificant so far as the final formation of prices is concerned. There is, thus, little
guidance on what to expect in transient states of adjustment to equilibrium in an
imperfect market setting with irrational traders and frictions. Behavioral finance has
made an important breakthrough by stressing out the importance and persistence of non-
equilibrium states caused by market frictions (or else limits to arbitrage) and irrational
trading. As we show later in this paper, it has also taken one step further by giving
indications on what patterns of irrational trading and equilibrium restoration are likely
to hold in practice. However, many of these ideas have not yet been fully accommodated
in current econometric practice, which makes it hard to empirically validate both the
statistical and the economic significance of the various hypotheses.
Many of the real-world economic and financial time-series drift in a stochastic or
unpredictable manner, meaning that they cannot be turned into stationary by simply
substracting a deterministic function of time. However, by taking first differences these
* Associate Professor, Department of Financial and Management Engineering, The School of Business Studies,
University of the AEGEAN, Greece. E-mail: nthomaid{g.dounias}@fme.aegean.gr
* * Ph.D Student, Department of Financial and Management Engineering, University of AEGEAN, Greece.
E-mail: g.dounias@aegean.gr
Arbitrage is the financial term for free-lunch or making money out of nothing. An
opportunity for arbitrage arises when two similar, in terms of final pay-offs, assets sell
at different prices in the market. By selling the expensive and buying the cheap one, one
holds a riskless portfolio, which costs nothing to the investor but still produces
non-negative profits. Apparently, such a portfolio constitutes a rather attractive
investment which many rational profit-hunting investors will seek to buy. Hence, in the
long-run, market activity will bring the prices of the two assets so close that arbitrage is
no more feasible. The necessity for the near-equality of the two prices is often referred
to as the Law of One Price.
Nowadays, it is common practice in finance to use (no) arbitrage-type arguments to
derive equilibrium or fair prices for various financial products. This approach to pricing
has the advantage of not making strong assumptions on the interaction of agents or their
attitude towards risk. Popular applications of the concept of arbitrage are in the pricing
of stocks and derivatives (futures, options, swaps, etc.), as well as in explaining the term
structure of interest rates. We attempt a brief review of the rationale of these no-arbitrage
theories.
One of the most popular conceptual framework for pricing common stocks is the
Arbitrage Pricing Theory (APT), which goes back to the seminar works Ross (1996, 1997).
One- or many-factor arbitrage models pose an equilibrium relationship for the price of a
traded stock which excludes any arbitrage opportunity. To derive the equilibrium model,
APT rests on two assumptions. The first one concerns the return-generating process and
reflects the intuition that the return on each security is driven by certain economic or
market factors and the sensitivity of the return to these factors. If ri is the ith securitys
return at the end of the investment period then in the case of a K-factor model the
return-generating process is assumed to follow the linear specification
ri ri i1 r f1 i 2 rf 2 i k rf k t ...(1)
where
E(ri)
E(rB) A
rf
E
i1
E
1
E
i2
Rational
fluctuation
bands
Fundamental value
Irrational Time
Deviation
Figure 2 gives a rough approximation of how real-world prices behave. In some markets
irrational deviations from the fundamental value may appear less often than in others, or
dont even last for long. This results in coarser price patterns with occasional spikes. The
area of inaction around the fundamental value has also to do with the nature and the
characteristics of the market and the security under study (liquidity, magnitude of
transaction costs, limitations to short-selling). One expects that the size of irrational
deviations and the time passes until those are eliminated may change with the market or
the security. Still, the analysis above can serve as a general methodological tool for
studying the dynamics of financial prices, which mainly dictates that fundamental indices
are inadequate to explain much of the short-term variation of security prices. Next sections
shows how these heuristic ideas can be implemented into a solid econometric
methodology.
if the amount t = f(xit, x2t) by which actual observations deviate from this equilibrium
is a stationary process with zero median; i.e., the error or discrepancy between outcome
and postulated equilibrium has a fixed distribution, centered on zero, which does not
change overtime.8 Intuitively, the concept of stationarity is a natural requirement that the
error cannot grow indefinitely; if it did, the two variables would not have a tendency to
satisfy (3) and hence it could not be an equilibrium relationship.
Most economic and financial time-series exhibit a trending behavior and are thus
non-stationary. However, econometricians have noticed that by taking their first
difference 'x t x t x t 1 , the resulting time-series normally become stationary. This
implies that the original time-series trend like random walks, where past shocks to the
series do not die out but continuously accumulate. A time-series which can be made
stationary by differencing it once is said to be integrated of order one, denoted by I/(1), or
to have a unit root.9 For a long time it was common practice in econometrics to estimate
equilibrium equations involving unit-root variables by straightforward linear regression.
However, is has been proven that testing hypotheses about the coefficients of the
regression using standard statistical inference might lead to completely spurious results!
In an influential paper, Granger et al.(1974) demonstrated that tests of such a regression may
often suggest a statistically significant relationship between variables where none in fact exists.
Over time, econometricians have suggested many simple solutions to the spurious
regression problem. If econometric relationships are specified after proper transformation
of the original data, e.g. taking first differences, logs or detrending, the statistical
difficulties due to non-stationarity are often avoided because transformed variables
become near stationary (this is actually the standard procedure suggested in many
econometric textbooks, see e.g. Box et al. (1994)). Although data transformations provide
an ad hoc technical solution to the problem of spurious repression, they cause other
problems in the modeling procedure. If an economic or financial theory is stated in terms
of levels of variables, then in estimating models relating transformed variables one
typically loses much of the theoretical guidance. Not to mention that with such models
it may no longer be feasible to empirically check the validity of hypotheses postulated by
the theory, (e.g., certain restrictions in variabless coefficients, etc.).
To overcome the problem of spurious regressions, modern econometric science focuses
on the concept of cointegration and the detection of long-run relationships among
cointegrated variables. Two variables xt and yt which are both I(1) are said to be
8
In the rest of the paper, we adopt a more practical definition of stationarity which requires that the main moments
of a processmean, variance and autocovariancedo not change over time. This is often termed as weak
stationarity.
9
If a discrete-time stochastic process can be made weakly stationary by differencing it d times, it is called integrated
of order d, or I(d). Weakly stationary stochastic processes are thus I(0).
...(4b)
where , are white noise. We also assume that the two variables are tied-up to
the long-run relationship yt , so that yt xt ~ I(0). The Granger Representation
Theorem states that in this case the system can be equivalently written as:
p-1 p 1
t 1 yt 1 1 1j t j 1j yt j 1t ...(5a)
j 1 j 1
p-1 p 1
yt 2 yt 1 1 2j t j 2j yt j 2t ...(5b)
j 1 j 1
where at least one of the parameters 1, 2 are significantly far from zero. The term
yt xt in (5) is the disequilibrium error and 1, 2 measure the strength or speed of the
disequilibrium correction. A specification such as (5), which includes variable differences
and an error-correction term, is generally called an equilibrium- or error-correction model.
There are a few important things to note about ECMs. First, as seen by (5) all variables
10
Strictly speaking, the components of a vector xt = (x1t, x2t, ..., xnt) are said to be cointegrated of order d, b, denoted
by xt ~ CI(d,b), if a) all components of xt are integrated of order d and b) there exists a vector = ( 1, 2, ..., n)'
such that a linear combination '. x = 1x1t + 2x2t + ... + nxnt is integrated of order (d b), where b > 0. Note
that if xt has n components there may be as many as n 1 linearly independent cointegrating vectors. The number
of cointegrating vectors is called the cointegrating rank of xt. Clearly if xt contains only two variables, there can be
at most one independent cointegrating vector.
11
An alternative procedure for testing and estimating cointegrating relationships and error-correction systems,
which tends to become an econometric common-practice, has been developed by Johansen (1988, 1991). This one
builds on simultaneous estimation of the full error-correction system (including the cointegrating vector) using
maximum likelihood and is particularly useful with more than three cointegrated variables. When three or more
cointegrated variables are considered it is possible that there be more than one cointegrating long-run
relationships. In this case, Johansens trace test is used to determine the effective number of cointegrating vectors
is here used as a notation for a vector and not as an operator on a vector.
...(6a)
if < Zt1<
i1 i
...(6b)
where denotes the difference operator dt(.) is a n x 1 vector of constants. In the case
of a two-regime model i = 1, 2 ( 0 = , 2 =+ ) and in the case of three regimes
i = 1, 2, 3 ( 0 = , 3 = + ).12 Generally, {ci(L), Gi(Zti)} depend on the regime i and
12
One can also assume a delay in the correction of mispricing and thus use Ztd in the place of Zt1, for some low
integer d.
and
-d1, Zt 1 d1W
di 0, 1W Zt 1 d2W ...(6d)
d3 , Zt 1 ! 2W
where in the central regime (surrounding the equilibrium) the coefficient of the error
term is 0, incorporating a band of inaction in the region [ 1W
, 2W
].
Various specifications for TECMs have been examined in by Balke et al. (1997), Enders
et al. (1998, 2001), Hansen et al. (2000, 2002), Tsay (1998). A comprehensive presentation
and evaluation of many of these models is given by Clements et al. (2003) in the context
of forecasting the spread between short- and long-term US interest rates.
For specifying a TECM one typically starts with testing for cointegration and especially
the hypothesis of a non-linear threshold adjustment against the linear one. Such tests
have been proposed in the references given above. The thresholds can be determined by
estimating the coefficients of each regime model (6a) over a grid of permissable threshold
values, ensuring that a minimum number of observations falls into each regime. Estimates
of thresholds are obtained by minimizing the total sum of squared errors. To determine
the number of regimes in a TECM one naturally turns to theoretical considerations
regarding the specific market or security, although formal statistical procedures, such as
the Akaikes Information Criterion, can be also employed, (see e.g., Tsay (1998)).
3.2.2. Smooth Transition Error-Correction Models (STECM)
While the threshold cointegration model is appealing since it reflects the intuition that
investors only respond to large deviations from equilibrium, the thresholds introduce
difficulties into the modeling process because they implicitly assume that agents respond
in a homogeneous way to disequilibrium errors. In practice, investors are unlikely to have
common activation thresholds because they encounter different transaction costs
associated with portfolio adjustments (brokerss fees, commissions, tax liabilities, etc.), and
also differ in the time needed to detect the deviation and organize the trading. The fact
that different investors might have different thresholds of inaction suggests that the
overall market thresholds may become blurred as one aggregates across individual
investors. Thus, the resulting equilibrium adjustment might not simply be an on/off
process, as a TECM suggests. Smooth transition and gradual weakening of adjustment as
the market moves closer towards equilibrium might provide a more realistic representation
of the aggregate adjustment process.
T2h (T)
1
G(T) = 1 e , h(T) ...(8b)
2 e
Figure 3: The Symmetric Error Adjustment where , > 0 and T R.
Function (8a) for Different Values of Figures 3 and 4 depict several plots
1 of (8a) and (8b) for different
values of the parameters. Note
0.9
that is a measure of the overall
0.8 tendency to return to equilibrium,
0.7 while in h(.) accounts for asymmetries
0.6 in the adjustment processes. When
> 0 (< 0) the speed of
G(T)
0.5
correction is slower (faster) for
0.4 positive deviations than for
0.3 negative ones. Such asymmetries
=100
0.2 in the adjustment process may
=50
0.1 well hold in practice.
=10
0 The smooth adjustment
0.5 0 0.5 functions (8) can be imbedded in
T
a more general framework to obtain
where s is the activation function, usually specified as sigmoidal, and i0 is the offset
or bias term. See [14] for details and proofs on the above.
To illustrate the application of cointegration methodology in CI, let us suppose that
variables Yt, Xt are I(1), in that their changes are stationary, and also co- integrated so that
there exist a vector of parameters such that Zt = Yt ' . Xt is a stationary variable.
Let us assume that we possess a sample of observations (or a training sample as often called
in CI literature) for Yt and Wt, where Wt ( Yti, i = 1, 2, ..., p; Xtj, j = 1, 2, ...,
q ; Zt1).13 Based on the training sample, CI methods are capable of adaptively learning the
underlying relationship between Yt and Wt , offering parametrization of the form
Yt f( Wt t ...(12)
The main advantage of Cl-based equilibrium correction models is that they offer rich
specifications which can possibly accommodate arbitrary nonlinear features of the
underlying data-generating process (especially the non-linearity in the response of Yt to
the disequilibrium error). Theoretically speaking, (12) can be considered as a generalization
of threshold and smooth transition error correction models (6), (7) and (9) presented in
section 3.2.
Nevertheless, the nonparametric feature of intelligent methods is often the source of
many problems. First of all, nonparametric methods are always prone to overfitting,
meaning that they can be designed to fit a certain data sample as well as possible. However,
when applied to unseen data their performance is poorer than a simple parsimonious model.
Several heuristic procedures have been proposed in the literature for robustifying a
nonparametric model and thus avoid overfitting, but none is superior to others.
Another disadvantage of nonparametric methods stems from the fact they perform
arbitrary non-linear transformations to the original variables which are solely driven by
data and not by theory. One should be concerned about whether these types of nonlinear
transformations make sense from a financial or economic point of view. To illustrate the
point, take for example the discussion on the response of a cointegrating variable to the
disequilibrium error. From a financial point of view, one would reasonably expect that as
the value of Zt becomes higher it leads to more abrupt corrections in future price and
hence have a larger impact on the target variable Yt. In other words, the partial derivative
of the output signal of an intelligent method to an input variable should resemble an
inverse bell function, like the ones depicted in figures 8a and 8b, which grows
exponentially as the value of the input variable becomes larger in absolute terms. It is by
no means guaranteed that the parametrization offered by intelligent models can effectively
capture this property, without being necessary to adding much complexity.
13
W is here used as a notation for a vector and not as an operator on a vector.
15
In the context of cointegrating relationships smooth transition models represent a special type of
equilibrium relationship departing from a long run linear equilibrium relationship and smoothly adjusting
to a new one.
Acknowledgment: I am grateful to Dr Uzay Kaymak, Erasmus University Rotterdam, and Dr Georg Dorffner,
Austrian Research Institute for Artificial Intelligence (OFAI), for comments and suggestions on earlier versions
of this paper. Remaining errors, omissions and opinions are at my responsibility. This research is financially
supported by the Public Benefit Foundation Alexander S. Onassis (www.onassis.gr), under the 2003-2004
scholarships program, and by a grant from Empeirikion foundation (www.empirikion.gr).
Reference # 36J-2006-09-03-01
References
1. Abu-Mostafa Y, Atiya A, Magdon-Ismail M and White H (2001), IEEE Transactions
on Neural Networks, Vol. 12, No. 4, pp. 653-949. Special Issue on Financial
Engineering.
2. Anderson H M (1997), Transaction Costs and Non-linear Adjustment Towards
Equilibrium in the US Treasury Bill Market, Oxford Bulletin of Economics and
Statistics, Vol. 59, No. 4, pp. 465-484.
3. Anderson J A (1995), An Introduction to Neural Networks, The M. I. T. Press,
Cambridge, MA.