SSRN 1206162
SSRN 1206162
Processes∗
Luc Bauwens
Université catholique de Louvain, CORE†
Nikolaus Hautsch
Humboldt-Universität zu Berlin, CASE, CFS‡
November 2007
Abstract
In this paper, we give an overview of the state-of-the-art in the econometric literature
on the modeling of so-called financial point processes. The latter are associated with the
random arrival of specific financial trading events, such as transactions, quote updates,
limit orders or price changes observable based on financial high-frequency data. After
discussing fundamental statistical concepts of point process theory, we review duration-
based and intensity-based models of financial point processes. Whereas duration-based
approaches are mostly preferable for univariate time series, intensity-based models pro-
vide powerful frameworks to model multivariate point processes in continuous time.
We illustrate the most important properties of the individual models and discuss major
empirical applications.
Keywords: Financial point processes, dynamic duration models, dynamic intensity mod-
els.
JEL Classification: C22, C32, C41
1 Introduction
Since the seminal papers by Hasbrouck (1991) and Engle and Russell (1998) the modelling of
financial data at the transaction level is an ongoing topic in the area of financial economet-
rics. This has created a new body of literature which is often referred to as ”the econometrics
of (ultra-)high-frequency finance” or ”high-frequency econometrics”. The consideration of
the peculiar properties of financial transaction data, such as the irregular spacing in time,
∗
The paper is written as a contribution to the Handbook of Financial Time Series, Springer, 2008. This
research was supported by the Deutsche Forschungsgemeinschaft through the SFB 649 ”Economic Risk”.
†
Université catholique de Louvain and CORE. Address: Voie du Roman Pays 34, 1348, Louvain-la-Neuve,
Belgium. Email: [email protected]
‡
Institute for Statistics and Econometrics and CASE – Center for Applied Statistics and Economics,
Humboldt-Universität zu Berlin as well as Center for Financial Studies (CFS), Frankfurt. Address: Span-
dauer Str. 1, D-10178 Berlin, Germany. Email: [email protected].
Electronic
Electroniccopy
copyavailable
availableat:
at:https://ssrn.com/abstract=1206162
http://ssrn.com/abstract=1206162
the discreteness of price changes, the bid-ask bounce as well as the presence of serial de-
pendence, provoked the surge of new econometric approaches. One important string of the
literature deals with the irregular spacing of data in time. Taking into account the latter
is indispensable whenever the full amount of information in financial transaction data has
to be exploited and no loss of information due to fixed-interval aggregation schemes can
be accepted. Moreover, it has been realized that the timing of trading events, such as the
arrival of particular orders and trades, and the frequency in which the latter occur have
information value for the state of the market and play an important role in market mi-
crostructure analysis, for the modelling of intraday volatility as well as the measurement of
liquidity and implied liquidity risks.
Taking into account the irregular occurrence of transaction data requires to consider it
as a point process, a so-called financial point process. Depending on the type of the financial
”event” under consideration, we can distinguish between different types of financial point
processes or processes of so-called financial durations. The most common types are trade
durations and quote durations as defined by the time between two consecutive trade or quote
arrivals, respectively. Price durations correspond to the time between absolute cumulative
price changes of given size and can be used as an alternative volatility measure. Similarly,
a volume duration is defined as the time until a cumulative order volume of given size is
traded and captures an important dimension of market liquidity. For more details and
illustrations, see Bauwens and Giot (2001) or Hautsch (2004).
One important property of transaction data is that market events are clustered over time
implying that financial durations follow positively autocorrelated processes with a strong
persistence. Actually, it turns out that the dynamic properties of financial durations are
quite similar to those of daily volatilities. Taking into account these properties leads to
different types of dynamic models on the basis of a duration representation, an intensity
representation or a counting representation of a point process.
In this chapter, we review duration-based and intensity-based models of financial point
processes. In Section 2, we introduce the fundamental concepts of point process theory and
discuss major statistical tools. In Section 3, we review the class of dynamic duration models.
Specifying a (dynamic) duration model is presumably the most intuitive way to characterize
a point process in discrete time and has been suggested by Engle and Russell (1998), which
was the starting point for a huge body of literature. Nevertheless, Russell (1999) realized
that a continuous-time setting on the basis of the intensity function constitutes a more
flexible framework which is particularly powerful for the modelling of multivariate processes.
Different types of dynamic intensity models are presented in Section 4.
Electronic
Electroniccopy
copyavailable
availableat:
at:https://ssrn.com/abstract=1206162
http://ssrn.com/abstract=1206162
2 Fundamental Concepts of Point Process Theory
In this section, we discuss important concepts and relationships in point process theory
which are needed throughout this chapter. In Section 2.1, we introduce the notation and
basic definitions. The fundamental concepts of intensity functions, compensators and hazard
rates are defined in Section 2.2, whereas in Section 2.3 different classes and representations
of point processes are discussed. Finally, in Section 2.4, we present the random time change
theorem which yields a powerful result for the construction of diagnostics for point process
models. Most concepts discussed in this section are based upon Chapter 2 of Karr (1991).
Let {ti }i∈{1,...,n} denote a random sequence of increasing event times 0 < t1 < . . . < tn
P
associated with an orderly (simple) point process. Then, N (t) := i≥1 1l {ti ≤t} defines the
right-continuous (càdlàg) counting function. Throughout this chapter, we consider only
point processes which are integrable, i.e. E[N (t)] < ∞ ∀ t ≥ 0. Furthermore, {Wi }i∈{1,...,n}
denotes a sequence of {1, . . . , K}-valued random variables representing K different types
of events. Then, we call the process {ti , Wi }i∈{1,...,n} an K-variate marked point process
on (0, ∞) as represented by the K sequences of event-specific arrival times {tki }i∈{1,...,nk } ,
k = 1, . . . , K, with counting functions N k (t) := i≥1 1l {ti ≤t} 1l {Wi =k} .
P
The internal history of an K-dimensional point process N (t) is given by the filtration FtN
with FtN = σ(N k (s) : 0 ≤ s ≤ t, k ∈ Ξ), N k (s) = i≥1 1l {ti ≤s} 1l {Wi ∈Ξ} , where Ξ denotes
P
the σ-field of all subsets of {1, . . . , K}. More general filtrations, including e.g. also processes
of explanatory variables (covariates) {zi }i∈{1,...,n} are denoted by Ft with FtN ⊆ Ft .
Define xi := ti − ti−1 with i = 1, . . . , n and t0 := 0 as the inter-event duration from
P
ti−1 until ti . Furthermore, x(t) with x(t) := t − tN̆ (t) , with N̆ (t) := i≥1 1l {ti <t} denoting
the left-continuous counting function, is called the backward recurrence time. It is a left-
continuous function that grows linearly through time with discrete jumps back to zero after
each arrival time ti . Finally, let θ ∈ Θ denote model parameters.
where Λ̃(t) is the (unique) compensator of N (t). This relationship emerges from the in-
terpretation of the compensator as integrated (conditional) hazard function. Consequently,
λ(t) can be also defined by the relation
Z s
E[N (s) − N (t)|Ft ] = E λ(u)du Ft (2)
t
which has to hold (almost surely) for all t, s with 0 ≤ t ≤ s. Letting s ↓ t leads to the
heuristic representation which is more familiar in classical duration analysis. Then, λ(t) is
obtained by
1
λ(t+) := lim E [ N (t + ∆) − N (t)| Ft ] , (3)
∆↓0 ∆
where λ(t+) := lim∆↓0 λ(t + ∆). In case of a stationary point process, λ̄ := E[dN (t)]/dt =
E[λ(t)] is constant.
Equation (3) manifests the close analogy between the intensity function and the hazard
function which is given by
1
h(x) := f (x)/S(x) = lim Pr[x ≤ X < x + ∆|X ≥ x] (4)
∆→0 ∆
with x denoting the (inter-event) duration as represented by the realization of a random
variable X with probability density function f (x), survivor function S(x) = 1 − F (x), and
cumulative distribution function (cdf) F (x) = Pr[X ≤ x]. Whereas the intensity function
is defined in (continuous) calendar time, the hazard rate is typically defined in terms of the
length of a duration x and is a key concept in (cross-section) survival analysis.
The simplest type of point process is the homogeneous Poisson process defined by
with ∆ ↓ 0. Then, λ > 0 is called the Poisson rate corresponding to the (constant)
intensity. Accordingly, equations (5) and (6) define the intensity representation of a Poisson
process. A well-known property of homogenous Poisson processes is that the inter-event
where θ = (γ1 , γ2 ), h0 (·) denotes the so-called baseline hazard rate and g(·) is a function
of the covariates z and parameters γ2 . The baseline hazard rate may be parameterized
in accordance with a certain distribution, like e.g., a Weibull distribution with parameters
λ, p > 0 implying
h0 (x|γ1 ) = λp(λx)p−1 . (8)
For p = 1 we obtain the exponential case h0 (x|γ1 ) = λ, implying a constant hazard rate.
Alternatively, if p > 1, ∂h0 (x|γ1 )/∂x > 0, i.e. the hazard rate is increasing with the length of
the spell which is referred to as ”positive duration dependence”. In contrast, p < 1 implies
”negative duration dependence”. Non-monotonic hazard rates can be obtained with more
flexible distributions, like the generalized F and particular cases thereof, including the
generalized gamma, Burr, Weibull and log-logistic distributions. We refer to the Appendix
to Chapter 3 of Bauwens and Giot (2001) and to the Appendix of Hautsch (2004) for
definitions and properties. Alternatively, the baseline hazard may be left unspecified and
can be estimated nonparametrically, see Cox (1975).
An alternative type of duration model is the class of accelerated failure time (AFT)
models given by
h(x|z; θ) = h0 [xg(z, γ2 )|γ1 ]g(z, γ2 ). (9)
Rt
where ω is a constant, w(s) denotes a non-negative weight function, and 0 w(s)dN (s)
is the stochastic Stieltjes integral of the process w with respect to the counting pro-
cess N (t). The process (10) was proposed by Hawkes (1971) and is therefore named a
Hawkes process. If w(s) declines with s, then, the process is self-exciting in the sense that
Cov[N (a, b), N (b, c)] > 0, where 0 < a ≤ b < c. Different types of Hawkes processes and
their applications to financial point processes are presented in Section 4.1. A further type
of intensity models which is relevant in the literature of financial point processes is given
by a specification where the intensity itself is driven by an autoregressive process which is
updated at each point of the process. This leads to a special type of point process models
which does not originate from the classical point process literature but originates from the
autoregressive conditional duration (ACD) literature reviewed in Section 2 and brings time
series analysis into play. Such a process is called an autoregressive conditional intensity
One fundamental result of martingale-based point process theory is the (multivariate) ran-
dom time change theorem by Meyer (1971) which allows to transform a wide class of point
processes to a homogeneous Poisson process:
Theorem (Meyer, 1971, Brown and Nair, 1988): Assume a multivariate point process
(N 1 (t), . . . , N K (t)) is formed from the event times {tki }i∈{1,...,nk } , k = 1, . . . , K, and has
continuous compensators (Λ̃(t)1 , . . ., Λ̃(t)K ) with Λ̃k (∞) = ∞ for each k = 1, . . . , K, then
the point processes formed from {Λ̃k (tki )}{i=1,...,nk } , k = 1, . . . , K, are independent Poisson
processes with unit intensity.
Proof: See Meyer (1971) or Brown and Nair (1988) for a more accessible and elegant
proof.
In this section, we discuss univariate dynamic models for the durations between consecutive
(financial) events. In Section 3.1, we review in detail the class of ACD models, which is
by far the most used class in the literature on financial point processes. In Section 3.2,
The class of ACD models has been introduced by Engle and Russell (1997, 1998) and Engle
(2000). In order to keep the notation simple, define xi in the following as the inter-event
duration which is standardized by a seasonality function s(ti ), i.e. xi := (ti − ti−1 )/s(ti ).
The function s(ti ) is typically parameterized according to a spline function capturing time-
of-day or day-of-week effects. Time-of-day effects arise because of systematic changes of the
market activity throughout the day and due to opening of other related markets. In most
approaches s(ti ) is specified according to a linear or cubic spline function and is estimated
separately in a first step yielding seasonality adjusted durations xi . Alternatively, a non-
parametric approach has been proposed by Veredas, Rodriguez-Poo, and Espasa (2002).
For more details and examples regarding seasonality effects in financial duration processes,
we refer the reader to Chapter 2 of Bauwens and Giot (2001) or to Chapter 3 of Hautsch
(2004).
The key idea of the ACD model is to model the (seasonally adjusted) durations {xi }i=1,...,n
in terms of a multiplicative error term model in the spirit of Engle (2002), i.e.
xi = Ψi i , (12)
where Ψi denotes a function of the past durations (and possible covariates), and εi defines
an i.i.d. random variable for which it is assumed that
E[i ] = 1, (13)
so that Ψi corresponds to the conditional duration mean (the so-called ”conditional dura-
tion”) with Ψi := E[xi |Fti−1 ]. The ACD model can be rewritten in terms of the intensity
function as
!
x(t) 1
λ(t|Ft ) = λ , (14)
ΨN̆ (t)+1 ΨN̆ (t)+1
where λ (s) denotes the hazard function of the ACD error term. This formulation clearly
demonstrates that the ACD model belongs to the class of AFT models. Assuming i to
be standard exponentially distributed yields the so-called Exponential ACD model. More
flexible specifications arise by assuming i to follow a more general distribution, see the
discusssion after equation (8). It is evident that the ACD model is the counter-part to the
GARCH model (Bollerslev (1986)) for duration processes. Not surprisingly, many results
and specifications from the GARCH literature have been adapted to the ACD literature.
Since Ψi must be positive, the restrictions ω > 0, α ≥ 0 and β ≥ 0 are usually imposed. It
is also assumed that β = 0 if α = 0, otherwise β is a redundant parameter. The process
defined by (12), (13) and (16) is known to be covariance-stationary if
where σ 2 := Var[i ] < ∞, and to have the following moments and autocorrelations:
α (1−β 2 −α β)
(3) ρ1 = 1−β 2 −2 α β
and ρn = (α + β)ρn−1 (n ≥ 2).
The condition (17) ensures the existence of the variance. These results are akin to those for
the GARCH(1,1) zero-mean process. They can be generalized to ACD(p,q) processes when
p, q > 1. It is usually found empirically that the estimates of the parameters are such that
α + β is in the interval (0.85,1) while α is in the interval (0.01,0.15). Since the ACD(1,1)
model can be written as
where g(i−1 ) is either ln i−1 (log-ACD of type I) or i−1 (type II). Using this setting, it
is convenient to augment Ψi by functions of covariates, see e.g. Bauwens and Giot (2001).
The stochastic process defined by (12), (13) and (19) is covariance-stationary if
Its mean, variance and autocorrelations are given in Section 3.2 in Bauwens and Giot (2001),
see also Fernandes and Grammig (2006) and Bauwens, Galli, and Giot (2008). Drost and
Werker (2004) propose to combine one of the previous ACD equations for the conditional
duration mean with an unspecified distribution for i , yielding a class of semi-parametric
ACD models.
The augmented ACD (AACD) model introduced by Fernandes and Grammig (2006)
provides a more flexible specification of the conditional duration equation than the previous
models. Here, Ψi is specified in terms of a Box-Cox transformation yielding
Ψδi 1 = ω + βΨδi−1
1
+ αΨδi−1
1
[|i−1 − ξ| − ρ(i−1 − ξ)]δ2 ,
where δ1 > 0, δ2 > 0, ξ, and ρ are parameters. The so-called news impact function
[|i−1 − ξ| − ρ(i−1 − ξ)]δ2 allows for a wide variety of shapes of the curve tracing the impact
of i−1 on Ψi for a given value of Ψi−1 and the remaining parameters. The parameter ξ is a
shift parameter and the parameter ρ is a rotation parameter. If ξ = ρ = 0, the linear ACD
model is obtained by setting δ1 = δ2 = 1, the type I logarithmic ACD model by letting
δ1 and δ2 tend to 0, and the type II version by letting δ1 tend to 0 and setting δ2 = 1.
Fernandes and Grammig (2006) compare different versions of the AACD model using IBM
price durations arising from trading at the New York Stock Exchange. Their main finding
is that ”letting δ1 free to vary and accounting for asymmetric effects (by letting ξ and ρ
free) seem to operate as substitute sources of flexibility”. Hautsch (2006) proposes an even
more general augmented ACD model that nests in particular the so-called EXponential
ACD model proposed by Dufour and Engle (2000) implying a kinked news impact function.
As a counterpart to the semiparametric GARCH model proposed by Engle and Ng (1993),
Hautsch (2006) suggests specifying the news impact function in terms of a linear spline
function based on the support of εi . He illustrates that the high flexibility of this model is
needed in order to appropriately capture the dynamic properties of financial durations.
10
(j) (j)
xi = Ψi i , (21)
(j)
Ψi = ω (j) + β (j) Ψi−1 + α(j) xi−1 (22)
when xi−1 ∈ [rj−1 , rj ), and 0 = r0 < r1 < . . . < rJ = ∞ are the threshold parameters.
The superscript (j) indicates that the distribution or the model parameters can vary with
the regime operating at observation i. This model can be viewed as a mixture of J ACD
models, where the probability to be in regime j at i is equal to 1 and the probabilities to
be in each of the other regimes is equal to 0. Hujer, Vuletic, and Kokot (2002) extend this
model to let the regime changes be governed by a hidden Markov chain.
While the TACD model implies discrete transitions between the individual regimes,
Meitz and Teräsvirta (2006) propose a class of smooth transition ACD (STACD) models
which generalize the linear and logarithmic ACD models in a specific way. Conditions for
strict stationarity, ergodicity, and existence of moments for this model and other ACD
models are provided in Meitz and Saikkonen (2004) using the theory of Markov chains. A
motivation for the STACD model is, like for the AACD, to allow for a nonlinear impact of
the past duration on the next expected duration.
The estimation of most ACD models can be easily performed by maximum likelihood (ML).
Engle (2000) demonstrates that the results by Bollerslev and Wooldridge (1992) on the
quasi-maximum likelihood (QML) property of the Gaussian GARCH(1,1) model extend to
the Exponential-ACD(1,1) model. Then, QML estimates are obtained by maximizing the
quasi-loglikelihood function given by
n
X xi
ln L θ; {xi }{i=1,...,n} =− ln Ψi + . (23)
Ψi
i=1
For more details we refer to Chapter 3 of Bauwens and Giot (2001), Chapter 5 of Hautsch
(2004), and to the survey of Engle and Russell (2005).
Residual diagnostics and goodness-of-fit tests are straightforwardly performed by eval-
uating the stochastic properties of the ACD residuals ˆi = xi /Ψ̂i . The dynamic properties
11
ACD models strongly resemble ARCH models. Therefore it is not surprising that Taylor’s
(1986) stochastic volatility model for financial returns has been a source of inspiration of
similar duration models. Bauwens and Veredas (2004) propose the stochastic conditional
duration model (SCD) as an alternative to ACD-type models. The SCD model relates to
the logarithmic ACD model in the same way as the stochastic volatility model relates to
the exponential GARCH model of Nelson (1991). Thus the model is defined by equations
(12), (13), and
ln Ψi = ω + β ln Ψi−1 + γi−1 + ui , (24)
where ui is iid N(0, σu2 ) distributed. The process {ui } is assumed to be independent of
the process {i }. The set of possible distributions for the duration innovations i is the
same as for ACD models. This model allows for a rich class of hazard functions for xi
through the interplay of two distributions. The latent variable Ψi may be interpreted as
being inversely related to the information arrival process which triggers bursts of activity
on financial markets. The ”leverage” term γi−1 in (24) is added by Feng, Jiang, and Song
(2004) to allow for an intertemporal correlation between the observable duration and the
conditional duration, and the correlation is found to be positive. Bauwens and Veredas
(2004) use a logarithmic transformation of (12) and employ QML estimation based on
the Kalman filter. Knight and Ning (2005) use the empirical characteristic function and
the method of generalized moments. Strickland, Forbes, and Martin (2003) use Bayesian
estimation with a Markov chain Monte Carlo algorithm. For the model with leverage term,
12
Ui H(1, F1i )
xi = = , (25)
aVi aH(b, F2i )
where Ui and Vi are two independent variables which are gamma(1,1) (i.e. exponential) and
gamma(b, b) distributed, respectively. The last ratio in (25) uses two independent Gaussian
factors F1i and F2i , and H(b, F ) = G(b, ϕ(F )), where G(b, .) is the quantile function of the
gamma(b, b) distribution and ϕ(.) the cdf of the standard normal distribution. Ghysels,
Gourieroux, and Jasiak (2004) extend this model to a dynamic setup through a VAR model
for the two underlying Gaussian factors. Estimation is relatively difficult and requires
simulation methods.
3.4 Applications
ACD models can be used to estimate and predict the intra-day volatility of returns from
the intensity of price durations. As shown by Engle and Russell (1998), a price intensity is
closely linked to the instantaneous price change volatility. The latter is given by
" #
p(t + ∆) − p(t) 2
2 1
σ̃ (t) := lim E Ft , (26)
∆↓0 ∆ p(t)
where p(t) denotes the price (or midquote) at t. By denoting the counting process associated
with the event times of cumulated absolute price changes of size dp by N dp (t), we can
formulate (26) in terms of the intensity function of the process of dp-price changes. Then,
the dp-price change instantaneous volatility can be computed as
dp 2
2 1
σ̃(dp) (t) = lim Pr [|p(t + ∆) − p(t)| ≥ dp |Ft ] ·
∆↓0 ∆ p(t)
1 h i dp 2
dp dp
= lim Pr (N (t + ∆) − N (t)) > 0 |Ft ·
∆↓0 ∆ p(t)
2
dp
:= λdp (t) · , (27)
p(t)
13
14
In this section, we review the most important types of dynamic intensity models which
are applied to model financial point processes. The class of Hawkes models and extensions
thereof are discussed in Section 4.1. In Section 4.2, we survey different autoregressive
intensity models. Statistical inference for intensity models is presented in Section 4.3,
whereas the most important applications in the recent literature are briefly discussed in
Section 4.4.
Hawkes processes originate from the statistical literature in seismology and are used to
model the occurrence of earthquakes, see e.g. Vere-Jones (1970), Vere-Jones and Ozaki
(1982), and Ogata (1988) among others. Bowsher (2006) was the first applying Hawkes
models to financial point processes. As explained in Section 3.2, Hawkes processes belong
to the class of self-exciting processes, where the intensity is driven by a weighted function
of the time distance to previous points of the process. A general class of univariate Hawkes
processes is given by
P
λ(t) = ϕ µ(t) + ti <t w(t − ti ) , (28)
15
where αj ≥ 0, βj > 0 for j = 1, . . . , P are model parameters, and P denotes the order of the
process and is selected exogenously (or by means of information criteria). The parameters
αj are scale parameters, whereas βj drive the strength of the time decay. For P > 1, the
intensity is driven by the superposition of differently parameterized exponentially decaying
weighted sums of the backward times to all previous points. In order to ensure identification
we impose the constraint β1 > . . . > βP . It can be shown that the stationarity of the process
R∞
requires 0 < 0 w(s)ds < 1, which is ensured only for Pj=1 αj /βj < 1, see Hawkes (1971).
P
H
w(t) = , (30)
(t + κ)p
with parameters H, κ, and p > 1 allows for a hyperbolic decay. Such weight functions are
typically applied in seismology (see e.g. Vere-Jones and Ozaki (1982) and Ogata (1988))
and allow to capture long range dependence. Since financial duration processes also tend to
reveal long memory behavior (see Jasiak (1998)), specification (30) might be an interesting
specification in financial applications.
Multivariate Hawkes models are obtained by a generalization of (28). Then, λ(t) is given
16
where wrk (s) is a k-type weight function of the backward time to all r-type events. Using
an exponential decay function, Hawkes (1971) suggests to parameterize wrk (s) as
P
k
k −βr,j t
X
wrk (t) = αr,j e , (32)
j=1
k ≥ 0 and β k > . . . > β k > 0 drive the influence of the time distance to past
where αr,j r,1 r,P
r-type events on the k-type intensity. Thus, in the multivariate case, λk (t) depends not
only on the distance to all k-type points, but also on the distance to all other points of the
pooled process. Hawkes (1971) provides a set of linear parameter restrictions ensuring the
stationarity of the process.
Bowsher (2006) proposes a generalization of the Hawkes model which allows to model
point processes which are interrupted by time periods where no activity takes place. In
high-frequency financial time series these effects occur because of trading breaks due to
trading halts, nights, weekends or holidays. In order to account for such effects, Bowsher
proposes to remove all non-activity periods and to concatenate consecutive activity periods
by a spill-over function.
Hamilton and Jordà (2002) establish a natural link between ACD models and intensity
models by extending the ACD model to allow for covariates which might change during a
duration spell (time-varying covariates). The key idea of their so-called autoregressive con-
ditional hazard (ACH) model is to rely on the fact that in the ACD model with exponential
error distribution, the intensity (or the hazard function, respectively) corresponds to the
inverse of the conditional duration, i.e. λ(t) = Ψ−1 . They extend this expression by a
N̆ (t)+1
function of variables which are known at time t − 1,
1
λ(t) = 0 γ, (33)
ΨN̆ (t)+1 + zt−1
where zt are time-varying covariates which are updated during a duration spell.
An alternative model which can be seen as a combination of a duration model and an
intensity model is introduced by Gerhard and Hautsch (2007). They propose a dynamic
extension of a Cox (1972) proportional intensity model, where the baseline intensity λ0 (t)
is non-specified. Their key idea is to exploit the stochastic properties of the integrated
intensity and to re-formulate the model in terms of a regression model with unknown left-
hand variable and Gumbel distributed error terms – see Kiefer (1988) for a nice illustration
17
where ΦN̆ (t)+1 captures the dynamic structure, λk0 (t) is a baseline intensity component cap-
turing the (deterministic) evolution of the intensity between two consecutive points and
sk (t) denotes a deterministic function of t capturing, for instance, possible seasonality ef-
fects. The function ΦN̆ (t) is indexed by the left-continuous counting function and is updated
instantaneously after the arrival of a new point. Hence, Φi is constant for ti−1 < t ≤ ti .
Then, the evolution of the intensity function between two consecutive arrival times is gov-
erned by λk0 (t) and sk (t).
In order to ensure the non-negativity of the process, the dynamic component Φki is
specified in log-linear form, i.e.
Φki = exp Φ̃ki + zi−1
0 γk , (35)
where zi denotes a vector of explanatory variables observed at arrival time ti and γ k the
corresponding parameter vector. Define εi as a (scalar) innovation term which is computed
from the integrated intensity function associated with the most recently observed process,
i.e.
K Z tk
N k (ti )
X
εi = 1 − λk (s; Fs )ds yik , (36)
k=1 tk
N k (ti )−1
where yik defines an indicator variable that takes the value 1 if the i-th point of the pooled
process is of type k. Using the random time change argument presented in Section 2.4, εi
corresponds to a random mixture of i.i.d. centered standard exponential variates and thus
18
where Ak = {akj } denotes a (K ×1) innovation parameter vector and B k = {bkij } is a (K ×K)
matrix of persistence parameters. Hence, the fundamental principle of the ACI model is that
at each event ti all K processes are updated by the realization of the integrated intensity with
respect to the most recent process, where the impact of the innovation on the K processes
can be different and also varies with the type of the most recent point. As suggested by
Bowsher (2006), an alternative specification of the ACI innovation term might be ε̃i =
1 − Λ(ti−1 , ti ), where Λ(ti−1 , ti ) := K k
P
k=1 Λ (ti−1 , ti ) denotes the integrated intensity of the
pooled process computed between the two most recent points. Following the arguments
above, ε̃i is also a zero mean i.i.d. innovation term. Because of the regime-switching nature
of the persistence matrix, the derivation of stationarity conditions is difficult. However, a
sufficient (but not necessary) condition is that the eigenvalues of the matrices B k for all
k = 1, . . . , K lie inside the unit circle.
As proposed by Hautsch (2004), the baseline intensity function λk0 (t) can be specified
as the product of K different Burr hazard rates, i.e.
K s
Y xr (t)pr −1
λk0 (t) = exp(ω ) k
, (psr > 0, ηrs ≥ 0). (38)
1 + ηrs xr (t)psr
r=1
According to this specification λk (t) is driven not only by the k-type backward recurrence
time but also by the time distance to the most recent point in all other processes r = 1, . . . , K
with r 6= k. A special case occurs when psr = 1 and ηrs = 0, ∀ r 6= s. Then, the k-th process
is affected only by its own backward recurrence time.
Finally, sk (t) is typically specified as a spline function in order to capture intraday
seasonalities. A simple parameterization which is used in most studies is given by a linear
spline function of the form sk (t) = 1 + Sj=1 νjk (t − τj ) · 1l {t>τj } , where τj , j = 1 . . . , S,
P
denote S nodes within a trading period and νj the corresponding parameters. A more
flexible parameterization is e.g. given by a flexible Fourier form (Gallant (1981)) as used by
Andersen and Bollerslev (1998) or Gerhard and Hautsch (2002) among others.
If K = 1 and η11 = 0, the ACI model and the ACD model coincide. Then, the ACI
model corresponds to a re-parameterized form of the Log-ACD model. If the ACI model is
extended to allow for time-varying covariates (see Hall and Hautsch (2007)), it generalizes
the approach by Hamilton and Jordà (2002). In this case, all event times associated with
(discrete time) changes of time-varying covariates are treated as another point process that
19
where λ∗N̆ (t)+1 := λ∗ (tN̆ (t)+1 ) denotes the common latent component which is updated
at each point of the (pooled) process {ti }i∈{1,...,n} . The direction and magnitude of the
process-specific impact of λ∗ is driven by the parameters σk∗ . The process-specific function
λo,k (t) := λo,k (t|Fto ) denotes a conditionally deterministic idiosyncratic k-type intensity
component given the observable history, Fto .
Bauwens and Hautsch (2006) assume that λ∗i has left-continuous sample paths with
right-hand limits and in logarithm is the zero mean AR(1) process given by
Because of the symmetry of the distribution of ln λ∗i , Bauwens and Hautsch impose an
identification assumption which restricts the sign of one of the scaling parameters σk∗ . The
observation-driven component λo,k (t) is specified in terms of an ACI parameterization as
described above. However, in contrast to the basic ACI model, in the SCI model, the
innovation term is computed based on the observable history of the process, i.e.
K n
X o
εi = −$ − ln Λo,k tkN k (ti )−1 , tkN k (ti ) yik , (41)
k=1
20
N (tki )−1
X −σ∗
= λ∗j k
Λk (tj , tj+1 ) (42)
j=N (tki−1 )
corresponding to the sum of (piecewise) integrated k-type intensities which are observed
through the duration spell and are standardized by the corresponding (scaled) realizations
of the latent component. This specification ensures that εi can be computed exclusively
based on past observables implying a distinct separation between the observation-driven
and the parameter-driven components of the model. Bauwens and Hautsch (2006) analyze
the probabilistic properties of the model and illustrate that the SCI model allows for a wide
range of (cross-)autocorrelation structures in multivariate point processes. In an application
to a multivariate process of price intensities, they find that the latent component captures
a substantial part of the cross-dependences between the individual processes resulting in a
quite parsimonious model. An extension of the SCI model to the case of multiple states
is proposed by Koopman, Lucas, and Monteiro (2005) and is applied to the modelling of
credit rating transitions.
Karr (1991) shows that valid statistical inference can be performed based on the intensity
function solely, see Theorem 5.2. in Karr (1991) or Bowsher (2006). Assume a K-variate
point process N (t) = {N k (t)}K
k=1 on (0, T ] with 0 < T < ∞, and the existence of a K-
variate Ft -predictable process λ(t) that depends on the parameters θ. Then, it can be
shown that a genuine log likelihood function is given by
K
"Z #
X T Z
k k k
ln L θ; {N (t)}t∈(0,T ] = (1 − λ (s))ds + ln λ (s)dN (s) ,
k=1 0 (0,T ]
Note that (43) differs from the standard log likelihood function of duration models by the
additive (integrating) constant T K which can be ignored for ML estimation. By apply-
ing the so-called exponential formula (Yashin and Arjas (1988)), the relation between the
integrated intensity function and the conditional survivor function is given by
21
with the k-type intensity component. Given that the parameters are variation free, the
log likelihood function can be computed as ln L θ; {N (t)}t∈(0,T ] = K k k
P
k=1 l (θ ) and can
be maximized by maximizing the individual k-type components lk (θk ) separately. This fa-
cilitates the estimation particularly when K is large. In contrast, ACI models require to
maximize the log likelihood function with respect to all the parameters jointly. This is due
to the fact that the ACI innovations are based on the integrated intensities which depend
on all individual parameters. The estimation of SCI models is computationally even more
demanding since the latent factor has to be integrated out resulting in a n-dimensional
integral. Bauwens and Hautsch (2006) suggest to evaluate the likelihood function numeri-
cally using the efficient importance sampling procedure introduced by Richard and Zhang
(2005). Regularity conditions for the maximum likelihood estimation of stationary simple
point processes are established by Ogata (1981). For more details, see Bowsher (2006).
Diagnostics for intensity based point process models can be performed by exploiting
the stochastic properties of compensators (see Bowsher (2006)) and integrated intensities
given in Section 2.4. The model goodness-of-fit can be straightforwardly evaluated through
the estimated integrated intensities of the K individual processes, eki,1 := Λ̂k (tki−1 , tki ), the
integrated intensity of the pooled process ei,2 := Λ̂(ti−1 , ti ) = K k
P
k=1 Λ̂ (ti−1 , ti ), or of the
PK
(non-centered) ACI residuals ei,3 := k=1 Λ̂k (tki−1 , tki ) yik . Under correct model specifica-
tion, all three types of residuals must be i.i.d. standard exponential. Then, model evaluation
is done by testing the dynamic and distributional properties. The dynamic properties are
easily evaluated with Portmanteau statistics or tests against independence such as proposed
by Brock, Scheinkman, Scheinkman, and LeBaron (1996). The distributional properties can
be evaluated using Engle and Russell’s (1998) test against excess dispersion (see Section
22
4.4 Applications
For financial point processes, dynamic intensity models are primarily applied in multivari-
ate frameworks or whenever a continuous-time setting is particularly required, like, for
instance, in order to allow for time-varying covariates. One string of applications focusses
on the modelling of trading intensities of different types of orders in limit order books. Hall
and Hautsch (2007) apply a bivariate ACI model to study the intensities of buy and sell
transactions in the electronic limit order book market of the Australian Stock Exchange
(ASX). The buy and sell intensities are specified to depend on time-varying covariates cap-
turing the state of the market. On the basis of the buy and sell intensities, denoted by
λB (t) and λS (t), Hall and Hautsch (2007) propose a measure of the continuous net buy
pressure defined by ∆B (t) := ln λB (t) − ln λS (t). Because of the log-linear structure of the
ACI model, the marginal change of ∆B (t) induced by a change of the covariates is computed
as γ B − γ S , where γ B and γ S denote the coefficients associated with covariates affecting
the buy and sell intensity, respectively (see eq. (35)). Hall and Hautsch (2006) study the
determinants of order aggressiveness and traders’ order submission strategy at the ASX by
applying a six-dimensional ACI model to study the arrival rates of aggressive market orders,
limit orders as well as cancellations on both sides of the market. In a related paper, Large
(2007) studies the resiliency of an electronic limit order book by modelling the processes
of orders and cancellations on the London Stock Exchange using a ten-dimensional Hawkes
process. Finally, Russell (1999) analyzes the dynamic interdependences between the sup-
ply and demand for liquidity by modelling transaction and limit order arrival times at the
NYSE using a bivariate ACI model.
Another branch of the literature focusses on the modelling of the instantaneous price
change volatility which is estimated on the basis of price durations, see (27) in Section
3.4. This relation is used by Bauwens and Hautsch (2006) to study the interdependence
between instantaneous price change volatilities of several blue chip stocks traded at the
New York Stock Exchange (NYSE) using a SCI model. In this setting, they find a strong
evidence for the existence of a common latent component as a major driving force of the
instantaneous volatilities on the market. In a different framework, Bowsher (2006) analyzes
the two-way interaction of trades and quote changes using a two-dimensional generalized
Hawkes process.
23
Bauwens, L., F. Galli, and P. Giot (2008): “The Moments of Log-ACD Models,”
Quantitative and Qualitative Analysis in Social Sciences, forthcoming.
Bauwens, L., and P. Giot (2000): “The Logarithmic ACD Model: An Application to
the Bid/Ask Quote Process of two NYSE Stocks,” Annales d’Economie et de Statistique,
60, 117–149.
Bauwens, L., and D. Veredas (2004): “The Stochastic Conditional Duration Model: A
Latent Factor Model for the Analysis of Financial Durations,” Journal of Econometrics,
119, 381–412.
Bisière, C., and T. Kamionka (2000): “Timing of Orders, Order Aggressiveness and the
Order Book at the Paris Bourse,” Annales d’Economie et de Statistique, 60, 43–72.
24
Brown, T. C., and M. G. Nair (1988): “A Simple Proof of the Multivariate Random
Time Change Theorem for Point Processes,” Journal of Applied Probability, 25, 210–214.
Cox, D. R. (1972): “Regression Models and Life Tables,” Journal of the Royal Statistical
Society, Series B, 34, 187–220.
Daley, D., and D. Vere-Jones (2003): An Introduction to the Theory of Point Processes,
vol. 1. Springer, New York.
De Luca, G., and G. Gallo (2004): “Mixture Processes for Financial Intradaily Du-
rations,” Studies in Nonlinear Dynamics and Econometrics, 8 (2), Downloadable under
http://www.bepress.com/snde/vol8/iss2/art8.
De Luca, G., and P. Zuccolotto (2003): “Finite and Infinite Mixtures for Financial
Durations,” Metron, 61, 431–455.
Dionne, G., P. Duchesne, and M. Pacurara (2005): “Intraday Value at Risk (IVaR)
Using Tick-by-Tick Data with Application to the Toronto Stock Exchange,” Mimeo, HEC
Montréal.
Dufour, A., and R. F. Engle (2000): “The ACD Model: Predictability of the Time
between Consecutive Trades,” Working Paper, ISMA Centre, University of Reading.
25
Easley, D., and M. O‘Hara (1992): “Time and Process of Security Price Adjustment,”
The Journal of Finance, 47, 577–605.
(2002): “New Frontiers for ARCH Models,” Journal of Applied Econometrics, 17,
425–446.
Engle, R. F., and A. Lunde (2003): “Trades and Quotes: A Bivariant Point Process,”
Journal of Financial Econometrics, 11, 159–188.
Engle, R. F., and V. K. Ng (1993): “Measuring and Testing the Impact of News on
Volatility,” Journal of Finance, 48, 1749–1778.
Feng, D., G. J. Jiang, and P. X.-K. Song (2004): “Stochastic Conditional Duration
Models with ‘Leverage Effect’ for Financial Tranbsaction Data,” Journal of Financial
Econometrics, 2, 390–421.
Fernandes, M., and J. Grammig (2005): “Non-parametric Specification Tests for Con-
ditional Duration Models,” Journal of Econometrics, 127, 35–68.
Gallant, R. A. (1981): “On the Bias in Flexible Functional Forms and an Essential
Unbiased Form: The Fourier Flexible Form,” Journal of Econometrics, 15, 211–245.
Gerhard, F., and N. Hautsch (2002): “Volatility Estimation on the Basis of Price
Intensities,” Journal of Empirical Finance, 9, 57–89.
26
Ghysels, E., and J. Jasiak (1998): “GARCH for Irregularly Spaced Financial Data: The
ACD-GARCH Model,” Studies in Nonlinear Dynamics and Econometrics, 2, 133–149.
Giesecke, K., and P. Tomecek (2005): “Dependent Events and Changes of Time,”
Working Paper, Cornell University.
Giot, P. (2005): “Market Risk Models for Intraday Data,” European Journal of Finance,
11, 187–212.
Grammig, J., and M. Wellner (2002): “Modeling the Interdependence of Volatility and
Inter-Transaction Duration Process,” Journal of Econometrics, 106, 369–400.
Hall, A. D., and N. Hautsch (2006): “Order Aggressiveness and Order Book Dynamics,”
Empirical Economics, 30, 973–1005.
(2007): “Modelling the Buy and Sell Intensity in a Limit Order Book Market,”
Journal of Financial Markets, 10, 249–286.
Hamilton, J. D., and O. Jordà (2002): “A Model of the Federal Funds Rate Target,”
Journal of Political Economy, 110, 1135–1167.
Han, A., and J. A. Hausman (1990): “Flexible Parametric Estimation of Duration and
Competing Risk Models,” Journal of Applied Econometrics, 5, 1–28.
Hawkes, A. G. (1971): “Spectra of Some Self-Exciting and Mutually Exciting Point Pro-
cesses,” Biometrika, 58, 83–90.
27
Hujer, R., and S. Vuletic (2005): “Econometric Analysis of Financial Trade Processes
by Discrete Mixture Duration Models,” Available at http://ssrn.com/abstract=766664.
Hujer, R., S. Vuletic, and S. Kokot (2002): “The Markov Switching ACD Model,” Fi-
nance and Accounting Working Paper 90, Johann Wofgang Goethe-University, Frankfurt.
Available at SSRN: http://ssrn.com/abstract=332381.
Karr, A. F. (1991): Point Processes and their Statistical Inference. Dekker, New York.
Knight, J., and C. Ning (2005): “Estimation of the Stochastic Conditional Duration
Model via Alternative Methods – ECF and GMM,” Mimeo, University of Western On-
tario.
Koopman, S. J., A. Lucas, and A. Monteiro (2005): “The Multi-State Latent Factor
Intensity Model for Credit Rating Transitions,” Discussion Paper TI2005-071/4, Tinber-
gen Institute.
Large, J. (2007): “Measuring the Resiliency of an Electronic Limit Order Book,” Journal
of Financial Markets, 10, 1–25.
Meddahi, N., E. Renault, and B. J. Werker (2006): “GARCH and Irregularly Spaced
Data,” Economics Letters, 90, 200–204.
28
Møller, J., and J. Rasmussen (2004): “Perfect Simulation of Hawkes Processes,” Work-
ing Paper, Aalborg University.
Ogata, Y. (1981): “On Lewis’ Simulation Method for Point Processes,” IEEE Transactions
of Information Theory, IT-27, 23–31.
Ogata, Y. (1988): “Statistical Models for Earthquake Occurrences and Residual Analysis
for Point Processes,” Journal of the American Statistical Association, 83, 9–27.
Quoreshi, A. S. (2006): “Long Memory, Count Data, Time Series Modelling for Financial
Application,” Umea Economic Studies 673, Department of Economics, Umea University.
29
Vere-Jones, D., and T. Ozaki (1982): “Some Examples of Statistical Inference Applied
to Earthquake Data,” Annals of the Institute of Statistical Mathematics, 34, 189–207.
Veredas, D., J. Rodriguez-Poo, and A. Espasa (2002): “On the (Intradaily) Sea-
sonality, Dynamics and Durations Zero of a Financial Point Process,” CORE Discussion
Paper 2002/23, Louvain-La-Neuve.
Yashin, A., and E. Arjas (1988): “A Note on Random Intensities and Conditional
Survival Functions,” Journal of Applied Probability, 25, 630–635.
30