Artículos de investigación
Abstract: We describe, thought of a particular example, the meaning of initial enlargement of filtration, this means, when we expand a filtration with a σ−algebra.
Resumen: En este trabajo presentamos, a través de un ejemplo particular, una breve introducción al tema de engrosamiento de filtraciones, en particular, el caso inicial, es decir, cuando expandimos una filtración con una σ−algebra.
Palabras clave: Filtration, Stochastic processes, Brownian bridge.
1. Introduction
Whenever we want to model a random experiment, we need to
define a probability space (Ω, , P) where the element is a
σ algebra. This structure contains the events for which it is possible to calculate its probability.
Now, consider and , both sub σ algebras, and
suppose that This means that contains more events than . For
example, if represents the result of the n-th flip of a coin and A is the event to get two heads, then
, however,
.
This motivates us to consider a filtration. A filtration is an infinite increasing succession of σ algebras contained in , which have the historical information, but no the future information, of a random phenomenon.
By enlargement of filtration, we refer that we expand the filtration with additional information until we get another filtration. The study of enlargement of filtration began with the work of Itoˆ in 1976 (see [16]), when he gave meaning to the integral
expanding the natural filtration of W with the variable and proving that the integral in this new filtration was well defined.
Two years later, in 1978, Barlow (see [4]) studied the problem of how to enlarge the filtration, in a minimal way, when L is a positive random variable. More precisely, how to make L a stopping time (see Definition 2.9) and preserve the welldefinedness of the stochastic integrals. Since then, many studies about this topic have been addressed, see, for instance, [19], [20] or the related chapters of the books [25] and [26].
The results of enlargement of filtrations have been extensively used in finance to study the consequences of insider trading (see the references of Section 4).
In this work, we will make a brief introduction to this topic through a simple example. Firstly, we review some basic definitions of stochastic processes in Section 2. Then, in Section 3, we will enlarge the filtration with the σ algebra generated by a Brownian motion evaluated t = 1 and, from this, we will explain the basic ideas of expanding a filtration. Finally, in Section 4, we present a simple example of application
2. Basic concepts
In this section, we will make a review of basic concepts which will be needed in the remainder of this paper.
2.1. Stochastic processes and filtrations
A probability space (Ω,F, P) is a shortlist of three elements. Here, Ω is a nonempty set, F is a σ algebra of subsets of Ω and P is a probability measure. The elements of Ω represent all the possible results in a random experiment, while those in F are the possible events of the experiment.
Definition
2.1. A random variable is a function such that, where B(R)
is the σ algebra of Borel.
As an example, let us consider the experiment of throwing a coin. Let X be the random variable that takes the value 1 if the result is head, and 0 if the result is tail. For this experiment, we have Ω = {head, tail} and F = {∅, Ω, {head}, {tail}}.
Definition 2.2. Let (E, B) be a measurable space. A stochastic process, is a family of random variables, such that
i.e., it is considered as function of two variables so that for each couple (t, ω) corresponds X(t, ω). For each
fixed, the function
is called the path of the process.
These processes represent random phenomena that evolve in time, for example, the capital of an insurance company, the price of a stock, or the number of cars passing through a toll booth in any given period of time. When we observe a random phenomenon, we can be interested in the events that occur before or inclusive until instant t. For all fixed t, these events form a σ algebra that is contained in F. Thus, if we are interested in studying random phenomena and we want to include the time factor, we can add to our probability space a filtration.
Definition 2.3. A filtration of F is a non-decreasing family of σ−algebras {Ft}t≥0, namely,
Fs ⊂ Ft, ∀s, t > 0 such as s < t.
Note that the non-decreasing condition makes sense due to the fact that, as time evolves, we have more information about the random phenomenon.
Definition 2.4. A stochastic processes X is adapted to the filtration {Ft}t≥0 if Xt is Ft−measurable ∀ t ≥ 0.
This means that the process can not anticipate the future or, in other words, the possible events up to a given instant t only depend on the past events. In particular, note that whether a process is measurable or not, only depends on F, while to be adapted depends on the filtration. In general, measurable does not imply adapted.
Definition 2.5. The filtration is defined by
and it is called filtration generated by X.
If {Ft}t≥ 0 is a filtration in a probability space (Ω,F, P), we can associate to it other possible filtrations, for example:
1 Ft+ = ∩s>tFs, t ≥ 0.
2 If the probability space is complete, we can denote Ft the smallest σ−algebra that contains to Ft and all elements of F of probability equals to zero.
If F t = Ft+ for each t, it is said the filtration is right cont
If F t = Ft+ for each t, it is said the filtration is right continuous.
We say that a filtration satisfies the usual conditions if it is right continuous and Ft = Ft.
2.2. Some examples of stochastic processes
The different types of stochastic processes are obtained by considering the different characteristics, mainly, the relations of dependence between the random variables that make up the process. For the purposes of this work, we will focus on defining martingales, semimartingales and we will discuss about two important processes: the Brownian motion and the brownian bridge.
We begin with martingales. These processes have a simple interpretation in terms of fair games: if Xs is the capital of a gamer at time s, the expected value (see Definition 2.6) tells us that the average fortune at the future time t, given that the history of the game is known until time s, is its capital at time t, that is, the game is fair because on average the player does not lose or win.
Definition 2.6. An adapted process X is called a martingale with respect to the filtration {Ft}t≥0, if E(|Xt|) < ∞, and if s ≤ t then E(Xt|Fs) = Xs.
If a process X is a martingale, this means that its expected value, in a future time t, is the value of the process in the last moment when it was observed.
Definition 2.7. When the sample paths of a stochastic process X are, almost sure, right continuous with left limits it is said that X is cadlag.
Definition 2.8. If X is a cadlag martingale and, in addition, E( Xt 2) < for all t 0, we will say that X is an square integrable martingale.
To define the next concept, we need some additional notation. If T is a nonnegative random variable and Yt is any process, we define
where T ∧ t stands for the minimum between T and t.
Definicion 2.9. A stopping time τ is a random variable such that, for each t ≥ 0, it is verified that {τ ≤ t} ∈ Ft.
A stopping time registers the moment in which an event of the process occurs. More precisely, at any time t, we can know if this event has occurred or not.
Definition 2.10. A process X is a local martingale with respect to {Ft}t≥0, if there are stopping times {Tn} such that Tn ↑ ∞ so that XTn is a martingale with respect to {Ft∧Tn }t≥0.
Local martingales are important for many reasons, one of those being for the integrability of stochastic proceses.
The classic Ito integral [17] is defined for adapted processes to the filtration for which the integrator is a brownian motion (see Definition 2.13). The idea of definition of this integral is similar to the integral of a function with respet to a measure. First, it is defined for simple processes and then, by approximation, for more general processes. The interested reader can consult
[2] and [21] for a detailed exhibition about the construction of this integral.
Remark 2.11. The idea of Itoˆ integral has been generalized to more general integrators as local martingales and processes of finite variation, namely those processes whose sample paths are functions of finite variation (see Definition 2.17). In addition, it can be shown (it is beyond the scope of these notes, though) that linear combinations of those are, in a sense, all processes that can be used as integrators. These processes are called “good integrators” or semimartingales, reducing the integrability requirements.
Definition 2.12. A processes X is said to be a continuous semimartingale if X can be written as X = M + A ,where M is a continuous local martingale and A is a ca`dla`g adapted process with finite variation on [0, t] for t > 0.
1 Importar tabla For a detailed exposition of the previous concepts, as well as of those of stochastic integration, the interested reader can consult [7], [12], [25] and the references therein.
2.3. Brownian motion
Now, we give a brief introduction to the brownian motion W (see for instance [21], [12], [26], [27]), one of the most important processes in stochastic calculus due to its applications in different disciplines of science (we refer the interested reader to [3], [23]). It was observed, for the first time in 1828, by the botanist Robert Brown (see [8],[9]), when he studied the movement of pollen grains suspended in a certain substance. Based on Brown’s observations, several theories emerged, but it was Einstein [13], in 1905, who gave the correct explanation. The first rigorous mathematical construction of brownian motion is due to Wiener [28].
Definition 2.13. A brownian motion W = Wt t≥0 is an adap ted stochastic process with continuous paths that satisfies the following properties:
1 W0 = 0 with pro bability 1.
2 For s ≤ t, Wt − Ws is independent of Fs.
3 For s ≤ t, Wt − Ws has a normal distribution with mean equal to 0 and variance t − s.
In what follows, we will mention only the properties of Brownian motion that we will be useful in this work.
Proposition 2.14. The brownian motion is an square integrable martingale.
Demostracion. From Property 2 of Definition 2.13, we see
E(Wt − Ws|Fs) = E(Wt − Ws) = 0.
Proposition 2.15. The quadratic variation of brownian motion in the interval [0, t] is t.
In order to establish the proof of this result, it is necessary to give some definitions.
Definition 2.16. The collections of points P = {t0, . . . tn} is a partition of an interval [a, b] if a = t0 < t1 < · · · < tn = b holds.
Definition 2.17. Let f : [a, b] → R be a function. The variation Vf of f over [a, b] is defined to be
P = {t0, . . . tn} is a partition of [a, b]} .
Now, we proceed with the proof of Proposition 2.15.
Demostración. Let be a succession of partitions of [0, t], such that
Then
The following theorem is known as Levy theorem and establishes the conditions that characterize brownian motion. Its proof can be found, for instance in, [7], [12].
Theorem 2.18. Let X be a continuous local martingale such that {X2 –t} t≥0 is a local martingale and X0 = 0. Then, X is a brownian motion with respect to the filtration for which X is a local martingale.
Remark 2.19. Formally speaking, the quadratic variation of a process X is a stochastic process which is denoted as [X] and it is defined by
As a consequence, Proposition 2.15 establishes that [W]t = t.
2.4. Brownian Bridge
This process is derived from brownian motion by requiring an extra constraint and it is used, for example, in Statistics to derivate the Kolmogorov-Smirnov test (the interested reader can consult [11], [22]).
Definition 2.20. The brownian bridge Xt 0≤t≤1 is a Gaussian process (i.e., for any finite sub family (t0, t1, . . . , tn) Rn, the random vector (Xt0 , . . . Xtn ) has a multivariate normal distribution) that satisfies the following:
1 X0 = 0 and X1 = 0 with probability 1.
2 E(Xt) = 0
3 For s ≤ t, cov(Xs, Xt) = s(1 − t).
A natural question is whether such a process exists. The answer is yes. We observe the following result.
Proposition 2.21. Let W be a brownian motion. The process Xt = Wt − tW1 is a brownian bridge.
Demostracion.
E(Xt) = E(Wt − tW1) = E(Wt) − tE(W1) = 0.
if s ≤ t ≤ 1.
cov(Xs, Xt) = E(XsXt) = E((Ws − sW1)(Wt − tW1))
= E(WsWt) − sE(W1Wt) − tE(WsW1) + stE(W1W1)
= s − st − ts + st = s(1 − t).
Proposition 2.21 give us a continuous version of the brownian bridge. In [26], the reader can see how the brownian bridge may be viewed as brownian motion conditioned to W1= 0, namely {Wt , 0 ≤ t ≤ 1| = 0}.
We may consider the brownian bridge between 0 and y for 0 ≤ t ≤ 1 by setting,
Xt = Wt − t(W1 − y).
And more generally, between x and y for 0 t T , by expressing,
For Section 4, we will need the following concept.
2.5. Stochastic differential equations
We consider the non-autonomous stochastic differential equation written formally as
X0 = x0.
The interpretation of this equation is
Here, W is a real-valued
Brownian motion defined on the completed probability space (Ω,F , P), equipped
with the filtration ( : 0 ≤ t ≤ T) that satisfies the usual
conditions. The coefficients b, σ : [0, T] × R → R are measurable functions and
are named drift and diffusion coefficients, respectively, and ∈ R stands for the initial condition. In equation (2), the stochastic
integral is defined in Ito’s sense. The process X = {Xt : t ≥ 0} that solves
(2) is called Ito’s process and its coefficients satisfy certain regularity
conditions, i.e., σ belongs to the class of processes such that: i) for t ∈ (0, T] the relation (s, ω) → σs(ω) defined in (0, t]xΩ is measurable respect of the σ algebra , where
is the Borel σ-algebra on [0, t], and ii)
are satisfied, while b belongs to the class of
processes such that i) and ii’)
hold.
To know about the existence and uniqueness of the solutions of these equations, as well as more details, the reader can
consult, for example [2], [12], [24].
3. Enlargement of filtrations
As we briefly mentioned in the above section, a filtration contains the information related to a given random phenomenon up to the moment t. If we suppose the arrival of new information, this induces to consider a different filtration
such that
for each
For example, in financial markets there are different types of traders whose behavior is induced by different types of information. It is commonly assumed that all traders have the same information. If we assume that the market can accept differences in information between traders, then we will have to consider a new filtration to approach this problem (see, for instance [10] and references therein).
The enlargement of filtration is split into two cases: initial enlargements, for which and progressive enlargement, where
In both cases, σ(L) and
are assumed to represent the new information.
When a filtration, is enlarged some questions need to be answered: in which cases a semimartingale remains a semimartigale and what is the new decomposition in G.
We want to focus our interest in inicial enlargement. For such reason, rather than studying the general problem, we discuss an interesting problem.
3.1. Brownian Bridge
Let us begin with a brownian motion W = {Wt }t≥0 defined in a complete space of probability (Ω, F, P) that satisfies the usual conditions and it is equipped with the natural filtration . Now, we define a new filtration that satisfies the usual conditions as
We know that W is a martingale in Ft, (see Proposition 2.14) however it is not a martingale in the filtration Gt. In fact, let be t ≤ 1 and note that W1 is measurable with respect to G, then
Nevertheless, W is a semimartingale in Gt. The following result presents us the decomposition of W as G− semimartingale.
Proposition 3.1. The process B which is defined as
is a −martingale and a brownian motion
in this same filtration.
Demostración. We begin by computing the following expectation. Let 0 ≤ s < t < 1. By means of the property of independent increments, we have
Notice that the
brownian bridge is independent of
, indeed, cov
and we know that gaussian variables are
independent if they have zero covariance. Thus,
On the other
hand, since is independent of
we have,
Now, we compute
and by Fubini’s Theorem for conditional expectation
Note that we do not have problems in t = 1 since
for some constant c and for all s ∈ [0, 1], therefore
Accordingly, and B is a G−martingale. Also by properties of
the quadratic variation we have
and the Theorem 2.18 gives B is a brownian
motion.
For more details and general version of the results presented in this section the reader can consult [25] and [18].
4. Example
The results of enlargement of filtrations have been used in finance to study the consequences of insider trading ([1], [5], [15], [14]). The insider trading is when some participants of the market have material information which they do not share with the rest of the market. The insider knowledge can produce the existence of arbitrage (that is, the possibility to get profit without risk), or changes in the prices dynamic.
For example, in the Black-Scholes model [6], which it is used to determine the value of certain financial assets called European options, the price S of underlaying asset, due to different factors, evolves randomly, i.e., it is represented by the following stochastic differential equation:
where µ and σ are constants. It is assumed that the asset has a constant interest rate r. The agent invests his money in an investment portfolio, thus his wealth is
We assume that r is a constant interest. The first addition of the right side of the equality (4) represents the investment without risk, while in the second adding πˆ is the number of shares of the risky asset. If we
include as the proportion of wealth invested in the
risky asset and
. The above equation is
Using Itoˆ’s formula (see [2], [24]) and properties of expectation value it is estimated that in the agent can maximize his expected wealth E(ln(XT)) and
Now, we suppose
that we enlarge the filtration with or equivalently with
. Using Proposition 3.1, the
dynamics of S and X, for t < 1, are
Here . Following a similar procedure, it is found
that
where with π ad. in H, we refer to admissible strategies in the σ−algebra H, this means that, once the new price of underlaying asset is announced, the investors readjust their portfolio without adding or consuming money. Note that if T = 1, the the maximum of the expected wealth is infinity, thus there is an arbitrage opportunity.
r
S. Ankirchner, S. Dereich, P. Imkeller, Enlargement of filtrations and continuous Girsanov-type embeddings, Springer (2004).
L. Arnold, Stochastic Differential Equations: Theory and Applications. John Wiley & Sons, New York, (1974)
L. Bachelier, M. Davis, A. Etheridge, Louis Bachelier?s Theory of Specu lation: The Origins of Modern Finance, Princeton University Press, Prin ceton (NJ) (2006).
M. T. Barlow, Study of a filtration expanded to include an honest time, Z. Wahrscheinlichkeitstheorie verw. Gebiete, 44, 307-323, (1978).
F. Baudoin, Modeling anticipations on financial markets, Lecture Notes in Mathematics, Springer- Verlag, 1814, 43-92, (2003).
F. Black, M. Scholes, The Pricing of Options and Corporate Liabilities, The Journal of Political Economy, 81 3, 637-654, (1973).
T. Bojdecki, Teor´ıa General de Procesos e Integracio´n Estoca´stica. Aportaciones Matema´ticas, Serie Textos 6, Soc. Mat. Mex., (1995).
R. Brown, A brief account of Microscopical Observations made in the Months of June, July, and August, 1827, on the Particles contained in the Pollen of Plants; and on the general Existence of active Molecules in Organic and Inorganic Bodies, Philosophical Magazine N. S. 4, 161-173, 1, (828).
R. Brown Additional remarks on active molecules. The Philosophical Mgazine and Annals of Philosophy (new Series) 6 161-166, (1828)
A. M. Corcuera, P. Imkeller, A. Kohatsu-Higa, D. Nualart Additional utility of insiders with imperfect dynamical information. Finance and Stochastics, 8, 437-450, (2004).
W. C. Chow, Brownian Bridge, John Wiley & Sons, Inc. WIREs Comp Stat, 1, 325-332, (2009).
R. Durret, Stochastic Calculus: A practical Introduction. CRC Press, (1996).
A. Einstein, U¨ ber die von der molekularkinetischen Theorie der Wa¨rme geforderte Bewegung von in ruhenden Flu¨ssigkeiten suspendierten Teilchen, Ann. Phys. 17, 549-560, (1905).
A. Grorud, M. Pontier, Insider trading in a continuous time market model, International Journal of Theoretical and Applied Finance, 1, 331-347, (1998).
P. Imkeller, Random times at which insiders can have free lunches, Stochastics and Stochastics Reports, 74, 465-487, (2002).
K. Itoˆ, Extension of stochastic integrals, In Proceedings of the International Symposium on Stochastic Differential Equations, New York, 95-109, (1978).
K. Itoˆ, Stochastic integral. Proc. Imp. Acad., 20, 519-524, (1944).
M. Jeanblanc, M. Yor and M. Chesney, Mathematics methods for financial Markets, Springer-Verlag. (2009).
T. Jeulin, Semi-martingales et grossissement de filtration, Lecture Notes in Mathematics. Springer-Verlag, 833, (1980).
T. Jeulin, M Yor, Grossissements de filtrations: exemples et applications, Lecture Notes in Mathematics. Springer-Verlag, 1118, (1985).
I. Karatzas, S. E. Shreve, Brownian Motion and Stochastic Calculus. Springer-Verlag, New York, (1998).
D. Kwiatkowski, P. C.B. Phillips, et.al., Testing the null hypothesis of stationarity against the alternative of a unit root, Journal of Econometrics. North-Holland, 54, 159-178, (1992) .
D. K. C. MacDonald, Noise and Fluctuations. Wiley, New York (1962). Reprinted by Dover, Mineola (NY) (2006).
B. Øksendal, Stochastic Differential Equations: an introduction with applications, sixth edition. Springer, Berlin (2007).
P. Protter, Stochastic Integration and Differential Equations, second edi tion. Springer, Berlin, (2004).
D. Revuz, M. Yor Continuos Martingales and Brownian Motion, third editon, Springer, Berlin, (1999).
R. L. Schilling, L. Partzsch Brownian Motion. An introduction to stocha- tic processes, D Gruyter, Germany, (2012).
N. Wiener, Differential-space, J. Math.&Phys. 58, 131-174, (1923).