jagomart
digital resources
picture1_Stochastic Calculus Pdf 171682 | Stochastic Calculus And Applications


 132x       Filetype PDF       File size 0.47 MB       Source: dec41.user.srcf.net


File: Stochastic Calculus Pdf 171682 | Stochastic Calculus And Applications
part iii stochastic calculus and applications based on lectures by r bauerschmidt notes taken by dexter chua lent 2018 these notes are not endorsed by the lecturers and i have ...

icon picture PDF Filetype PDF | Posted on 26 Jan 2023 | 2 years ago
Partial capture of text on file.
                   Part III — Stochastic Calculus and Applications
                             Based on lectures by R. Bauerschmidt
                                  Notes taken by Dexter Chua
                                       Lent 2018
                      These notes are not endorsed by the lecturers, and I have modified them (often
                     significantly) after lectures. They are nowhere near accurate representations of what
                       was actually lectured, and in particular, all errors are almost surely mine.
                    – Brownian motion. Existence and sample path properties.
                    – Stochastic calculus for continuous processes. Martingales, local martingales, semi-
                      martingales, quadratic variation and cross-variation, Itˆo’s isometry, definition of
                      the stochastic integral, Kunita–Watanabe theorem, and Itˆo’s formula.
                    – Applications to Brownian motion and martingales. L´evy characterization of
                      Brownian motion, Dubins–Schwartz theorem, martingale representation, Gir-
                      sanov theorem, conformal invariance of planar Brownian motion, and Dirichlet
                      problems.
                    – Stochastic differential equations. Strong and weak solutions, notions of existence
                      and uniqueness, Yamada–Watanabe theorem, strong Markov property, and
                      relation to second order partial differential equations.
                  Pre-requisites
                  Knowledge of measure theoretic probability as taught in Part III Advanced Probability
                  will be assumed, in particular familiarity with discrete-time martingales and Brownian
                  motion.
                                          1
                                 Contents                                    III Stochastic Calculus and Applications
                                 Contents
                                 0 Introduction                                                                          3
                                 1 The Lebesgue–Stieltjes integral                                                       6
                                 2 Semi-martingales                                                                      9
                                     2.1   Finite variation processes     . . . . . . . . . . . . . . . . . . . . . .    9
                                     2.2   Local martingale . . . . . . . . . . . . . . . . . . . . . . . . . . .       11
                                     2.3   Square integrable martingales . . . . . . . . . . . . . . . . . . . .        15
                                     2.4   Quadratic variation . . . . . . . . . . . . . . . . . . . . . . . . . .      17
                                     2.5   Covariation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .      22
                                     2.6   Semi-martingale      . . . . . . . . . . . . . . . . . . . . . . . . . . .   24
                                 3 The stochastic integral                                                              25
                                     3.1   Simple processes . . . . . . . . . . . . . . . . . . . . . . . . . . .       25
                                     3.2   Itˆo isometry . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .    26
                                     3.3   Extension to local martingales . . . . . . . . . . . . . . . . . . . .       28
                                     3.4   Extension to semi-martingales . . . . . . . . . . . . . . . . . . . .        30
                                     3.5   Itˆo formula   . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .   32
                                     3.6   The L´evy characterization . . . . . . . . . . . . . . . . . . . . . .       35
                                     3.7   Girsanov’s theorem . . . . . . . . . . . . . . . . . . . . . . . . . .       37
                                 4 Stochastic differential equations                                                     41
                                     4.1   Existence and uniqueness of solutions        . . . . . . . . . . . . . . .   41
                                     4.2   Examples of stochastic differential equations . . . . . . . . . . . .         45
                                     4.3   Representations of solutions to PDEs . . . . . . . . . . . . . . . .         47
                                 Index                                                                                  51
                                                                             2
                                 0 Introduction                              III Stochastic Calculus and Applications
                                 0      Introduction
                                 Ordinary differential equations are central in analysis. The simplest class of
                                 equations tend to look like
                                                                      x˙ (t) = F(x(t)).
                                 Stochastic differential equations are differential equations where we make the
                                 function F “random”. There are many ways of doing so, and the simplest way
                                 is to write it as
                                                                  x˙ (t) = F(x(t)) + η(t),
                                 where η is a random function. For example, when modeling noisy physical
                                 systems, our physical bodies will be subject to random noise. What should we
                                 expect the function η to be like? We might expect that for |t − s| ≫ 0, the
                                 variables η(t) and η(s) are “essentially” independent. If we are interested in
                                 physical systems, then this is a rather reasonable assumption, since random noise
                                 is random!
                                     In practice, we work with the idealization, where we claim that η(t) and
                                 η(s) are independent for t 6= s. Such an η exists, and is known as white noise.
                                 However, it is not a function, but just a Schwartz distribution.
                                     To understand the simplest case, we set F = 0. We then have the equation
                                                                           x˙ = η.
                                 Wecan write this in integral form as
                                                                x(t) = x(0)+Z tη(s) ds.
                                                                                 0
                                 To make sense of this integral, the function η should at least be a signed measure.
                                 Unfortunately, white noise isn’t. This is bad news.
                                     Weignore this issue for a little bit, and proceed as if it made sense. If the
                                 equation held, then for any 0 = t < t < ···, the increments
                                                                       0    1
                                                                                 Z t
                                                                                     i
                                                             x(ti) − x(ti−1) = t       η(s) ds
                                                                                    i−1
                                 should be independent, and moreover their variance should scale linearly with
                                 |ti − ti−1|. So maybe this x should be a Brownian motion!
                                     Formalizing these ideas will take up a large portion of the course, and the
                                 work isn’t always pleasant. Then why should we be interested in this continuous
                                 problem, as opposed to what we obtain when we discretize time? It turns out
                                 in some sense the continuous problem is easier. When we learn measure theory,
                                 there is a lot of work put into constructing the Lebesgue measure, as opposed
                                 to the sum, which we can just define. However, what we end up is much easier
                                                                                P
                                —it’s easier to integrate 1 than to sum            ∞ 1. Similarly, once we have set
                                                              x3                   n=1 n3
                                 up the machinery of stochastic calculus, we have a powerful tool to do explicit
                                 computations, which is usually harder in the discrete world.
                                     Another reason to study stochastic calculus is that a lot of continuous
                                 time processes can be described as solutions to stochastic differential equations.
                                 Compare this with the fact that functions such as trigonometric and Bessel
                                 functions are described as solutions to ordinary differential equations!
                                                                              3
                                0 Introduction                            III Stochastic Calculus and Applications
                                    There are two ways to approach stochastic calculus, namely via the Itˆo
                                integral and the Stratonovich integral. We will mostly focus on the Itˆo integral,
                                which is more useful for our purposes. In particular, the Itˆo integral tends to
                                give us martingales, which is useful.
                                    To give a flavour of the construction of the Itˆo integral, we consider a simpler
                                scenario of the Wiener integral.
                                Definition (Gaussian space). Let (Ω,F,P) be a probability space. Then a
                                subspace S ⊆ L2(Ω,F,P) is called a Gaussian space if it is a closed linear
                                subspace and every X ∈ S is a centered Gaussian random variable.
                                    An important construction is
                                Proposition. Let H be any separable Hilbert space. Then there is a probability
                                space (Ω,F,P) with a Gaussian subspace S ⊆ L2(Ω,F,P) and an isometry
                                I : H → S. In other words, for any f ∈ H, there is a corresponding random
                                variable I(f) ∼ N(0,(f,f)H). Moreover, I(αf + βg) = αI(f) + βI(g) and
                                (f,g)H = E[I(f)I(g)].
                                Proof. By separability, we can pick a Hilbert space basis (e )∞            of H. Let
                                                                                                     i i=1
                                (Ω,F,P) be any probability space that carries an infinite independent sequence
                                of standard Gaussian random variables X ∼ N(0,1). Then send e to X , extend
                                                                            i                          i     i
                                by linearity and continuity, and take S to be the image.
                                    In particular, we can take H = L2(R+).
                                Definition(Gaussianwhitenoise). AGaussian white noise onR+ isanisometry
                                             2
                                WNfromL (R+) into some Gaussian space. For A ⊆ R+, we write WN(A) =
                                WN(1A).
                                Proposition.
                                    – For A ⊆ R+ with |A| < ∞, WN(A) ∼ N(0,|A|).
                                    – For disjoint A,B ⊆ R+, the variables WN(A) and WN(B) are indepen-
                                      dent.
                                    – If A = S∞ Ai for disjoint sets Ai ⊆ R+, with |A| < ∞,|Ai| < ∞, then
                                                i=1
                                                                      ∞
                                                                      X                  2
                                                         WN(A)=           WN(Ai) in L and a.s.
                                                                      i=1
                                Proof. Only the last point requires proof. Observe that the partial sum
                                                                          n
                                                                 M =XWN(A)
                                                                    n
                                                                         i=1
                                                                        2
                                is a martingale, and is bounded in L as well, since
                                                               n                   n
                                                     EM2=XEWN(Ai)2=X|Ai|≤|A|.
                                                         n
                                                              i=1                 i=1
                                So we are done by the martingale convergence theorem. The limit is indeed
                                                         P
                                WN(A)because 1 =            ∞ 1 .
                                                    A             A
                                                            n=1    i
                                                                           4
The words contained in this file might help you see if this file matches what you are looking for:

...Part iii stochastic calculus and applications based on lectures by r bauerschmidt notes taken dexter chua lent these are not endorsed the lecturers i have modied them often signicantly after they nowhere near accurate representations of what was actually lectured in particular all errors almost surely mine brownian motion existence sample path properties for continuous processes martingales local semi quadratic variation cross it o s isometry denition integral kunita watanabe theorem formula to l evy characterization dubins schwartz martingale representation gir sanov conformal invariance planar dirichlet problems dierential equations strong weak solutions notions uniqueness yamada markov property relation second order partial pre requisites knowledge measure theoretic probability as taught advanced will be assumed familiarity with discrete time contents introduction lebesgue stieltjes finite square integrable covariation simple extension girsanov examples pdes index ordinary central a...

no reviews yet
Please Login to review.