120x Filetype PDF File size 0.66 MB Source: www.shivajicollege.ac.in
16 Markov Chains The preceding chapter focused on decision making in the face of uncertainty about one future event (learning the true state of nature). However, some decisions need to take into account uncertainty about many future events. We now begin laying the groundwork for decision making in this broader context. In particular, this chapter presents probability models for processes that evolve over time in a probabilistic manner. Such processes are called stochastic processes.After briefly introducing general stochastic processes in the first section, the remainder of the chapter focuses on a special kind called a Markov chain. Markov chains have the special prop- erty that probabilities involving how the process will evolve in the future depend only on the present state of the process, and so are independent of events in the past. Many processes fit this description, so Markov chains provide an especially important kind of probability model. 16.1 STOCHASTIC PROCESSES A stochastic process is defined to be an indexed collection of random variables {Xt}, where the index t runs through a given set T. Often T is taken to be the set of non- negative integers, and Xt represents a measurable characteristic of interest at time t. For example, Xt might represent the inventory level of a particular product at the end of week t. Stochastic processes are of interest for describing the behavior of a system operating over some period of time. A stochastic process often has the following structure. The current status of the system can fall into any one of M 1 mutually exclusive cate- gories called states. For notational convenience, these states are labeled 0, 1, . . . , M. The random variable Xt represents the state of the system at time t, so its only possible values are 0, 1, . . . , M. The system is observed at particular points of time, labeled t 0, 1, 2, . . . . Thus, the stochastic process {X } {X , X , X , . . .} provides a math- t 0 1 2 ematical representation of how the status of the physical system evolves over time. This kind of process is referred to as being a discrete time stochastic process with a finite state space. Except for Sec. 16.8, this will be the only kind of stochastic process consid- ered in this chapter. (Section 16.8 describes a certain continuous time stochastic process.) 802 ▲| ▲ | e-Text Main Menu | Textbook Table of Contents | 16.2 MARKOV CHAINS 803 An Inventory Example Consider the following inventory problem. A camera store stocks a particular model cam- era that can be ordered weekly. Let D , D , . . . represent the demand for this camera (the 1 2 number of units that would be sold if the inventory is not depleted) during the first week, second week, . . . , respectively. It is assumed that the D are independent and identically i distributed random variables having a Poisson distribution with a mean of 1. Let X rep- 0 resent the number of cameras on hand at the outset, X the number of cameras on hand 1 at the end of week 1, X the number of cameras on hand at the end of week 2, and so on. 2 Assume that X 3. On Saturday night the store places an order that is delivered in time 0 for the next opening of the store on Monday. The store uses the following order policy: If there are no cameras in stock, the store orders 3 cameras. However, if there are any cameras in stock, no order is placed. Sales are lost when demand exceeds the inventory on hand. Thus, {X} for t 0, 1, . . . is a stochastic process of the form just described. t The possible states of the process are the integers 0, 1, 2, 3, representing the possible num- ber of cameras on hand at the end of the week. The random variables X are dependent t and may be evaluated iteratively by the expression max{3D ,0} if X 0 X t1 t t1 max{X D ,0} if X 1, t t1 t for t 0, 1, 2, . . . . This example is used for illustrative purposes throughout many of the following sec- tions. Section 16.2 further defines the particular type of stochastic process considered in this chapter. 16.2 MARKOV CHAINS Assumptions regarding the joint distribution of X , X , . . . are necessary to obtain ana- 0 1 lytical results. One assumption that leads to analytical tractability is that the stochastic process is a Markov chain, which has the following key property: A stochastic process {X} is said to have the Markovian property if P{X jX t t1 0 k , X k , . . . , X k ,X i}P{X jX i}, for t 0, 1, . . . and every 0 1 1 t1 t1 t t1 t sequence i, j, k , k , . . . , k . 0 1 t1 In words, this Markovian property says that the conditional probability of any future “event,” given any past “event” and the present state X i, is independent of the past t event and depends only upon the present state. A stochastic process {Xt} (t 0, 1, . . .) is a Markov chain if it has the Markovian property. The conditional probabilities P{X jX i} for a Markov chain are called (one- t1 t step) transition probabilities. If, for each i and j, P{X jX i}P{X jX i}, for all t 1, 2, . . . , t1 t 1 0 then the (one-step) transition probabilities are said to be stationary. Thus, having sta- tionary transition probabilities implies that the transition probabilities do not change ▲ | ▲ | e-Text Main Menu | Textbook Table of Contents | 804 16 MARKOV CHAINS over time. The existence of stationary (one-step) transition probabilities also implies that, for each i, j, and n (n 0, 1, 2, . . .), P{X jX i}P{X jX i} tn t n 0 for all t 0, 1, . . . . These conditional probabilities are called n-step transition proba- bilities. To simplify notation with stationary transition probabilities, let p P{X jX i}, ij t1 t (n) p P{X jX i}. ij tn t (n) Thus, the n-step transition probability pij is just the conditional probability that the sys- tem will be in state j after exactly n steps (time units), given that it starts in state i at any (1) 1 time t. When n 1, note that p p . ij ij (n) Because the pij are conditional probabilities, they must be nonnegative, and since the process must make a transition into some state, they must satisfy the properties (n) pij 0, for all i and j; n 0, 1, 2, . . . , and M (n) pij 1 for all i; n 0, 1, 2, . . . . j0 A convenient way of showing all the n-step transition probabilities is the matrix form State 0 1 … M (n) (n) … (n) 0 p p p 00 01 0M (n) (n) … (n) 1 p p p 10 11 1M (n) ………………………… P , for n 0, 1, 2, . . . (n) (n) … (n) Mpp p M0 M1 MM or, equivalently, the n-step transition matrix State 0 1 … M (n) (n) … (n) 0 p00 p01 p0M (n) (n) … (n) (n) 1 p10 p11 p1M P … … … … (n) (n) … (n) M pM0 pM1 pMM Note that the transition probability in a particular row and column is for the transition from the row state to the column state. When n 1, we drop the superscript n and sim- ply refer to this as the transition matrix. 1 (0) For n 0, p is just P{X jX i} and hence is 1 when i j and is 0 when i j. ij 0 0 ▲ | ▲ | e-Text Main Menu | Textbook Table of Contents | 16.2 MARKOV CHAINS 805 The Markov chains to be considered in this chapter have the following properties: 1. A finite number of states. 2. Stationary transition probabilities. We also will assume that we know the initial probabilities P{X i} for all i. 0 Formulating the Inventory Example as a Markov Chain Returning to the inventory example developed in the preceding section, recall that X is t the number of cameras in stock at the end of week t (before ordering any more), where X represents the state of the system at time t. Given that the current state is X i,the t t expression at the end of Sec. 16.1 indicates that X depends only on D (the demand t1 t1 in week t 1) and X. Since X is independent of any past history of the inventory sys- t t1 tem, the stochastic process {X } (t 0, 1, . . .) has the Markovian property and so is a t Markov chain. Now consider how to obtain the (one-step) transition probabilities, i.e., the elements of the (one-step) transition matrix State 0 1 2 3 0 p00 p01 p02 p03 1 p p p p P 10 11 12 13 2 p20 p21 p22 p23 3 p p p p 30 31 32 33 given that D has a Poisson distribution with a mean of 1. Thus, t1 n 1 (1) e P{Dt1n}, for n 0, 1, . . . , n! so 1 P{D 0}e 0.368, t1 1 P{D 1}e 0.368, t1 1 1 P{D 2}e 0.184, t1 2 P{D 3}1P{D 2}1(0.3680.3680.184)0.080. t1 t1 For the first row of P, we are dealing with a transition from state X 0 to some state t X . As indicated at the end of Sec. 16.1, t1 X max{3D ,0} if X 0. t1 t1 t Therefore, for the transition to X 3 or X 2 or X 1, t1 t1 t1 p03 P{Dt1 0} 0.368, p02 P{Dt1 1} 0.368, p01 P{Dt1 2} 0.184. ▲ | ▲ | e-Text Main Menu | Textbook Table of Contents |
no reviews yet
Please Login to review.