首页 miller1997kalman

miller1997kalman

举报
开通vip

miller1997kalman Kalman �ltering for linear systems with coeÆcients driven by a hidden Markov jump process* Boris M. Miller Institute for Information Transmission Problems Russian Academy of Sciences B. Karetny per. 19, 101447 GSP-4 Moscow, Russia and Wolfgang J. Rung...

miller1997kalman
Kalman �ltering for linear systems with coeÆcients driven by a hidden Markov jump process* Boris M. Miller Institute for Information Transmission Problems Russian Academy of Sciences B. Karetny per. 19, 101447 GSP-4 Moscow, Russia and Wolfgang J. Runggaldier Dipartimento di Matematica Pura e Applicata Universit�a di Padova Via Belzoni 7, I - 35131 - Padova, Italy Abstract. We present an explicit algorithm for the solution of the nonlinear �ltering problem concerning a linear, partially observed di�usion-type model, where the coef- �cients are driven by a hidden Markov jump process, of which one can observe the occurence of a jump. The (recursive) �lter is of a "branching" type with its (�nite) dimension growing for each jump of the hidden Markov process. Key words: linear and nonlinear �ltering, �nite dimensional �lters, hidden Markov models, Markovian switching coeÆcients, Kalman �ltering. 1. INTRODUCTION The �ltering problem for di�usion processes, where the coeÆcients depend on an unobservable Markov jump process, has many applications in various �elds such as telecomunications, image processing, tracking of maneuvering objects, and many others. The traditional problem setting considers a partially observable process (x; y; �) = f(x t ; y t ; � t ; t � 0g, generally described by a linear (in x) Ito equation dx t = a t (� t )x t dt+ b t (� t ) dw (1) t ; x 0 � N (m 0 ;� 0 ); dy t = A t (� t )x t dt+ B t dw (2) t ; y 0 = 0; (1:1) where x t and y t are vectors, and � t is a right continuous Markov process with �nite num- ber of states � t 2 f1; � � � ; ng, having matrix of transition intensities � t = f� i;j t g i;j=1;���;n ; w (i) = fw (i) t ; t � 0g; i = 1; 2 are standard vector Wiener processes, a t (�); b t (�); A t (�); B t are matrix functions of appropriate dimensions, and x 0 ; � t ; w (1) ; w (2) are independent. * Work partially performed during a stay of B.M. Miller in Padova, supported by INTAS Grants 93-2622 and 94-697. 1 It is known that, in general, the �lter equations for (x; �), based on the observations y, namely the equations for Efx t j y t 0 g and EfIf� t = ig j y t 0 g where I(�) denotes the indicator function, are nonlinear and in�nite-dimensional (see [1],[8]) just as in the general nonlinear case ([10]). It is also known that the problem of estimating the state of a Markov process on the basis of continuous observations admits a solution in terms of a �nite-dimensional �lter system only if the observation process y t admits a representation of the form y t = Z t 0 C(� s ) ds+ Z t 0 B s dw s ; (1:2) where C(�) is a function of the current state of the process � t , and w t is a Wiener process independent of � t (see [1],[8]). In [8] the authors consider estimation problems for more general unobserved state processes and this latter monograph covers all known cases admitting a �nite dimensional �lter; still, it does not include models of the form (1.1). On the other hand, in the discrete time case the �ltering problem for models of the type (1.1) can always be reduced to a �nite-dimensional recursive system of equations, which, as shown in [12], has dimension n K , where n denotes the number of possible states of the Markov chain and K the maximal value of the discrete time index. The reason is that in the generic time instant k � 1 the set of possible trajectories of the Markov chain that lead to any of the current n possible states is equal to n k and, consequently, in order to obtain a closed system of �lter equations, it suÆces to have a procedure to compute the conditional probabilitites for each trajectory. As shown in [12] (see also [3]), for given transition probabilities these conditional probabilities can be computed with the use of Bayes formula and as a result one obtains a �nite-dimensional system of �lter equations. These considerations are also at the basis of the present study, where, as in [5], we extend the observation process by adding to it the information about the instants of state transition of � t . This means, we can observe that something has happened, but we do not know exactly what (for a justi�cation of this see [5]). If therefore one can observe the process N t , describing the number of these state transitions up to the generic moment t > 0, then for each given state � t = i; i = 1; � � � ; n; the number of possible trajectories leading to this state is equal to n(n� 1) N t . The problem then consists only in computing the conditional probabilities for each of the possible trajectories which, generally speaking, leads to a smoothing problem. Using the information given by the observation of the process N t , one can construct a system of branching equations for the computation of these probabilities where the branching points coincide with the moments of jump of the process N t . The outline of the paper is as follows. In the next section 2 we formulate the problem and give the basic de�nitions. In the third section we derive a �nite-dimensional system of equations for the estimation of the state of the process � t on the basis of the observations of the process N t together with a continuous process y t of the type as in (1.1) which we may rewrite as y t = Z t 0 g s (!; � s )ds+ w t (1:3) 2 for a suitable adapted process g s (!; � s ) and a Wiener process w t . For simplicity we shall assume y to be scalar and w standard Wiener, the extensions to the vector case implying only technical complications. These equations contain terms of the form Efg t (!; � t ) j F Y t g, where F Y t is the �� algebra generated up to time t by the process-pair Y t = (N t ; y t ), and that can be represented as Efg t (!; � t ) j F Y t g = EfEfg t (!; � t ) j F �;Y t g jF Y t g = X � t 0 g^ t (Y t 0 ; � t 0 )Pf� t 0 j F Y t g; (1:4) where the sum extends over the (�nite) set of admissible trajectories � t 0 that correspond to the observation of the process N t 0 = fN s ; 0 � s � tg, and Pf� t 0 j F Y t g are the conditional probabilities of these trajectories. In the fourth section we derive the system of branching forward equations for the computation of Pf� t 0 j F Y t g leading to a �nite- dimensional (for each t > 0 such that N t < +1) closed system of �lter equations, provided we have an explicit way to compute g^ t (Y t 0 ; � t 0 ). Finally, in the �fth section we obtain the solution of the original problem in the form of a �nite-dimensional (w.P.1) system of �lter equations depending on g^. The original �ltering problem is thus fully solved every time one can get an explicit expression for g^. In the linear case, as in model (1.1), g^ can be computed via a Kalman �lter. 2. PROBLEM SETTING Given a probability space ( ;F ; P ) with a right continuous �ltration F t , consider a nonstationary Markov process � t with �nite state space � t 2 f1; � � � ; ng having transition intensity matrix � t = f� ij t g i;j=1;���;n with deterministic and continuous � ij t . Putting X i t := If� t = ig, where IfAg is the indicator function of the set A, consider the vector X t := fX 1 t ; � � � ; X n t g. Then (see [10; Ch.9]) the process X = fX t ; t � 0g admits the representation X t = X 0 + Z t 0 � � s X s ds+M t ; (2:1) where M = fM t ; t � 0g with M t := fM 1 t ; � � � ;M n t g is a square integrable (P;F t )� martingale with quadratic variation (see [4], [9]) hMi t = � Z t 0 [� � s diagX s + diagX s � s ] ds+ Z t 0 diag (� � s X s ) ds; (2:2) where diagX denotes the matrix with diagonal entries X 1 ; � � � ; X n and � the transpose. Suppose that one can observe the two-component process Y = fY t ; t � 0g, where Y t = � y 1 t = N t y 2 t = R t 0 g s (!; � s )ds+ w t � : (2:3) The process N t in (2.3) denotes the number of state transitions of the Markov process � t that occur up to the moment t, w = fw t ; t � 0g is a standard F t � Wiener process, and g s (!; �) is a given scalar F t � adapted random function, which in particular may correspond to an observation equation as in (1.1) (see also (1.3)). Consider also the 3 sequence of �� algebras F Y t := �fY s ; 0 � s � tg � F t and the stopping times f� k g; k = 0; 1; � � � ; such that � 0 = 0; � k = inffs : N s = kg. Since f� k < tg 2 F Y t , the random variables � k are F Y t � measurable and so with each stopping time � k one can associate the �� algebra F Y � k , generated by the events A : A T f� k � tg 2 F Y t . Remark 2.1. It turns out that F Y � k = F Y � � k . In fact, F Y � k = F Y � � k W �f�N � k g. On the other hand, Pf�N � k = 1g = Pf g = 1 and so F Y � � k di�ers from F Y � k only by events of probability zero. Therefore, after completion of F Y � � k by the null sets, we get the required equality, which also implies F Y t = F Y t� ; 8t: We introduce the following de�nitions : for a random variable h(!), the symbol ^ h t denotes ^ h t = Efh(!) j F Y t g; if h = h t (!); then ^ h t = Efh t (!) j F Y t g. Here we keep the notations from [8]. Notice also that g t (!; � t ) = n X i=1 g t (!; i)If� t = ig = n X i=1 g t (!; i)X i t = hg � t (!); X t i: (2:4) Our purpose is to derive an equation for ^ X t = EfX t j F Y t g: (2:5) Notice that for a model of type (1.1), combining our results with a Kalman �lter (see section 5, in particular (5.4)), from ^ X we can also obtain x^ t = Efx t j F Y t g. 3. DERIVATION OF THE FILTER EQUATIONS To derive the �lter equation in closed form we need the following Theorem 3.1. Let the random processes g t (!; i) be square integrable for any t and i = 1; � � � ; n and continuous in t for �xed ! and i. The process ^ X t = EfX t j F Y t g satis�es then the equation d ^ X t = � � t ^ X t� dt+ 1 t� (dy 1 t + h� � t ; ^ X t� idt) + 2 t (dy 2 t � d z }| { hg � t (!); X t� idt); (3:1) where 1 t =Ifh� � t ; ^ X t� i 6= 0g fdiag(� t )� � � t g ^ X t� h� � t ; ^ X t� i � ^ X t� 2 t = d z }| { diag(g � t (!))X t� � I d z }| { hg � t (!); X t� i ^ X t� ; (3:2) with I standing for the unit matrix of dimension (n � n) ,h�; �i denoting the scalar product, � � s := (� 11 s ; � � � ; � nn s ) ; g � s (!) := (g s (!; 1); � � � ; g s (!; n)): (3:3) and with diag(a) denoting the diagonal matrix with elememts a 1 ; � � � ; a n ; and � meaning transpose. Remark 3.2 : Notice that what Theorem 3.1. really does is generalizing and then combining results from [8; part III] and [2] as well as [6] (concerning 1 ) with 4 results from [10] (concerning 2 ). There is a priori no clear way to do this directly. The results of the theorem can be shown to follow from the theory of �ltration of general semimartingales [11, Th.10.1, Ch.4]. A direct proof is possible but, since it has only technical interest, it is moved to Appendix. The �lter equations (3.1), (3.2) contain terms of the form d z }| { diag(g � t (!))X t� = 0 @ d z }| { g i t (!)X i t� 1 A i=1;���;n = 0 @ d z }| { g i t (!)If� t� = ig 1 A i=1;���;n and d z }| { hg � t (!); X t� i = n X i=1 d z }| { g i t (!)If� t� = ig Analogously to (1.4), recalling also Remark 2.1., we obtain d z }| { g i t (!)If� t� = ig = X � � N t �1 0 g^ t � Y t 0 ; � � N t �1 0 ; � � N t = i � P � � � N t �1 0 j F Y t� ; (3:4) where � � N t �1 0 is any trajectory of the process � t from 0 to � N t �1 compatible with the observations of the process y 1 t = N t , namely having the switching points at the random time instants � k and keeping constant values between them. Generally speaking, the computation of an expression of the form P � � � N t �1 0 j F Y t� leads to a smoothing problem for the Markov process � t . We shall however be able to suggest a recursive procedure allowing the computation of these probabilities by means of a system of forward equations that are analogous to the �lter equations (3.1),(3.2). 4. ALGORITHM FOR THE COMPUTATION OF THE CONDITIONAL PROBABILITY OF A GIVEN TRAJECTORY OF THE HIDDEN MARKOV PROCESS As underlying basis for the algorithm, that will be described below, we give the following auxiliary result Proposition 4.1. Given a time instant t and a sequence i 0 ; � � � ; i N t �1 , let C(!) = Q N t �1 k=0 If� � k = i k g and F Y t := F X;Y � � N t W F Y t for � N t � t < � N t +1 . Then, for � N t � t < � N t +1 , we have that X t C := EfX t C j F Y t g is solution of the following �lter equation X t C = X � N t C + Z t � N t � � s X s Cds+ Z t � N t 1 s� Cd ~ y 1 s + Z t � N t 2 s Cd ~ y 2 s (4:1) where X � N t C =E n X � N t C j F Y � N t o = P n � � N t = i N t j � � N t �1 = i N t �1 ; � � � ; � � 0 = i 0 ; F Y � N t o i N t =1;���;n (4:2) 5 ~ y 1 t = y 1 t � y 1 � N t + Z t � N t h� � s ; X s� Cids; (4:3) ~ y 2 t = y 2 t � y 2 � N t � Z t � N t hg � s (!); X s Cids; (4:4) 1 t = Ifh� � t ; X t� Ci 6= 0g fdiag(� t )� � � t )gX t� C h� � t ; X t� Ci �X t� C; (4:5) 2 t = diag(g � t (!))X t� C � Ihg � t (!); X t� CiX t� C; (4:6) with hg � t (!); X t� Ci = n X i=1 g^ t � Y t 0 ; � � N t �1 0 ; � � N t = i � X i t� C; (4:7) diag(g � t (!))X t� C = � g^ t � Y t 0 ; � � N t �1 0 ; � � N t = i � X i t� C � i=1;���;n ; (4:8) where � � N t �1 0 represents the given sequence i 0 ; � � � ; i N t �1 and g^ t � Y t 0 ; � � N t �1 0 ; � � N t = i � = E n g t � !; � � N t = i � j F Y t� o : (4:9) Proof : The proof of this statement can be obtained in complete analogy to that of Theorem 3.1.(see Appendix), with a suitable adjustment of (2.1) to represent the process X t C on [� N t ; � N t +1 ) and taking also into account that, on one hand, by properties of the indicator function, C 2 = C, on the other that C is F Y � N t � measurable. To determine 1 t ; 2 t one obtains then an equation analogous to (A.9). The Proposition leads to a closed system of �lter equations for X t C on the generic interval [� N t ; � N t +1 ) , that can be solved provided one has the knowledge of the initial condition X � N t C as well as of g^ t . While the problem of determining g^ t will be discussed in the next section 5, the following result shows that, with the help of the system of �lter equations (4.1), one can also obtain the initial condition for the solution of the �lter equations for X t C on the following interval [� N t +1 ; � N t +2 ). Proposition 4.2. Under the same assumptions as for Proposition 4.1, the matrix of conditional transition probabilities of the Markov jump process � t at the moment � N t +1 is D � � � N t +1 ; X � � N t +1 C � = 8 > < > : I; if h� � � N t +1 ; X � � N t +1 Ci = 0 diag(� � N t +1 )�� � � N t +1 h� � � N t +1 ;X � � N t +1 Ci ; if h� � � N t +1 ; X � � N t +1 Ci 6= 0 : (4:10) Proof : The solution of the �lter equations (4.1) at time � N t +1 is X � N t +1 C = 1 � � N t +1 C +X � � N t +1 C = D � � � N t +1 ; X � � N t +1 C � X � � N t +1 C (4:11) 6 which de�nes implicitly D as the matrix of conditional transition probabilities. We show now that it is given by (4.10). To this e�ect notice that the elements of the vector X � � N t +1 C are (taking into account Remark 2.1) X i � � N t +1 C = P n � � N t = i; � � N t �1 = i N t �1 ; � � � ; � � 0 = i 0 j F Y � � N t +1 o = P n � � N t = i; � � N t �1 = i N t �1 ; � � � ; � � 0 = i 0 j F Y � N t +1 o : (4:12) On the other hand, for the components of the vector X j � N t +1 C we have X j � N t +1 C = P n � � N t +1 = j; � � N t �1 = i N t �1 ; � � � ; � � 0 = i 0 j F Y � N t +1 o = n X i=1 P n � � N t +1 = j j � � N t = i; � � N t �1 = i N t �1 ; � � � ; � � 0 = i 0 ; F Y � N t +1 o � P n � � N t = i; � � N t �1 = i N t �1 ; � � � ; � � 0 = i 0 j F Y � N t +1 o ; (4:13) where the term for i = j is zero w.P.1. From here, taking (4.11), (4.12) into account, we obtain the following representation for the elements of the matrix D D ji � � � N t +1 ; X � � N t +1 C � = P n � � N t +1 = j j � � N t = i; � � N t �1 = i N t �1 ; � � � ; � � 0 = i 0 ; F Y � N t +1 o = P n � � N t +1 = j j � � N t = i; � � N t �1 = i N t �1 ; � � � ; � � 0 = i 0 ; F Y � N t +1 o (4:14) since all information about the past history of the process � t up to the moment � � N t +1 is contained in the expression representing the conditioning. This accomplishes the proof of the Proposition. Since on the interval [� 0 ; � 1 ) the initial condition is given by the initial distribu- tion Pf� � 0 = ig i=1;���;n , with the use of Propositions 4.1 and 4.2 we can now derive a (forward) recursive �ltering procedure. We obtain in fact the following Algorithm for the solution of the general �ltering problem - First de�ne the set of random quantities: - on the interval [� 0 ; � 1 ) let C i 0 := If� � 0 = i 0 g; i 0 = 1; � � � ; n, - on the generic interval [� k ; � k+1 ) let C i k ;i k�1 ;���;i 0 := If� � k = i k ; � � � ; � � 0 = i 0 g; i k 6= i k�1 ; i k�1 6= i k�2 ; � � � ; i 0 = 1; � � � ; n. In this way on the k�th interval we obtain n(n� 1) k di�erent random quantities. - The algorithm now starts on the interval [� 0 ; � 1 ), where the elements of the vector X t C coincide with the random variables C i 0 , namely X i 0 t = C i 0 . - The solution of n �lter equations of the type (4.1) leads to the conditional probabilities EfX t C j F Y t g = Pf� � 0 = i 0 j F Y t g i 0 =1;���;n ; for t < � 1 ; 7 - at the time point t = � 1 , in accordance with Proposition 4.2, we obtain n(n�1) initial conditions for the solution of n(n� 1) �lter equations of type (4.1) for the conditional probabilities E � C i 1 i 0 j F Y t � i 0 ; i 1 = 1; � � � ; n i 1 6= i 0 � = E � X i 1 t C i 0 j F Y t � i 0 ; i 1 = 1; � � � ; n i 1 6= i 0 � - The algorithm then proceeds analogously in the obvious way over the various in- tervals [� k ; � k+1 ). In this way, recalling also Remark 2.1., we have obtained a recursive algorithm for the computation of the conditional probabilities Pf� t 0 j F Y t� gthat, by (3.4), are needed for the solution of the general �lter equations (3.1), (3.2). For this recursive algorithm we need at every time point t 2 [� N t ; � N t +1 ) to solve n(n� 1) N t �lter equations of type (4.1), and then to compute the coeÆcients in the equations (3.1), (3.2) by means of (3.4), using the current values of g^ t (Y t 0 ; � t 0 ) as well as of Pf� t 0 j F Y t� g. Remark 4.3 In the case n = 2 the system of �lter equations is
本文档为【miller1997kalman】,请使用软件OFFICE或WPS软件打开。作品中的文字与图均可以修改和编辑, 图片更改请在作品中右键图片并更换,文字修改请直接点击文字进行修改,也可以新增和删除文档中的内容。
该文档来自用户分享,如有侵权行为请发邮件ishare@vip.sina.com联系网站客服,我们会及时删除。
[版权声明] 本站所有资料为用户分享产生,若发现您的权利被侵害,请联系客服邮件isharekefu@iask.cn,我们尽快处理。
本作品所展示的图片、画像、字体、音乐的版权可能需版权方额外授权,请谨慎使用。
网站提供的党政主题相关内容(国旗、国徽、党徽..)目的在于配合国家政策宣传,仅限个人学习分享使用,禁止用于任何广告和商用目的。
下载需要: 免费 已有0 人下载
最新资料
资料动态
专题动态
is_876915
暂无简介~
格式:pdf
大小:197KB
软件:PDF阅读器
页数:0
分类:
上传时间:2014-01-08
浏览量:21