Abstract
The present paper is the sequel to a previous paper by the author. The main purpose of the present paper is to justify the claim made curlier that the approach to stochastic optimal control previously employed can be extended to the case of noisy observations of the state. The partial differential equation, for the conditional probability density of the present state given the past history of the noisy observations, is discussed. Using this equation, it is shown that the stochastic optimal control problem can be viewed as a problem in the theory of control of distributed parameter systems. Dynamic programming is then applied to this distributed parameter problem, to obtain a stochastic Hamilton-Jacobi equation in function space.

This publication has 4 references indexed in Scilit: