登入選單
返回Google圖書搜尋
Markov Decision Problems with Expected Utility Criteria
註釋Finite state and action Markov decision problems with expected utility criteria are analyzed. A Markov decision chain (or sequential decision process) is defined in the usual manner. But instead of seeking to maximize the expected sum (or product) of rewards, the objective is maximization of the expectation of some cardinal utility function defined on the sequence of rewards.