IDEAS home Printed from https://ideas.repec.org/a/eee/spapps/v49y1994i1p159-177.html
   My bibliography  Save this article

Finite state Markov decision models with average reward criteria

Author

Listed:
  • Feinberg, Eugene A.
  • Park, Haechurl

Abstract

This paper deals with a discrete time Markov decision model with a finite state space, arbitrary action space, and bounded reward function under the average reward criteria. We consider four average reward criteria and prove the existence of persistently nearly optimal strategies in various classes of strategies for models with complete state information. We show that such strategies exist in any class of strategies satisfying the following condition: along any trajectory at different epochs the controller knows different information about the past. Though neither optimal nor stationary nearly optimal strategies may exist, we show that for some nonempty set of states the described nearly optimal strategies may be chosen either stationary or optimal.

Suggested Citation

  • Feinberg, Eugene A. & Park, Haechurl, 1994. "Finite state Markov decision models with average reward criteria," Stochastic Processes and their Applications, Elsevier, vol. 49(1), pages 159-177, January.
  • Handle: RePEc:eee:spapps:v:49:y:1994:i:1:p:159-177
    as

    Download full text from publisher

    File URL: http://www.sciencedirect.com/science/article/pii/0304-4149(94)90116-3
    Download Restriction: Full text for ScienceDirect subscribers only
    ---><---

    As the access to this document is restricted, you may want to search for a different version of it.

    Citations

    Citations are extracted by the CitEc Project, subscribe to its RSS feed for this item.
    as


    Cited by:

    1. Matthew Sobel, 2013. "Discounting axioms imply risk neutrality," Annals of Operations Research, Springer, vol. 208(1), pages 417-432, September.
    2. Rolando Cavazos-Cadena & Eugene A. Feinberg & Raúl Montes-de-Oca, 2000. "A Note on the Existence of Optimal Policies in Total Reward Dynamic Programs with Compact Action Sets," Mathematics of Operations Research, INFORMS, vol. 25(4), pages 657-666, November.

    Corrections

    All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:eee:spapps:v:49:y:1994:i:1:p:159-177. See general information about how to correct material in RePEc.

    If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.

    We have no bibliographic references for this item. You can help adding them by using this form .

    If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.

    For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: Catherine Liu (email available below). General contact details of provider: http://www.elsevier.com/wps/find/journaldescription.cws_home/505572/description#description .

    Please note that corrections may take a couple of weeks to filter through the various RePEc services.

    IDEAS is a RePEc service. RePEc uses bibliographic data supplied by the respective publishers.