IDEAS home Printed from https://ideas.repec.org/a/inm/ormnsc/v22y1975i2p148-157.html
   My bibliography  Save this article

On the Optimality of Structured Policies in Countable Stage Decision Processes

Author

Listed:
  • Evan L. Porteus

    (Stanford University)

Abstract

Multi-stage decision processes are considered, in notation which is an outgrowth of that introduced by Denardo [Denardo, E. 1967. Contraction mappings in the theory underlying dynamic programming. SIAM Rev. 9 165-177.]. Certain Markov decision processes, stochastic games, and risk-sensitive Markov decision processes can be formulated in this notation. We identify conditions sufficient to prove that, in infinite horizon nonstationary processes, the optimal infinite horizon (present) value exists, is uniquely defined, is what is called "structured," and can be found by solving Bellman's optimality equations: \epsilon -optimal strategies exist: an optimal strategy can be found by applying Bellman's optimality criterion; and a specially identified kind of policy, called a "structured" policy is optimal in each stage. A link is thus drawn between (i) studies such as those of Blackwell [Blackwell, D. 1965. Discounted dynamic programming. Ann. Math. Stat. 36 226-235.] and Strauch [Strauch, R. 1966. Negative dynamic programming. Ann. Math. Stat. 37 871-890.], where general policies for general processes are considered, and (ii) other studies, such as those of Scarf [Scarf, H. 1963. The optimality of (S, s) policies in the dynamic inventory problem. H. Scarf, D. Gilford, M. Shelly, eds. Mathematical Methods in the Social Sciences . Stanford University Press, Stanford.] and Derman [Derman, C. 1963. On optimal replacement rules when changes of state are Markovian. R. Bellman, ed. Mathematical Optimization Techniques. University of California Press. Berkeley.] where structured policies for special processes are considered. Those familiar with dynamic programming models (e.g., inventory, queueing optimization, replacement, optimal stopping) will be well acquainted with the use of what we call structured policies and value functions. The infinite stage results are built on finite stage results. Results for the stationary infinite horizon case are also included. For an application, we provide conditions sufficient to prove that an optimal stationary strategy exists in a discounted stationary risk sensitive Markov decision process with constant risk aversion. In Porteus [Porteus, E. On the optimality of structured policies in countable stage decision processes. Research Paper No. 141, Graduate School of Business, Stanford University, 71 pp., 1973, 1974, unabridged version of present paper.], of which this is a condensation, we also (i) show how known conditions under which a Borel measurable policy is optimal in an infinite horizon, nonstationary Markov decision process, fit into our framework, and (ii) provide conditions under which a generalized (s, S) policy [Porteus, E. 1971. On the optimality of generalized (s, S) policies. Management Sci. 17 411-426.] is optimal in an infinite horizon nonstationary inventory process.

Suggested Citation

  • Evan L. Porteus, 1975. "On the Optimality of Structured Policies in Countable Stage Decision Processes," Management Science, INFORMS, vol. 22(2), pages 148-157, October.
  • Handle: RePEc:inm:ormnsc:v:22:y:1975:i:2:p:148-157
    DOI: 10.1287/mnsc.22.2.148
    as

    Download full text from publisher

    File URL: http://dx.doi.org/10.1287/mnsc.22.2.148
    Download Restriction: no

    File URL: https://libkey.io/10.1287/mnsc.22.2.148?utm_source=ideas
    LibKey link: if access is restricted and if your library uses this service, LibKey will redirect you to where you can use your library subscription to access this item
    ---><---

    Citations

    Citations are extracted by the CitEc Project, subscribe to its RSS feed for this item.
    as


    Cited by:

    1. Lucy Gongtao Chen & Daniel Zhuoyu Long & Melvyn Sim, 2015. "On Dynamic Decision Making to Meet Consumption Targets," Operations Research, INFORMS, vol. 63(5), pages 1117-1130, October.
    2. Monahan, George E. & Sobel, Matthew J., 1997. "Risk-Sensitive Dynamic Market Share Attraction Games," Games and Economic Behavior, Elsevier, vol. 20(2), pages 149-160, August.
    3. Zeynep Erkin & Matthew D. Bailey & Lisa M. Maillart & Andrew J. Schaefer & Mark S. Roberts, 2010. "Eliciting Patients' Revealed Preferences: An Inverse Markov Decision Process Approach," Decision Analysis, INFORMS, vol. 7(4), pages 358-365, December.
    4. Ketzenberg, Michael & Gaukler, Gary & Salin, Victoria, 2018. "Expiration dates and order quantities for perishables," European Journal of Operational Research, Elsevier, vol. 266(2), pages 569-584.
    5. Robert A. Shumsky & Fuqiang Zhang, 2009. "Dynamic Capacity Management with Substitution," Operations Research, INFORMS, vol. 57(3), pages 671-684, June.
    6. Eberly, Janice C. & Van Mieghem, Jan A., 1997. "Multi-factor Dynamic Investment under Uncertainty," Journal of Economic Theory, Elsevier, vol. 75(2), pages 345-387, August.
    7. Sen Lin & Bo Li & Antonio Arreola-Risa & Yiwei Huang, 2023. "Optimizing a single-product production-inventory system under constant absolute risk aversion," TOP: An Official Journal of the Spanish Society of Statistics and Operations Research, Springer;Sociedad de Estadística e Investigación Operativa, vol. 31(3), pages 510-537, October.
    8. Miehling, Erik & Teneketzis, Demosthenis, 2020. "Monotonicity properties for two-action partially observable Markov decision processes on partially ordered spaces," European Journal of Operational Research, Elsevier, vol. 282(3), pages 936-944.
    9. Sakine Batun & Andrew J. Schaefer & Atul Bhandari & Mark S. Roberts, 2018. "Optimal Liver Acceptance for Risk-Sensitive Patients," Service Science, INFORMS, vol. 10(3), pages 320-333, September.
    10. Martin A. Lariviere & Evan L. Porteus, 1999. "Stalking Information: Bayesian Inventory Management with Unobserved Lost Sales," Management Science, INFORMS, vol. 45(3), pages 346-363, March.

    More about this item

    Statistics

    Access and download statistics

    Corrections

    All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:inm:ormnsc:v:22:y:1975:i:2:p:148-157. See general information about how to correct material in RePEc.

    If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.

    We have no bibliographic references for this item. You can help adding them by using this form .

    If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.

    For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: Chris Asher (email available below). General contact details of provider: https://edirc.repec.org/data/inforea.html .

    Please note that corrections may take a couple of weeks to filter through the various RePEc services.

    IDEAS is a RePEc service. RePEc uses bibliographic data supplied by the respective publishers.