IDEAS home Printed from https://ideas.repec.org/a/spr/mathme/v83y2016i2d10.1007_s00186-015-0524-y.html
   My bibliography  Save this article

On undiscounted semi-Markov decision processes with absorbing states

Author

Listed:
  • Prasenjit Mondal

    (Chandernagore Government College)

Abstract

Limiting ratio average (undiscounted) reward finite (state and action spaces) semi-Markov decision processes (SMDPs) with absorbing states are considered where all but one states are absorbing. We propose a realistic inspection model that suitably fits into the class of undiscounted SMDPs with absorbing states. Existence of an optimal semi-stationary policy (i.e., a semi-Markov policy independent of decision epoch counts) is proved. A linear programming algorithm is provided to compute such an optimal policy.

Suggested Citation

  • Prasenjit Mondal, 2016. "On undiscounted semi-Markov decision processes with absorbing states," Mathematical Methods of Operations Research, Springer;Gesellschaft für Operations Research (GOR);Nederlands Genootschap voor Besliskunde (NGB), vol. 83(2), pages 161-177, April.
  • Handle: RePEc:spr:mathme:v:83:y:2016:i:2:d:10.1007_s00186-015-0524-y
    DOI: 10.1007/s00186-015-0524-y
    as

    Download full text from publisher

    File URL: http://link.springer.com/10.1007/s00186-015-0524-y
    File Function: Abstract
    Download Restriction: Access to the full text of the articles in this series is restricted.

    File URL: https://libkey.io/10.1007/s00186-015-0524-y?utm_source=ideas
    LibKey link: if access is restricted and if your library uses this service, LibKey will redirect you to where you can use your library subscription to access this item
    ---><---

    As the access to this document is restricted, you may want to search for a different version of it.

    References listed on IDEAS

    as
    1. Chen, Dongyan & Trivedi, Kishor S., 2005. "Optimization for condition-based maintenance with semi-Markov decision process," Reliability Engineering and System Safety, Elsevier, vol. 90(1), pages 25-29.
    2. Tapas K. Das & Abhijit Gosavi & Sridhar Mahadevan & Nicholas Marchalleck, 1999. "Solving Semi-Markov Decision Problems Using Average Reward Reinforcement Learning," Management Science, INFORMS, vol. 45(4), pages 560-574, April.
    3. Cyrus Derman, 1962. "On Sequential Decisions and Markov Chains," Management Science, INFORMS, vol. 9(1), pages 16-24, October.
    4. Manfred Schäl, 1992. "On the Second Optimality Equation for Semi-Markov Decision Models," Mathematics of Operations Research, INFORMS, vol. 17(2), pages 470-486, May.
    5. Prasenjit Mondal & Sagnik Sinha, 2015. "Ordered Field Property for Semi-Markov Games when One Player Controls Transition Probabilities and Transition Times," International Game Theory Review (IGTR), World Scientific Publishing Co. Pte. Ltd., vol. 17(02), pages 1-26.
    6. L. Jianyong & Z. Xiaobo, 2004. "On Average Reward Semi-Markov Decision Processes with a General Multichain Structure," Mathematics of Operations Research, INFORMS, vol. 29(2), pages 339-352, May.
    7. J. Flesch & F. Thuijsman & O. J. Vrieze, 1996. "Recursive Repeated Games with Absorbing States," Mathematics of Operations Research, INFORMS, vol. 21(4), pages 1016-1022, November.
    8. P. J. Schweitzer & A. Federgruen, 1978. "The Functional Equations of Undiscounted Markov Renewal Programming," Mathematics of Operations Research, INFORMS, vol. 3(4), pages 308-321, November.
    9. Prasenjit Mondal, 2015. "Linear Programming and Zero-Sum Two-Person Undiscounted Semi-Markov Games," Asia-Pacific Journal of Operational Research (APJOR), World Scientific Publishing Co. Pte. Ltd., vol. 32(06), pages 1-20, December.
    Full references (including those not matched with items on IDEAS)

    Citations

    Citations are extracted by the CitEc Project, subscribe to its RSS feed for this item.
    as


    Cited by:

    1. Prasenjit Mondal, 2020. "Computing semi-stationary optimal policies for multichain semi-Markov decision processes," Annals of Operations Research, Springer, vol. 287(2), pages 843-865, April.
    2. Prasenjit Mondal, 2018. "Completely mixed strategies for single controller unichain semi-Markov games with undiscounted payoffs," Operational Research, Springer, vol. 18(2), pages 451-468, July.

    Most related items

    These are the items that most often cite the same works as this one and are cited by the same works as this one.
    1. Prasenjit Mondal, 2020. "Computing semi-stationary optimal policies for multichain semi-Markov decision processes," Annals of Operations Research, Springer, vol. 287(2), pages 843-865, April.
    2. L. Jianyong & Z. Xiaobo, 2004. "On Average Reward Semi-Markov Decision Processes with a General Multichain Structure," Mathematics of Operations Research, INFORMS, vol. 29(2), pages 339-352, May.
    3. Prasenjit Mondal, 2018. "Completely mixed strategies for single controller unichain semi-Markov games with undiscounted payoffs," Operational Research, Springer, vol. 18(2), pages 451-468, July.
    4. Prasenjit Mondal, 2015. "Linear Programming and Zero-Sum Two-Person Undiscounted Semi-Markov Games," Asia-Pacific Journal of Operational Research (APJOR), World Scientific Publishing Co. Pte. Ltd., vol. 32(06), pages 1-20, December.
    5. Andriotis, C.P. & Papakonstantinou, K.G., 2019. "Managing engineering systems with large state and action spaces through deep reinforcement learning," Reliability Engineering and System Safety, Elsevier, vol. 191(C).
    6. Oscar Volij & Casilda Lasso de la Vega, 2016. "The Value Of A Draw In Quasi-Binary Matches," Working Papers 1601, Ben-Gurion University of the Negev, Department of Economics.
    7. Oguzhan Alagoz & Lisa M. Maillart & Andrew J. Schaefer & Mark S. Roberts, 2007. "Determining the Acceptance of Cadaveric Livers Using an Implicit Model of the Waiting List," Operations Research, INFORMS, vol. 55(1), pages 24-36, February.
    8. Lodewijk Kallenberg, 2013. "Derman’s book as inspiration: some results on LP for MDPs," Annals of Operations Research, Springer, vol. 208(1), pages 63-94, September.
    9. Flesch, J. & Thuijsman, F. & Vrieze, O.J., 2007. "Stochastic games with additive transitions," European Journal of Operational Research, Elsevier, vol. 179(2), pages 483-497, June.
    10. Eilon Solan & Nicholas Vieille, 2001. "Quitting Games - An Example," Discussion Papers 1314, Northwestern University, Center for Mathematical Studies in Economics and Management Science.
    11. Ohno, Katsuhisa & Boh, Toshitaka & Nakade, Koichi & Tamura, Takayoshi, 2016. "New approximate dynamic programming algorithms for large-scale undiscounted Markov decision processes and their application to optimize a production and distribution system," European Journal of Operational Research, Elsevier, vol. 249(1), pages 22-31.
    12. Alexander Zadorojniy & Guy Even & Adam Shwartz, 2009. "A Strongly Polynomial Algorithm for Controlled Queues," Mathematics of Operations Research, INFORMS, vol. 34(4), pages 992-1007, November.
    13. David T. Abdul‐Malak & Jeffrey P. Kharoufeh & Lisa M. Maillart, 2019. "Maintaining systems with heterogeneous spare parts," Naval Research Logistics (NRL), John Wiley & Sons, vol. 66(6), pages 485-501, September.
    14. Shmaya, Eran & Solan, Eilon & Vieille, Nicolas, 2003. "An application of Ramsey theorem to stopping games," Games and Economic Behavior, Elsevier, vol. 42(2), pages 300-306, February.
    15. Casilda Lasso de la Vega & Oscar Volij, 2020. "The value of a draw," Economic Theory, Springer;Society for the Advancement of Economic Theory (SAET), vol. 70(4), pages 1023-1044, November.
    16. VIEILLE, Nicolas & SOLAN, Eilon, 2001. "Stopping games: recent results," HEC Research Papers Series 744, HEC Paris.
    17. Zhao, Yunfei & Huang, Linan & Smidts, Carol & Zhu, Quanyan, 2020. "Finite-horizon semi-Markov game for time-sensitive attack response and probabilistic risk assessment in nuclear power plants," Reliability Engineering and System Safety, Elsevier, vol. 201(C).
    18. Zhou, Zhi-Jie & Hu, Chang-Hua & Xu, Dong-Ling & Chen, Mao-Yin & Zhou, Dong-Hua, 2010. "A model for real-time failure prognosis based on hidden Markov model and belief rule base," European Journal of Operational Research, Elsevier, vol. 207(1), pages 269-283, November.
    19. Panagiotidou, S. & Tagaras, G., 2012. "Optimal integrated process control and maintenance under general deterioration," Reliability Engineering and System Safety, Elsevier, vol. 104(C), pages 58-70.
    20. Yang, Hongbing & Li, Wenchao & Wang, Bin, 2021. "Joint optimization of preventive maintenance and production scheduling for multi-state production systems based on reinforcement learning," Reliability Engineering and System Safety, Elsevier, vol. 214(C).

    Corrections

    All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:spr:mathme:v:83:y:2016:i:2:d:10.1007_s00186-015-0524-y. See general information about how to correct material in RePEc.

    If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.

    If CitEc recognized a bibliographic reference but did not link an item in RePEc to it, you can help with this form .

    If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.

    For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: Sonal Shukla or Springer Nature Abstracting and Indexing (email available below). General contact details of provider: http://www.springer.com .

    Please note that corrections may take a couple of weeks to filter through the various RePEc services.

    IDEAS is a RePEc service. RePEc uses bibliographic data supplied by the respective publishers.