IDEAS home Printed from https://ideas.repec.org/a/eee/oprepe/v8y2021ics2214716021000221.html
   My bibliography  Save this article

A review of approximate dynamic programming applications within military operations research

Author

Listed:
  • Rempel, M.
  • Cai, J.

Abstract

Sequences of decisions that occur under uncertainty arise in a variety of settings, including transportation, communication networks, finance, defence, etc. The classic approach to find an optimal decision policy for a sequential decision problem is dynamic programming; however its usefulness is limited due to the curse of dimensionality and the curse of modelling, and thus many real-world applications require an alternative approach. Within operations research, over the last 25 years the use of Approximate Dynamic Programming (ADP), known as reinforcement learning in many disciplines, to solve these types of problems has increased in popularity. These efforts have resulted in the successful deployment of ADP-generated decision policies for driver scheduling in the trucking industry, locomotive planning and management, and managing high-value spare parts in manufacturing. In this article we present the first review of applications of ADP within a defence context, specifically focusing on those which provide decision support to military or civilian leadership. This article’s main contributions are twofold. First, we review 18 decision support applications, spanning the spectrum of force development, generation, and employment, that use an ADP-based strategy and for each highlight how its ADP algorithm was designed, evaluated, and the results achieved. Second, based on the trends and gaps identified we discuss five topics relevant to applying ADP to decision support problems within defence: the classes of problems studied; best practices to evaluate ADP-generated policies; advantages of designing policies that are incremental versus complete overhauls when compared to currently practiced policies; the robustness of policies as scenarios change, such as a shift from high to low intensity conflict; and sequential decision problems not yet studied within defence that may benefit from ADP.

Suggested Citation

  • Rempel, M. & Cai, J., 2021. "A review of approximate dynamic programming applications within military operations research," Operations Research Perspectives, Elsevier, vol. 8(C).
  • Handle: RePEc:eee:oprepe:v:8:y:2021:i:c:s2214716021000221
    DOI: 10.1016/j.orp.2021.100204
    as

    Download full text from publisher

    File URL: http://www.sciencedirect.com/science/article/pii/S2214716021000221
    Download Restriction: Full text for ScienceDirect subscribers only

    File URL: https://libkey.io/10.1016/j.orp.2021.100204?utm_source=ideas
    LibKey link: if access is restricted and if your library uses this service, LibKey will redirect you to where you can use your library subscription to access this item
    ---><---

    As the access to this document is restricted, you may want to search for a different version of it.

    References listed on IDEAS

    as
    1. Hugo P. Simão & Jeff Day & Abraham P. George & Ted Gifford & John Nienow & Warren B. Powell, 2009. "An Approximate Dynamic Programming Algorithm for Large-Scale Fleet Management: A Case Application," Transportation Science, INFORMS, vol. 43(2), pages 178-197, May.
    2. Warren B. Powell & Belgacem Bouzaiene-Ayari & Coleman Lawrence & Clark Cheng & Sourav Das & Ricardo Fiorillo, 2014. "Locomotive Planning at Norfolk Southern: An Optimizing Simulator Using Approximate Dynamic Programming," Interfaces, INFORMS, vol. 44(6), pages 567-578, December.
    3. Eyal Pecht & Asher Tishler & Nir Weingold, 2013. "ON THE CHOICE OF MULTI-TASK R&D DEFENSE PROJECTS: A CASE STUDY OF The ISRAELI MISSILE DEFENSE SYSTEM," Defence and Peace Economics, Taylor & Francis Journals, vol. 24(5), pages 429-448, October.
    4. Kyohong Shin & Taesik Lee, 2020. "Emergency medical service resource allocation in a mass casualty incident by integrating patient prioritization and hospital selection problems," IISE Transactions, Taylor & Francis Journals, vol. 52(10), pages 1141-1155, October.
    5. Robbins, Matthew J. & Jenkins, Phillip R. & Bastian, Nathaniel D. & Lunday, Brian J., 2020. "Approximate dynamic programming for the aeromedical evacuation dispatching problem: Value function approximation utilizing multiple level aggregation," Omega, Elsevier, vol. 91(C).
    6. Jenkins, Phillip R. & Robbins, Matthew J. & Lunday, Brian J., 2021. "Approximate dynamic programming for the military aeromedical evacuation dispatching, preemption-rerouting, and redeployment problem," European Journal of Operational Research, Elsevier, vol. 290(1), pages 132-143.
    7. Phillip R. Jenkins & Matthew J. Robbins & Brian J. Lunday, 2021. "Approximate Dynamic Programming for Military Medical Evacuation Dispatching Policies," INFORMS Journal on Computing, INFORMS, vol. 33(1), pages 2-26, January.
    8. Warren B. Powell & Arun Marar & Jack Gelfand & Steve Bowers, 2002. "Implementing Real-Time Optimization Models: A Case Application From The Motor Carrier Industry," Operations Research, INFORMS, vol. 50(4), pages 571-581, August.
    9. Warren B. Powell, 2009. "What you should know about approximate dynamic programming," Naval Research Logistics (NRL), John Wiley & Sons, vol. 56(3), pages 239-249, April.
    10. Davis, Michael T. & Robbins, Matthew J. & Lunday, Brian J., 2017. "Approximate dynamic programming for missile defense interceptor fire control," European Journal of Operational Research, Elsevier, vol. 259(3), pages 873-886.
    11. Warren B. Powell, 2016. "Perspectives of approximate dynamic programming," Annals of Operations Research, Springer, vol. 241(1), pages 319-356, June.
    12. Richard Bellman, 1957. "On a Dynamic Programming Approach to the Caterer Problem--I," Management Science, INFORMS, vol. 3(3), pages 270-278, April.
    13. Martijn R. K. Mes & Arturo Pérez Rivera, 2017. "Approximate Dynamic Programming by Practical Examples," International Series in Operations Research & Management Science, in: Richard J. Boucherie & Nico M. van Dijk (ed.), Markov Decision Processes in Practice, chapter 0, pages 63-101, Springer.
    14. Hugo P. Simão & Abraham George & Warren B. Powell & Ted Gifford & John Nienow & Jeff Day, 2010. "Approximate Dynamic Programming Captures Fleet Operations for Schneider National," Interfaces, INFORMS, vol. 40(5), pages 342-352, October.
    15. Gregory A. Godfrey & Warren B. Powell, 2002. "An Adaptive Dynamic Programming Algorithm for Dynamic Fleet Management, II: Multiperiod Travel Times," Transportation Science, INFORMS, vol. 36(1), pages 40-54, February.
    16. Rebekah S. McKenna & Matthew J. Robbins & Brian J. Lunday & Ian M. McCormack, 2020. "Approximate dynamic programming for the military inventory routing problem," Annals of Operations Research, Springer, vol. 288(1), pages 391-416, May.
    17. Warren H. Hausman, 1969. "Sequential Decision Problems: A Model to Exploit Existing Forecasters," Management Science, INFORMS, vol. 16(2), pages 93-111, October.
    18. Gregory A. Godfrey & Warren B. Powell, 2002. "An Adaptive Dynamic Programming Algorithm for Dynamic Fleet Management, I: Single Period Travel Times," Transportation Science, INFORMS, vol. 36(1), pages 21-39, February.
    19. Gerald G. Brown & Robert F. Dell & Alexandra M. Newman, 2004. "Optimizing Military Capital Planning," Interfaces, INFORMS, vol. 34(6), pages 415-425, December.
    20. Walker, Warren E. & Rahman, S. Adnan & Cave, Jonathan, 2001. "Adaptive policies, policy analysis, and policy-making," European Journal of Operational Research, Elsevier, vol. 128(2), pages 282-289, January.
    21. Powell, Warren B., 2019. "A unified framework for stochastic optimization," European Journal of Operational Research, Elsevier, vol. 275(3), pages 795-821.
    22. Rettke, Aaron J. & Robbins, Matthew J. & Lunday, Brian J., 2016. "Approximate dynamic programming for the dispatch of military medical evacuation assets," European Journal of Operational Research, Elsevier, vol. 254(3), pages 824-839.
    23. Warren Powell & Andrzej Ruszczyński & Huseyin Topaloglu, 2004. "Learning Algorithms for Separable Approximations of Discrete Stochastic Optimization Problems," Mathematics of Operations Research, INFORMS, vol. 29(4), pages 814-836, November.
    24. Stasko, Timon H. & Oliver Gao, H., 2012. "Developing green fleet management strategies: Repair/retrofit/replacement decisions under environmental regulation," Transportation Research Part A: Policy and Practice, Elsevier, vol. 46(8), pages 1216-1226.
    25. Evan E. Anderson & Yu‐Min Chen, 1988. "A decision support system for the procurement of military equipment," Naval Research Logistics (NRL), John Wiley & Sons, vol. 35(4), pages 619-632, August.
    26. Daniel R. Jiang & Warren B. Powell, 2018. "Risk-Averse Approximate Dynamic Programming with Quantile-Based Risk Measures," Mathematics of Operations Research, INFORMS, vol. 43(2), pages 554-579, May.
    27. Mark Zais & Dan Zhang, 2016. "A Markov chain model of military personnel dynamics," International Journal of Production Research, Taylor & Francis Journals, vol. 54(6), pages 1863-1885, March.
    Full references (including those not matched with items on IDEAS)

    Citations

    Citations are extracted by the CitEc Project, subscribe to its RSS feed for this item.
    as


    Cited by:

    1. José Javier Galán & Ramón Alberto Carrasco & Antonio LaTorre, 2022. "Military Applications of Machine Learning: A Bibliometric Perspective," Mathematics, MDPI, vol. 10(9), pages 1-27, April.

    Most related items

    These are the items that most often cite the same works as this one and are cited by the same works as this one.
    1. Liles, Joseph M. & Robbins, Matthew J. & Lunday, Brian J., 2023. "Improving defensive air battle management by solving a stochastic dynamic assignment problem via approximate dynamic programming," European Journal of Operational Research, Elsevier, vol. 305(3), pages 1435-1449.
    2. Heydar, Mojtaba & Mardaneh, Elham & Loxton, Ryan, 2022. "Approximate dynamic programming for an energy-efficient parallel machine scheduling problem," European Journal of Operational Research, Elsevier, vol. 302(1), pages 363-380.
    3. Michael F. Gorman & John-Paul Clarke & Amir Hossein Gharehgozli & Michael Hewitt & René de Koster & Debjit Roy, 2014. "State of the Practice: A Review of the Application of OR/MS in Freight Transportation," Interfaces, INFORMS, vol. 44(6), pages 535-554, December.
    4. Antoine Sauré & Jonathan Patrick & Martin L. Puterman, 2015. "Simulation-Based Approximate Policy Iteration with Generalized Logistic Functions," INFORMS Journal on Computing, INFORMS, vol. 27(3), pages 579-595, August.
    5. Warren B. Powell & Abraham George & Hugo Simão & Warren Scott & Alan Lamont & Jeffrey Stewart, 2012. "SMART: A Stochastic Multiscale Model for the Analysis of Energy Resources, Technology, and Policy," INFORMS Journal on Computing, INFORMS, vol. 24(4), pages 665-682, November.
    6. Powell, Warren B., 2019. "A unified framework for stochastic optimization," European Journal of Operational Research, Elsevier, vol. 275(3), pages 795-821.
    7. Warren B. Powell, 2016. "Perspectives of approximate dynamic programming," Annals of Operations Research, Springer, vol. 241(1), pages 319-356, June.
    8. Voelkel, Michael A. & Sachs, Anna-Lena & Thonemann, Ulrich W., 2020. "An aggregation-based approximate dynamic programming approach for the periodic review model with random yield," European Journal of Operational Research, Elsevier, vol. 281(2), pages 286-298.
    9. Mercedes Esteban-Bravo & Jose M. Vidal-Sanz & Gökhan Yildirim, 2014. "Valuing Customer Portfolios with Endogenous Mass and Direct Marketing Interventions Using a Stochastic Dynamic Programming Decomposition," Marketing Science, INFORMS, vol. 33(5), pages 621-640, September.
    10. Zolfagharinia, Hossein & Haughton, Michael, 2018. "The importance of considering non-linear layover and delay costs for local truckers," Transportation Research Part E: Logistics and Transportation Review, Elsevier, vol. 109(C), pages 331-355.
    11. Zhou, Shaorui & Zhang, Hui & Shi, Ning & Xu, Zhou & Wang, Fan, 2020. "A new convergent hybrid learning algorithm for two-stage stochastic programs," European Journal of Operational Research, Elsevier, vol. 283(1), pages 33-46.
    12. Baris Yildiz & Martin Savelsbergh, 2019. "Provably High-Quality Solutions for the Meal Delivery Routing Problem," Transportation Science, INFORMS, vol. 53(5), pages 1372-1388, September.
    13. Schmid, Verena, 2012. "Solving the dynamic ambulance relocation and dispatching problem using approximate dynamic programming," European Journal of Operational Research, Elsevier, vol. 219(3), pages 611-621.
    14. Meissner, Joern & Senicheva, Olga V., 2018. "Approximate dynamic programming for lateral transshipment problems in multi-location inventory systems," European Journal of Operational Research, Elsevier, vol. 265(1), pages 49-64.
    15. Ulmer, Marlin W. & Thomas, Barrett W., 2020. "Meso-parametric value function approximation for dynamic customer acceptances in delivery routing," European Journal of Operational Research, Elsevier, vol. 285(1), pages 183-195.
    16. Michael F. Gorman & John-Paul Clarke & René Koster & Michael Hewitt & Debjit Roy & Mei Zhang, 2023. "Emerging practices and research issues for big data analytics in freight transportation," Maritime Economics & Logistics, Palgrave Macmillan;International Association of Maritime Economists (IAME), vol. 25(1), pages 28-60, March.
    17. Dong‐Ping Song & Jonathan Carter, 2008. "Optimal empty vehicle redistribution for hub‐and‐spoke transportation systems," Naval Research Logistics (NRL), John Wiley & Sons, vol. 55(2), pages 156-171, March.
    18. Jenkins, Phillip R. & Robbins, Matthew J. & Lunday, Brian J., 2021. "Approximate dynamic programming for the military aeromedical evacuation dispatching, preemption-rerouting, and redeployment problem," European Journal of Operational Research, Elsevier, vol. 290(1), pages 132-143.
    19. Alexandre Forel & Martin Grunow, 2023. "Dynamic stochastic lot sizing with forecast evolution in rolling‐horizon planning," Production and Operations Management, Production and Operations Management Society, vol. 32(2), pages 449-468, February.
    20. Yan, Pengyu & Yu, Kaize & Chao, Xiuli & Chen, Zhibin, 2023. "An online reinforcement learning approach to charging and order-dispatching optimization for an e-hailing electric vehicle fleet," European Journal of Operational Research, Elsevier, vol. 310(3), pages 1218-1233.

    Corrections

    All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:eee:oprepe:v:8:y:2021:i:c:s2214716021000221. See general information about how to correct material in RePEc.

    If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.

    If CitEc recognized a bibliographic reference but did not link an item in RePEc to it, you can help with this form .

    If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.

    For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: Catherine Liu (email available below). General contact details of provider: http://www.journals.elsevier.com/operations-research-perspectives .

    Please note that corrections may take a couple of weeks to filter through the various RePEc services.

    IDEAS is a RePEc service. RePEc uses bibliographic data supplied by the respective publishers.