IDEAS home Printed from https://ideas.repec.org/a/nat/nature/v521y2015i7553d10.1038_nature14540.html
   My bibliography  Save this article

Reinforcement learning improves behaviour from evaluative feedback

Author

Listed:
  • Michael L. Littman

    (Brown University)

Abstract

Reinforcement learning is a branch of machine learning concerned with using experience gained through interacting with the world and evaluative feedback to improve a system's ability to make behavioural decisions. It has been called the artificial intelligence problem in a microcosm because learning algorithms must act autonomously to perform well and achieve their goals. Partly driven by the increasing availability of rich data, recent years have seen exciting advances in the theory and practice of reinforcement learning, including developments in fundamental technical areas such as generalization, planning, exploration and empirical methodology, leading to increasing applicability to real-life problems.

Suggested Citation

  • Michael L. Littman, 2015. "Reinforcement learning improves behaviour from evaluative feedback," Nature, Nature, vol. 521(7553), pages 445-451, May.
  • Handle: RePEc:nat:nature:v:521:y:2015:i:7553:d:10.1038_nature14540
    DOI: 10.1038/nature14540
    as

    Download full text from publisher

    File URL: https://www.nature.com/articles/nature14540
    File Function: Abstract
    Download Restriction: Access to the full text of the articles in this series is restricted.

    File URL: https://libkey.io/10.1038/nature14540?utm_source=ideas
    LibKey link: if access is restricted and if your library uses this service, LibKey will redirect you to where you can use your library subscription to access this item
    ---><---

    As the access to this document is restricted, you may want to search for a different version of it.

    Citations

    Citations are extracted by the CitEc Project, subscribe to its RSS feed for this item.
    as


    Cited by:

    1. Wenjing Guo & Cairong Yan & Ting Lu, 2019. "Optimizing the lifetime of wireless sensor networks via reinforcement-learning-based routing," International Journal of Distributed Sensor Networks, , vol. 15(2), pages 15501477198, February.
    2. Yuhong Wang & Lei Chen & Hong Zhou & Xu Zhou & Zongsheng Zheng & Qi Zeng & Li Jiang & Liang Lu, 2021. "Flexible Transmission Network Expansion Planning Based on DQN Algorithm," Energies, MDPI, vol. 14(7), pages 1-21, April.
    3. Gohar Gholamibozanjani & Mohammed Farid, 2021. "A Critical Review on the Control Strategies Applied to PCM-Enhanced Buildings," Energies, MDPI, vol. 14(7), pages 1-39, March.
    4. Vijendra Kumar & Hazi Md. Azamathulla & Kul Vaibhav Sharma & Darshan J. Mehta & Kiran Tota Maharaj, 2023. "The State of the Art in Deep Learning Applications, Challenges, and Future Prospects: A Comprehensive Review of Flood Forecasting and Management," Sustainability, MDPI, vol. 15(13), pages 1-33, July.
    5. Liang, Xuedong & Luo, Peng & Li, Xiaoyan & Wang, Xia & Shu, Lingli, 2023. "Crude oil price prediction using deep reinforcement learning," Resources Policy, Elsevier, vol. 81(C).
    6. Adrian Millea, 2021. "Deep Reinforcement Learning for Trading—A Critical Survey," Data, MDPI, vol. 6(11), pages 1-25, November.
    7. Li, Yanbin & Wang, Jiani & Wang, Weiye & Liu, Chang & Li, Yun, 2023. "Dynamic pricing based electric vehicle charging station location strategy using reinforcement learning," Energy, Elsevier, vol. 281(C).
    8. Chuhan Wu & Fangzhao Wu & Tao Qi & Wei-Qiang Zhang & Xing Xie & Yongfeng Huang, 2022. "Removing AI’s sentiment manipulation of personalized news delivery," Palgrave Communications, Palgrave Macmillan, vol. 9(1), pages 1-9, December.

    More about this item

    Statistics

    Access and download statistics

    Corrections

    All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:nat:nature:v:521:y:2015:i:7553:d:10.1038_nature14540. See general information about how to correct material in RePEc.

    If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.

    We have no bibliographic references for this item. You can help adding them by using this form .

    If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.

    For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: Sonal Shukla or Springer Nature Abstracting and Indexing (email available below). General contact details of provider: http://www.nature.com .

    Please note that corrections may take a couple of weeks to filter through the various RePEc services.

    IDEAS is a RePEc service. RePEc uses bibliographic data supplied by the respective publishers.