JOURNAL ARTICLE

Online Markov Decision Processes Under Bandit Feedback

Gergely NeuAndrás GyörgyCsaba SzepesváriAndrás Antos

Year: 2014 Journal:   IEEE Transactions on Automatic Control Vol: 59 (3)Pages: 676-691   Publisher: Institute of Electrical and Electronics Engineers

Abstract

International audience

Keywords:
Regret Hindsight bias Markov decision process Markov chain State (computer science) Computer science Markov process Function (biology) Mathematical economics Discrete mathematics Combinatorics Mathematical optimization Mathematics Artificial intelligence Algorithm Machine learning Statistics Psychology

Metrics

102
Cited By
4.88
FWCI (Field Weighted Citation Impact)
34
Refs
0.95
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Citation History

Topics

Advanced Bandit Algorithms Research
Social Sciences →  Decision Sciences →  Management Science and Operations Research
Reinforcement Learning in Robotics
Physical Sciences →  Computer Science →  Artificial Intelligence
Optimization and Search Problems
Physical Sciences →  Computer Science →  Computer Networks and Communications

Related Documents

JOURNAL ARTICLE

Online Markov Decision Processes

Eyal Even-DarSham M. KakadeYishay Mansour

Journal:   Mathematics of Operations Research Year: 2009 Vol: 34 (3)Pages: 726-736
JOURNAL ARTICLE

An ϵ-Greedy Multiarmed Bandit Approach to Markov Decision Processes

Isa MuqattashJiaqiao Hu

Journal:   Stats Year: 2023 Vol: 6 (1)Pages: 99-112
BOOK-CHAPTER

PAC Bounds for Multi-armed Bandit and Markov Decision Processes

Eyal Even-DarShie MannorYishay Mansour

Lecture notes in computer science Year: 2002 Pages: 255-270
JOURNAL ARTICLE

Markov decision processes under ambiguity

Bäuerle, NicoleRieder, Ulrich

Journal:   KITopen Year: 2020
JOURNAL ARTICLE

Markov decision processes under ambiguity

Nicole BäuerleUlrich Rieder

Journal:   Banach Center Publications Year: 2020 Vol: 122 Pages: 25-39
© 2026 ScienceGate Book Chapters — All rights reserved.