A Fitted-Q Algorithm for Budgeted MDPs
Document type :
Communication dans un congrès avec actes
Title :
A Fitted-Q Algorithm for Budgeted MDPs
Author(s) :
Carrara, Nicolas [Auteur]
Orange Labs [Lannion]
Sequential Learning [SEQUEL]
Laroche, Romain [Auteur]
Maluuba
Bouraoui, Jean-Léon [Auteur]
Orange Labs [Lannion]
Urvoy, Tanguy [Auteur]
Orange Labs [Lannion]
Pietquin, Olivier [Auteur]
Centre de Recherche en Informatique, Signal et Automatique de Lille - UMR 9189 [CRIStAL]
Orange Labs [Lannion]
Sequential Learning [SEQUEL]
Laroche, Romain [Auteur]
Maluuba
Bouraoui, Jean-Léon [Auteur]
Orange Labs [Lannion]
Urvoy, Tanguy [Auteur]
Orange Labs [Lannion]
Pietquin, Olivier [Auteur]
Centre de Recherche en Informatique, Signal et Automatique de Lille - UMR 9189 [CRIStAL]
Conference title :
EWRL 2018 - 14th European workshop on Reinforcement Learning
City :
Lille
Country :
France
Start date of the conference :
2018-10-01
Publication date :
2018
English keyword(s) :
Budgeted-MDP
Fitted-Q
Reinforcement Learning
Fitted-Q
Reinforcement Learning
HAL domain(s) :
Informatique [cs]/Intelligence artificielle [cs.AI]
English abstract : [en]
We address the problem of budgeted reinforcement learning, in continuous state-space, using a batch of transitions. To this extend, we introduce a novel algorithm called Budgeted Fitted-Q (BFTQ). Benchmarks show that BFTQ ...
Show more >We address the problem of budgeted reinforcement learning, in continuous state-space, using a batch of transitions. To this extend, we introduce a novel algorithm called Budgeted Fitted-Q (BFTQ). Benchmarks show that BFTQ performs as well as a regular Fitted-Q algorithm in a continuous 2-D world but also allows one to choose the right amount of budget that fits to a given task without the need of engineering the rewards. We believe that the general principles used to design BFTQ can be applied to extend others classical reinforcement learning algorithms for budgeted oriented applications.Show less >
Show more >We address the problem of budgeted reinforcement learning, in continuous state-space, using a batch of transitions. To this extend, we introduce a novel algorithm called Budgeted Fitted-Q (BFTQ). Benchmarks show that BFTQ performs as well as a regular Fitted-Q algorithm in a continuous 2-D world but also allows one to choose the right amount of budget that fits to a given task without the need of engineering the rewards. We believe that the general principles used to design BFTQ can be applied to extend others classical reinforcement learning algorithms for budgeted oriented applications.Show less >
Language :
Anglais
Peer reviewed article :
Oui
Audience :
Internationale
Popular science :
Non
Collections :
Source :
Files
- https://hal.archives-ouvertes.fr/hal-01928092/document
- Open access
- Access the document
- https://hal.archives-ouvertes.fr/hal-01928092/document
- Open access
- Access the document
- https://hal.archives-ouvertes.fr/hal-01928092/document
- Open access
- Access the document
- document
- Open access
- Access the document
- ewrl_14_2018_paper_67.pdf
- Open access
- Access the document
- document
- Open access
- Access the document
- ewrl_14_2018_paper_67.pdf
- Open access
- Access the document