Difference of Convex Functions Programming ...
Document type :
Communication dans un congrès avec actes
Title :
Difference of Convex Functions Programming for Reinforcement Learning
Author(s) :
Piot, Bilal [Auteur]
Sequential Learning [SEQUEL]
IMS : Information, Multimodalité & Signal
Geist, Matthieu [Auteur]
IMS : Information, Multimodalité & Signal
Pietquin, Olivier [Auteur]
Sequential Learning [SEQUEL]
Laboratoire d'Informatique Fondamentale de Lille [LIFL]
Institut universitaire de France [IUF]

Sequential Learning [SEQUEL]
IMS : Information, Multimodalité & Signal
Geist, Matthieu [Auteur]
IMS : Information, Multimodalité & Signal
Pietquin, Olivier [Auteur]
Sequential Learning [SEQUEL]
Laboratoire d'Informatique Fondamentale de Lille [LIFL]
Institut universitaire de France [IUF]
Conference title :
Advances in Neural Information Processing Systems (NIPS 2014)
City :
Montreal
Country :
Canada
Start date of the conference :
2014-12
HAL domain(s) :
Informatique [cs]
Sciences de l'ingénieur [physics]
Sciences de l'ingénieur [physics]
English abstract : [en]
Large Markov Decision Processes are usually solved using Approximate Dy-namic Programming methods such as Approximate Value Iteration or Ap-proximate Policy Iteration. The main contribution of this paper is to show that, ...
Show more >Large Markov Decision Processes are usually solved using Approximate Dy-namic Programming methods such as Approximate Value Iteration or Ap-proximate Policy Iteration. The main contribution of this paper is to show that, alternatively, the optimal state-action value function can be estimated using Difference of Convex functions (DC) Programming. To do so, we study the minimization of a norm of the Optimal Bellman Residual (OBR) T * Q − Q, where T * is the so-called optimal Bellman operator. Control-ling this residual allows controlling the distance to the optimal action-value function, and we show that minimizing an empirical norm of the OBR is consistant in the Vapnik sense. Finally, we frame this optimization problem as a DC program. That allows envisioning using the large related literature on DC Programming to address the Reinforcement Leaning problem.Show less >
Show more >Large Markov Decision Processes are usually solved using Approximate Dy-namic Programming methods such as Approximate Value Iteration or Ap-proximate Policy Iteration. The main contribution of this paper is to show that, alternatively, the optimal state-action value function can be estimated using Difference of Convex functions (DC) Programming. To do so, we study the minimization of a norm of the Optimal Bellman Residual (OBR) T * Q − Q, where T * is the so-called optimal Bellman operator. Control-ling this residual allows controlling the distance to the optimal action-value function, and we show that minimizing an empirical norm of the OBR is consistant in the Vapnik sense. Finally, we frame this optimization problem as a DC program. That allows envisioning using the large related literature on DC Programming to address the Reinforcement Leaning problem.Show less >
Language :
Anglais
Peer reviewed article :
Oui
Audience :
Internationale
Popular science :
Non
ANR Project :
Collections :
Source :
Files
- https://hal.inria.fr/hal-01104419/document
- Open access
- Access the document
- https://hal.inria.fr/hal-01104419/document
- Open access
- Access the document
- https://hal.inria.fr/hal-01104419/document
- Open access
- Access the document
- document
- Open access
- Access the document
- 5443-difference-of-convex-functions-programming-for-reinforcement-learning.pdf
- Open access
- Access the document
- document
- Open access
- Access the document
- 5443-difference-of-convex-functions-programming-for-reinforcement-learning.pdf
- Open access
- Access the document