Geometric Deep Reinforcement Learning for ...
Type de document :
Communication dans un congrès avec actes
Titre :
Geometric Deep Reinforcement Learning for Dynamic DAG Scheduling
Auteur(s) :
Grinsztajn, Nathan [Auteur]
Scool [Scool]
Beaumont, Olivier [Auteur]
High-End Parallel Algorithms for Challenging Numerical Simulations [HiePACS]
Jeannot, Emmanuel [Auteur]
Topology-Aware System-Scale Data Management for High-Performance Computing [TADAAM]
Preux, Philippe [Auteur]
Scool [Scool]
Scool [Scool]
Beaumont, Olivier [Auteur]
High-End Parallel Algorithms for Challenging Numerical Simulations [HiePACS]
Jeannot, Emmanuel [Auteur]
Topology-Aware System-Scale Data Management for High-Performance Computing [TADAAM]
Preux, Philippe [Auteur]
Scool [Scool]
Titre de la manifestation scientifique :
IEEE SSCI 2020 - Symposium Series on Computational Intelligence
Ville :
Canberra / Virtual
Pays :
Australie
Date de début de la manifestation scientifique :
2020-12-01
Titre de la revue :
SSCI 2020 proceedings
Date de publication :
2020-12
Mot(s)-clé(s) en anglais :
Reinforcement learning
scheduling
task graph
DAG
high performance computing
combinatorial optimization
scheduling
task graph
DAG
high performance computing
combinatorial optimization
Discipline(s) HAL :
Informatique [cs]/Intelligence artificielle [cs.AI]
Résumé en anglais : [en]
In practice, it is quite common to face combinatorial optimization problems which contain uncertainty along with non-determinism and dynamicity. These three properties call for appropriate algorithms; reinforcement learning ...
Lire la suite >In practice, it is quite common to face combinatorial optimization problems which contain uncertainty along with non-determinism and dynamicity. These three properties call for appropriate algorithms; reinforcement learning (RL) is dealing with them in a very natural way. Today, despite some efforts, most real-life combinatorial optimization problems remain out of the reach of reinforcement learning algorithms. In this paper, we propose a reinforcement learning approach to solve a realistic scheduling problem, and apply it to an algorithm commonly executed in the high performance computing community, the Cholesky factorization. On the contrary to static scheduling, where tasks are assigned to processors in a predetermined ordering before the beginning of the parallel execution, our method is dynamic: task allocations and their execution ordering are decided at runtime, based on the system state and unexpected events, which allows much more flexibility. To do so, our algorithm uses graph neural networks in combination with an actor-critic algorithm (A2C) to build an adaptive representation of the problem on the fly. We show that this approach is competitive with state-of-the-art heuristics used in high-performance computing runtime systems. Moreover, our algorithm does not require an explicit model of the environment, but we demonstrate that extra knowledge can easily be incorporated and improves performance. We also exhibit key properties provided by this RL approach, and study its transfer abilities to other instances.Lire moins >
Lire la suite >In practice, it is quite common to face combinatorial optimization problems which contain uncertainty along with non-determinism and dynamicity. These three properties call for appropriate algorithms; reinforcement learning (RL) is dealing with them in a very natural way. Today, despite some efforts, most real-life combinatorial optimization problems remain out of the reach of reinforcement learning algorithms. In this paper, we propose a reinforcement learning approach to solve a realistic scheduling problem, and apply it to an algorithm commonly executed in the high performance computing community, the Cholesky factorization. On the contrary to static scheduling, where tasks are assigned to processors in a predetermined ordering before the beginning of the parallel execution, our method is dynamic: task allocations and their execution ordering are decided at runtime, based on the system state and unexpected events, which allows much more flexibility. To do so, our algorithm uses graph neural networks in combination with an actor-critic algorithm (A2C) to build an adaptive representation of the problem on the fly. We show that this approach is competitive with state-of-the-art heuristics used in high-performance computing runtime systems. Moreover, our algorithm does not require an explicit model of the environment, but we demonstrate that extra knowledge can easily be incorporated and improves performance. We also exhibit key properties provided by this RL approach, and study its transfer abilities to other instances.Lire moins >
Langue :
Anglais
Comité de lecture :
Oui
Audience :
Internationale
Vulgarisation :
Non
Collections :
Source :
Fichiers
- https://hal.inria.fr/hal-03028981/document
- Accès libre
- Accéder au document
- http://arxiv.org/pdf/2011.04333
- Accès libre
- Accéder au document
- https://hal.inria.fr/hal-03028981/document
- Accès libre
- Accéder au document
- https://hal.inria.fr/hal-03028981/document
- Accès libre
- Accéder au document
- document
- Accès libre
- Accéder au document
- HPC_ADPRL.pdf
- Accès libre
- Accéder au document
- 2011.04333
- Accès libre
- Accéder au document
- document
- Accès libre
- Accéder au document
- HPC_ADPRL.pdf
- Accès libre
- Accéder au document