Revisiting Natural Actor-Critics with Value Function Approximation - CentraleSupélec Accéder directement au contenu
Communication Dans Un Congrès Année : 2010

Revisiting Natural Actor-Critics with Value Function Approximation

Matthieu Geist
Olivier Pietquin

Résumé

Actor-critics architectures have become popular during the last decade in the field of reinforcement learning because of the introduction of the policy gradient with function approximation theorem. It allows combining rationally actor-critic architectures with value function approximation and therefore addressing large-scale problems. Recent researches led to the replacement of policy gradient by a natural policy gradient, improving the efficiency of the corresponding algorithms. However, a common drawback of these approaches is that they require the manipulation of the so-called advantage function which does not satisfy any Bellman equation. Consequently, derivation of actor-critic algorithms is not straightforward. In this paper, we re-derive theorems in a way that allows reasoning directly with the state-action value function (or Q-function) and thus relying on the Bellman equation again. Consequently, new forms of critics can easily be integrated in the actor-critic framework.

Dates et versions

hal-00553870 , version 1 (10-01-2011)

Identifiants

Citer

Matthieu Geist, Olivier Pietquin. Revisiting Natural Actor-Critics with Value Function Approximation. MDAI 2010, Oct 2010, Perpignan, France. pp.207-218, ⟨10.1007/978-3-642-16292-3_21⟩. ⟨hal-00553870⟩
320 Consultations
0 Téléchargements

Altmetric

Partager

Gmail Facebook X LinkedIn More