SAC: Soft Actor Critic

This baseline comes from the paper: Soft Actor-Critic: Off-Policy Maximum Entropy Deep Reinforcement Learning with a Stochastic Actor


This module proposes an implementation of the SAC algorithm.

This is an old implementation that is probably not correct, it was included out of backward compatibility with earlier version (< 0.5.0) of this package

An example to train this model is available in the train function Example-sacold.

Exported class

You can use this class with:

from l2rpn_baselines.SACOld import train, evaluate, SACOld

Other non exported class

These classes need to be imported, if you want to import them with (non exhaustive list): .. code-block:: python

from l2rpn_baselines.SACOld.SACOld_NN import SACOld_NN from l2rpn_baselines.SACOld.SACOld_NNParam import SACOld_NNParam