SAC: Soft Actor Critic¶
This baseline comes from the paper: Soft Actor-Critic: Off-Policy Maximum Entropy Deep Reinforcement Learning with a Stochastic Actor
This module proposes an implementation of the SAC algorithm.
This is an old implementation that is probably not correct, it was included out of backward compatibility with earlier version (< 0.5.0) of this package
An example to train this model is available in the train function Example-sacold.
You can use this class with:
from l2rpn_baselines.SACOld import train, evaluate, SACOld
Other non exported class¶
These classes need to be imported, if you want to import them with (non exhaustive list): .. code-block:: python
from l2rpn_baselines.SACOld.SACOld_NN import SACOld_NN from l2rpn_baselines.SACOld.SACOld_NNParam import SACOld_NNParam