Skip to content
This repository was archived by the owner on Jun 8, 2020. It is now read-only.

Commit 53051f2

Browse files
authored
Update README.md
1 parent dc009e5 commit 53051f2

File tree

1 file changed

+2
-2
lines changed

1 file changed

+2
-2
lines changed

README.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -22,8 +22,8 @@ Implementations of important policy gradient algorithms in deep reinforcement le
2222

2323
- Soft Actor Critic (SAC)
2424

25-
- `(sac2018.py`) ["Soft Actor-Critic: Off-Policy Maximum Entropy Deep Reinforcement Learning with a Stochastic Actor" (Haarnoja et al., 2018)](https://arxiv.org/abs/1801.01290)
26-
- (`sac2019.py`) ["Soft Actor-Critic Algorithms and Applications" (Haarnoja et al., 2019)](https://arxiv.org/abs/1812.05905)
25+
- Papaer (`sac2018.py`): ["Soft Actor-Critic: Off-Policy Maximum Entropy Deep Reinforcement Learning with a Stochastic Actor" (Haarnoja et al., 2018)](https://arxiv.org/abs/1801.01290)
26+
- Paper (`sac2019.py`): ["Soft Actor-Critic Algorithms and Applications" (Haarnoja et al., 2019)](https://arxiv.org/abs/1812.05905)
2727
- Algorithm in 2018 paper uses value network, double Q networks, and Gaussian policy. Algorithm in 2019 paper uses double Q networks and Gaussian policy, and adds automatic entropy tuning.
2828
- TODO: SAC for discrete action space
2929

0 commit comments

Comments
 (0)