We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 7f8b329 commit 1318058Copy full SHA for 1318058
reinforcement_learning/actor_critic.py
@@ -63,7 +63,7 @@ def forward(self, x):
63
# critic: evaluates being in the state s_t
64
state_values = self.value_head(x)
65
66
- # return values for both actor and critic as a tupel of 2 values:
+ # return values for both actor and critic as a tuple of 2 values:
67
# 1. a list with the probability of each action over the action space
68
# 2. the value from state s_t
69
return action_prob, state_values
0 commit comments