mirror of
https://github.com/vale981/ray
synced 2025-03-06 10:31:39 -05:00
Use observation filter in compute_action for PPO. (#884)
This commit is contained in:
parent
5d72818ddc
commit
60d4d01d06
1 changed files with 1 additions and 0 deletions
|
@ -266,4 +266,5 @@ class PolicyGradient(Algorithm):
|
|||
for (a, o) in zip(self.agents, extra_data[4])])
|
||||
|
||||
def compute_action(self, observation):
|
||||
observation = self.model.observation_filter(observation)
|
||||
return self.model.common_policy.compute([observation])[0][0]
|
||||
|
|
Loading…
Add table
Reference in a new issue