Skip to content

Conversation

pierrelux
Copy link
Contributor

This example shows how to use the special TrajectoryRecorder decorator for saving trajectories. Furthermore, it uses IPython parallel to compute the performance of many agents in parallel.

It can be used over a guerilla SSH cluster over the computers in your lab, or more cleanly on an MPI, PBS or AWS cluster.

For local parallel computation, execute in a separate tab:
$ ipcluster start -n 4

followed by
$ python learn-flat-policy.py -a 100 -n 100 pinball_simple_single.cfg

Fix some inconsistencies with RL-Glue. Add the ability
to replay saved trajectories.
This example shows how to use the special TrajectoryRecorder
decorator for saving trajectories. Furthermore, it uses
IPython parallel to compute the performance of many agents
in parallel.

It can be used over a guerilla SSH cluster over the computers in your
lab, or more cleanly on an MPI, PBS or AWS cluster.

For local parallel computation, execute in a separate tab:
$ ipcluster start -n 4

followed by
$ python learn-flat-policy.py -a 100 -n 100 pinball_simple_single.cfg
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

1 participant