Through Deep Reinforcement Learning, this agent can autonomously train itself to play Out Run and potentially be modified to play other games or perform tasks other than gaming.
More details about the training process in this blogpost.
Built with Python and Tensorflow.
-
Clone the custom Cannonball fork, which contains the changes needed to access the emulator externally.
-
Build Cannonball:
# On Cannonball's root folder do: $ mkdir build $ cd build # Building on OSX / MacOS $ cmake -G "Unix Makefiles" -DTARGET:STRING=sdl2_macosx ../cmake # If building on Linux, execute this commented line instead: # $ cmake -G 'Unix Makefiles' -DTARGET:STRING=sdl2 ../cmake $ make
-
Copy the built shared object file (libcannonball.so) to agent-trainer's lib folder:
$ cp <cannonball-folder>/build/libcannonball.so <agent-trainer-folder>/lib/libcannonball.so
Copy your game roms to <agent-trainer-folder>/roms/
. More details here.
Set the train_results_root_folder
parameter in config.py. This will be the default folder used to write and read training results (can be overriden, see below)
On the root folder, run:
$ make
This installs the depencies via pip. Feel free to use virtual env wrapper (for example) to contain these.
Dependencies on a GPU enabled Linux machine
If you are running a Linux machine and want to make use of its GPU for training, use the following flag when installing dependencies:
$ USE_GPU=true make
You can use the trainer via make tasks:
# Run all the tests
$ make test
# Start new training session
$ make train-new
# Resume training session
$ SESSION_ID="201609272034" make train-resume
# Play a previously trained session
$ SESSION_ID="201609272034" make play
# Create a t-Distributed Stochastic Neighbor Embedding (t-SNE) visualization, placed on `<train-results-root-folder>/<session-id>/visualizations/t-SNE.png`
# Example output: https://github.com/lopespm/agent-trainer-results/blob/master/201609171218_175eps/visualizations/t-SNE_no_time_mode.png
$ SESSION_ID="201609272034" make visualize-tsne
# Export the metrics of the given session, as PNG images files. These will be placed on `<train-results-root-folder>/<session-id>/metrics-session/`
# Example output: https://github.com/lopespm/agent-trainer-results/tree/master/201609171218_175eps/metrics-session
$ SESSION_ID="201609272034" make metrics-export
# Show the session's metrics on-screen
$ SESSION_ID="201609272034" make metrics-show
For finer control, you can run the library module as a script. For example:
python -m agent play --ec2spot --resultspath /example/alternative-results-folder -s 201609261533
Actions: train-new
, train-resume
, play
, visualize-tsne
, metrics-show
or metrics-export
Options:
-s
: define the session ID--ec2spot
: when used, it will check on which episode if the spot instance is scheduled for termination, acting accordingly by saving the current session and halting training--resultspath
: overrides the defaulttrain_results_root_folder
parameter
You can deploy the agent to a generic remote Linux machine or to an AWS EC2 instance with the help of agent-trainer-deployer