Comments (4)
Try tensorboard --logdir=${PWD}
. You can launch tensorboard from any parent directory too. I'm using 0.8 and it works fine.
from deep_rl_ale.
First off, thanks for sharing this awesome repo. I am attempting to reproduce the deepmind results with TF and am slowly getting better results as I buff out the many subtle issues. Your repo has really helped me figure things out!
You're welcome! I'm excited that it's actually being used by others!
I wanted to know what results you get for space invaders (on a public post I saw you mentioned 1500), and exactly what args you are using. If haven't run deep_rl_ale for the full 200 epochs but at about 125 (with all default args) I was seeing average scores of about 1000-1100. Maybe I just need to let it run, but I wanted to make sure.
I got 1514 using --double_dqn and --gradient_clip=10. All other args were the defaults. DeepMind reports 1975 from the nature paper and 3154 with double dqn. These are significantly higher than my results, but I'm unaware of any differences between my implementation and theirs that might cause this. However, it looks like training progress hadn't yet converged and was still improving when the experiment ended:
It seems like I was getting 1000-1300 range around epoch 125.
I don't know if it would have kept improving, but it seems to be progressing much slower than theirs. I have changed the initialization of the moving averages in my rmsprop since then, which might make the training progress a little faster initially, but I don't think it would make a major difference.
Let me know what results you get with what settings! It takes a while to run these experiments, so the more data the better. You could try running it for more than 200 epochs if you have the time and see if it still improves. Let me know if you have any hypotheses about the performance differences as well. I plan on putting all of my results in the wiki when I have finished testing on a couple more games.
from deep_rl_ale.
Thats very helpful. I have a few other random questions if you are willing to contact me over mail (my github user name on gee male). I would also be interested in a more systematic hunt to reproduce the deepmind results with TF if you are game. Perhaps we could mount an offensive with help of others from the deep-q-learning google group.
from deep_rl_ale.
This question is off-topic to this discussion, but how did you use tensorboard to obtain the graph of score_per_game vs. epochs? I tried to use tensorboard --logdir <path/to/records/dir_containing_tf_events>
but was not able to see any graphs. Maybe this broke when I upgraded my tensorflow to 0.8?
from deep_rl_ale.
Related Issues (3)
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from deep_rl_ale.