Skip to content
This repository has been archived by the owner on Jun 13, 2019. It is now read-only.

improve after learning? #14

Closed
panda2048 opened this issue Nov 8, 2017 · 4 comments
Closed

improve after learning? #14

panda2048 opened this issue Nov 8, 2017 · 4 comments

Comments

@panda2048
Copy link

I am expecting the DQN was there to improve the "dice guess", so over time, the max loose streak will be lower down (i.e. lower down the min. satoish required to start)
But after running for 2 millions round. I do not actually seeing significant change
Is my understanding correct?
If so, when should we expect the improvement?

@mbithy
Copy link
Owner

mbithy commented Nov 8, 2017

yup, theoretically, this could be an indication that there is no advantage to be gained by using MA OR the ReinforceJS mathematical equations aren't capable of beating the randomness of provably fair OR we(me) from start are using the wrong "inputs" to try and get a MA algorithm to find a pattern... I'm leaning on the first one.

@mbithy
Copy link
Owner

mbithy commented Nov 9, 2017

@osavigne cool stuff, I would like to see

@mbithy mbithy closed this as completed Feb 24, 2018
Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants
@mbithy @panda2048 and others