-
Notifications
You must be signed in to change notification settings - Fork 231
Type of self.output in policy.py #38
Comments
I found in previous commits that output used to be, but later due to log_likelihood_cost refactor got deleted. output = tf.nn.softmax(tf.reshape(h_conv_final, [-1, go.N ** 2]) + b_conv_final) |
Hm. Sorry about that - work on this repo is continuing at https://github.com/tensorflow/minigo. I'll update the README.md |
Hey,
It might result in such error, to fix output issue you can add this line:
self.output = tf.nn.softmax(tf.reshape(h_conv_final, [-1, go.N ** 2]) +
b_conv_final)
at line 88 and it will work.
However as @brilee mentioned, MuGo is no longer developed it's all switched
to minigo.
You may want to check out leela zero and/or lizzie(very easy configuration
here).
czw., 31 maj 2018 o 10:43 JoeyQ Wu <notifications@github.com> napisał(a):
… Hello, @CezCz <https://github.com/CezCz> @brilee
<https://github.com/brilee>
I just met the same question "AttributeError: 'PolicyNetwork' object has
no attribute 'output'
" , and I want to ask you about whether it can result in the error "GTP
Stream was Closed" .
what should I do if I want this program can run the correct result ?
—
You are receiving this because you were mentioned.
Reply to this email directly, view it on GitHub
<#38 (comment)>, or mute
the thread
<https://github.com/notifications/unsubscribe-auth/ALPOFnDcWIVdpPk3U1aA7ZK54WC1TvZfks5t360wgaJpZM4SlQyF>
.
--
Cezary Czernecki
|
@CezCz yeah, thanks for your kind answer, |
Hey,
log_likelihood_cost is another problem, as now you need named logits
parameter. I am happy that you managed to fix it.
I am not sure what you are talking about when saying negative and bigger
value? Can you provide an example?
Cezary
wt., 5 cze 2018 o 09:34 Joeyq <notifications@github.com> napisał(a):
… @CezCz <https://github.com/CezCz> yeah, thanks for your kind answer,
actually, I fixed the line 88 with
" log_likelihood_cost =
tf.reduce_sum(tf.nn.softmax_cross_entropy_with_logits(logits=logits,
labels=y))" and it could work,
but I could not understand the output of mcts, why it often choose the
bigger value even if it is negative ?
I am confused about the result , I would appreciate if you can tell me the
reason @CezCz <https://github.com/CezCz>
—
You are receiving this because you were mentioned.
Reply to this email directly, view it on GitHub
<#38 (comment)>, or mute
the thread
<https://github.com/notifications/unsubscribe-auth/ALPOFuLIha42c3GOEL8RL3KYlx3dTolTks5t5jSdgaJpZM4SlQyF>
.
--
Cezary Czernecki
|
just like the first picture , the location of white is R4, and I get the value is -7.5, just as the second image, |
Hello,
Move that is to play is solely chosen based on visit count, what are they
for those two nodes?
I've prepared a silly image describing what might be going on:
[image](https://user-images.githubusercontent.com/11783702/40991674-cbbf834a-68f4-11e8-9f81-8ac087a70b55.png)
Let's consider only depth of 2 is checked and only 1 search is being done.
The state of tree in the picture is after backpropagation of this first
search, I marked selection with blue pen, then backpropagation poorly with
black. As you can see value network said the value of position is -0,98
(let's assume -1 is max). We can clearly see it is bad, however when the
final move is chosen, only the visit count N is considered. In the end
(1,1) node has the most visits therefor it is chosen.
Cezary
|
Hi , @CezCz |
@JoeyQWu |
@CezCz |
I also wrote http://www.moderndescartes.com/essays/deep_dive_mcts/ recently |
Hey @brilee ,
I'm trying to play against this wonderful library. However when I'm trying genmove b I'm getting:
File "\MuGo\policy.py", line 152, in run
probabilities = self.session.run(self.output, feed_dict={self.x: processed_position[None, :]})[0]
AttributeError: 'PolicyNetwork' object has no attribute 'output'
What should self.output be ?
The text was updated successfully, but these errors were encountered: