Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Make CPU reduction more precisse in float32 #1092

Closed
nouiz opened this issue Nov 23, 2012 · 2 comments
Closed

Make CPU reduction more precisse in float32 #1092

nouiz opened this issue Nov 23, 2012 · 2 comments
Milestone

Comments

@nouiz
Copy link
Member

nouiz commented Nov 23, 2012

See: https://groups.google.com/d/topic/theano-dev/g2w3LhJMSH8/discussion

Possible fix:

  • Do the reduction in float64. Maybe this won't slow down as it should be memory bound. Need test.
  • Do multiple level of summation. This would help for other dtype like int and float64. It would help in the int case to prevent overflow when there is a mix of positive and negative value.
    • We could do a sum per row. Probably easy code change, but won't help for big row.

I make it as a high priority as when running on the GPU in debugmode, this cause error, as the GPU sum do this, so it is more precise.

But solving the gpu sum useless error will probably create a new one: theano c_code vs numpy code as numpy code don't do this.

@nouiz
Copy link
Member Author

nouiz commented Jan 14, 2013

see numpy/numpy#2448 that want to be more general. Some of the suggestion seam fast to do! So why not fix that upstream?

@lamblin
Copy link
Member

lamblin commented Mar 26, 2013

Fixed by gh-1226.

@lamblin lamblin closed this as completed Mar 26, 2013
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

2 participants