You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
However, with 8 integer bits and 8 fractional bits, the original tensor values should definitely be representable, so the result should be the exact same. Would be great to take a look at this and see what's going on!
The text was updated successfully, but these errors were encountered:
So wl is an acronym of word length, which refers to the total number of bits used to represent the numbers. In other words, wl=1+integer bits+fractional bits.
so in this example, integer bits is actually -1, so the smallest representable number is 2^(-1) = -0.5.
Hi again! I'm experiencing some weird results in
fixed_point_quantize
that I think might not be correct behavior. Here is an example:The output of this is:
However, with 8 integer bits and 8 fractional bits, the original tensor values should definitely be representable, so the result should be the exact same. Would be great to take a look at this and see what's going on!
The text was updated successfully, but these errors were encountered: