Skip to content

Parameter transformation #943

Answered by lawrennd
patel-zeel asked this question in Q&A
Discussion options

You must be logged in to vote

Hi Zeel ... I think (if I remember well) logexp is implementing what people (nowadays) call the "soft relu" so log(1+exp(x))

It's nice because it's monotonic ... and it behaves linearly in the positive half space. But exp() or square() are also valid. Square is not monotonic, so if it's used (e.g.) in a Bayesian model ... then it causes a bimodality that wasn't there in the original model. exp() is good too but often will cause numerical overflow problems. If I remember well, the implementation is likely clipped to avoid this, but it's a little bit of a sludgy solution.

Neil

Replies: 1 comment

Comment options

You must be logged in to vote
0 replies
Answer selected by patel-zeel
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
2 participants