You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I need to test whether using the KL divergence between a "baseline" run and a "sampled" (resampled + reweighted) run can be used to determine convergence to the posterior distribution.
The text was updated successfully, but these errors were encountered:
Initial tests seem to indicate that the variation in the KL divergence (not necessarily the mean) between the "distribution" formed by a "sampled" run and a "baseline" run seems to behave as I had hoped. This makes some sense in retrospect: we expect there to be some typical divergence, so it's the stability in this divergence that actually matters. I might add some more notes here for (tentative) public posterity depending on how further tests turn out.
Although resampling complicates the procedure algorithmically, I managed to implement something that works. Essentially, we compute the KL divergence from the perspective of the "first" run (we're computing the KL divergence to this run), which may or may not include resampled positions, and then evaluate terms on the KL divergence one-by-one by position. I split this by unique particle IDs to speed up position checks. In the case where there are multiple positions in the "second" run (we're computing the KL divergence from this run) that match (leading to possible confusion), I choose the term that is closest to zero (which correctly gives zero overall in the case where we compare the same "resampled" run against itself).
After some testing using a 3-D correlated normal, I ended up setting the default stopping values to 2% fractional variation in the KL divergence and 0.1 standard deviation in lnz (~10%), which gave similar sample sizes to a standard nested sampling run with K=1000 live points under posterior-optimized and evidence-optimized estimation, respectively. I might change this later depending on more testing.
I need to test whether using the KL divergence between a "baseline" run and a "sampled" (resampled + reweighted) run can be used to determine convergence to the posterior distribution.
The text was updated successfully, but these errors were encountered: