Skip to content

Data loss when running compactor for fixed time range #6293

Closed Answered by v1jayr
v1jayr asked this question in Questions & Answers
Discussion options

You must be logged in to vote

After bit more digging, it appears that downsampling requires source blocks to be of certain size. In my case 5m downsample did occur but not 1hour. The 1h downsampling requires blocks size of 10 days and since the time range I ran the compactor for was not a multiple of 10 days the compacted blocks for the missing period were <10 days long. So downsampling skipped those blocks. Then retention hit and cleaned up the raw and 5m blocks. Its possible there are other edge cases I might be missing.

Conclusion:
I guess the safest possible path forward is to compact with retention time increased to a large enough value. I am also not sure what the best approach is to work through a large backlog…

Replies: 2 comments 3 replies

Comment options

You must be logged in to vote
3 replies
@v1jayr
Comment options

@v1jayr
Comment options

Answer selected by v1jayr
@BenoitPoulet
Comment options

Comment options

You must be logged in to vote
0 replies
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
2 participants