Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[0.13.0-1] large shards compaction never ends, influx_inspect on the tsm files panics #6683

Closed
t-orb opened this issue May 19, 2016 · 3 comments
Assignees
Labels
Milestone

Comments

@t-orb
Copy link

t-orb commented May 19, 2016

System info: [0.13.0-1 on Ubuntu 14.04.4 LTS]

Steps to reproduce:

  1. [use default retention policy of each shard covering 1 week]
  2. [load in more than a week of data creating big tsm files]
  3. [wait until 24hr compaction run starts]

Expected behavior:

If more than 3 TSM files were present I'd expect 1 compaction run across the files

Actual behavior:

It starts compacting and never stops, continues until restarting influxdb.

Additional info: [Include gist of relevant config, logs, etc.]

Here's a peek at one example shard, the logs from last night and influx_inspect failing (until now):

https://gist.github.com/t-orb/bbe0ea12a9bff7bb098728c77d347f23

I asked on Slack and Gunnar Aasen asked me to open an issue.

After listening to one of your webex training sessions I realised that for the amount of data I was putting in the default of each shard covering a week was too big so I switched to 24 hours per shard, all the new shards don't have this issue. So it's likely something going wrong with very large shards/tsm files.

I just realized that running influx_inspect on the files panics.. That's a hint I guess.

The files otherwise appears to work just fine as I can query data in the period just fine.

Thanks for taking a look.

@jwilder
Copy link
Contributor

jwilder commented May 19, 2016

You need to run influx_inspect dumptsmdev <path>. The dumptsm command is for an older version of TSM files with a different format which is why it panics.

@jwilder
Copy link
Contributor

jwilder commented May 19, 2016

@t-orb Are you deleting series or measurements by chance? If you delete data that is contained in these TSM files, they will get re-compacted to expunge that data.

@jwilder jwilder added this to the 1.0.0 milestone May 19, 2016
@t-orb
Copy link
Author

t-orb commented May 19, 2016

I just updated my gist with a comment containing the dumptsmdev output (no panics).

Never done a delete. No inserts to this database the last day either. It is just running the 24 hour compacting cycle.

The file size of 2148728539 is perhaps a problem?

@jwilder jwilder self-assigned this May 20, 2016
jwilder added a commit that referenced this issue May 23, 2016
The level planner would keep including the same TSM files to be
recompacted even if they were already quite compacted and split
across several TSM files.

Fixes #6683
jwilder added a commit that referenced this issue May 25, 2016
The level planner would keep including the same TSM files to be
recompacted even if they were already quite compacted and split
across several TSM files.

Fixes #6683
@timhallinflux timhallinflux modified the milestones: 1.0.0, 1.0.0 beta Dec 20, 2016
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Projects
None yet
Development

No branches or pull requests

3 participants