New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Remote pollers may sometimes fail to replicate data back to main system #2775
Comments
Which data collector? |
Ohh, good question! They are all on 192.168.252.3. |
Device IDs 493,494,495,496,497,498,499,512,513,532,536,537,538,542,545,546,547,548,576,577,578,579,580,581,605,608,612,613,614,615,619,651,671,672,673,675,678,679,718,719,720,721,723,724,725,726,794,812 |
So, is the re-indexing only not appearing to work on the remote data collectors? Can you confirm? |
Okay, found the culprit. |
Data Queries for remotely managed devices fail to push updated data to main data collector
Okay, just pushed two changes. What you can do to verify is that on the remote data collector you see the re-index being forced. Then, you should see updated information coming back to the main data collector from the remote. I'm not sure this will actually drive the automation, but this much is at least broken right now. So, there are a few ways you can verify, you could delete all the records for the device in the host_snmp_cache on the main table, and verify that they re-populate properly, which is what I did, or see if a new graph get's created. I've marked resolved, but I still have to trace a few things. |
Just walking through things, I still see a few issues. Still checking. |
Yea, automation is going to continue to fail. Let me see what I can do about that. |
Okay, so I just reviewed the code, and found that re-indexes that are performed by a remote device reboot, will not force an automation run. That's the way it's written right now. Looks like automation is only run from the main data collector. The question is in the case where we have a reboot, how can we schedule a run. I'll have to think about that. Please confirm what you are seeing on your end. Maybe delete a graph, then purge the host_snmp_cache for the device and data query in question on the main data collector, reboot the device to force a recache, then see that the data is repopulated in the host_snmp_cache after the data collector runs, then re-run automation on the main data collector to see that the graph is created again. |
Sorry for the delay I should be able to look at this in about 2-3 hours |
What I will do is simply change the name of a temperature sensor and then reboot the PDU. When the uptime goes backwards the Data query should reindex and update the name of the temperature sensor without any other intervention on my part. Should only take about 5 minutes to test. |
Cool. I know that I can force an automation run, but for now, we may just want to open another issue. |
Sorry, something went totally haywire over here and I've been trying to fix it. I think I merged too many updates and things got really quirky so I had to back-out of it |
OK. It worked
|
Okay, closing. |
Unfortunately I've also just found that data queries do not appear to be re-indexing when they should be. I have data queries setup with an Uptime re-index method, however I just restarted about 50 devices (after making specific changes to information the data query captures) and I had to re-run all the data queries by hand to refresh them.
Cacti is accurately obtaining the Uptime for these devices, at least I can see the correct Uptime listed in the Management->Devices pane.
The text was updated successfully, but these errors were encountered: