Replies: 1 comment 3 replies
-
Which shards are unassigned? For the above shard, what does this show? |
Beta Was this translation helpful? Give feedback.
3 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Version
2.4.60
Installation Method
Security Onion ISO image
Description
configuration
Installation Type
Distributed
Location
on-prem with Internet access
Hardware Specs
Exceeds minimum requirements
CPU
60
RAM
128
Storage for /
256
Storage for /nsm
256
Network Traffic Collection
span port
Network Traffic Speeds
Less than 1Gbps
Status
No, one or more services are failed (please provide detail below)
Salt Status
No, there are no failures
Logs
Yes, there are additional clues in /opt/so/log/ (please provide detail below)
Detail
Hey guys!
I've been fighting this issue for the past week or so. Starting out everything will work great, but then elasticsearch will show faulted in the grid consol. This is accompanied with a shard fail error in kibana and that same issue in Securityonion.log.
These shards have so far always shown in the unassigned report after running sudo so-elasticsearch-query _cat/shards | grep UN
I have quite a few, and it looks to be ever growing. I will move some out, just for some new ones to take its place. The kicker is these few that I have now, regardless of whether I issue the command to resolve them, they often stay. I have issued the commands with the appropriate users required as well.
What I have found that will work for me very briefly, is issuing that command then restarting my ENTIRE grid. Things will work for around 4ish hours, then this process repeats with a new shard failure on a new index. Which so far has always been unassigned. At least a replica of it anyway.
Currently this shard is:
ds-logs-endpoint.events.network-default-2024.04.30-000086
Sometimes I have also noticed this issue will resolve its self, but only for about that 4ish hour cycle described above.
Something I have also noticed is throughout this, the memory usage on the node will steadily climb much beyond normal operating range.
It might be worth noting my grid has shown pending for quite some time, with most of those indexes being older. I believe only these new indexes that are being dropped in there are the cause of the issue.
Any help is much appreciated as always!
Salt-call fails on the elastic-wait
Guidelines
Beta Was this translation helpful? Give feedback.
All reactions