-
Notifications
You must be signed in to change notification settings - Fork 359
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[BUG][MISP] Connector creating a runaway task loop - Leading to Platform Stall #958
Comments
@SamuelHassine @richard-julien This is still a problem in 5.5.2, and I have confirmed its the source of OpenCTI-Platform/opencti#2603 on my end |
@SamuelHassine this is now becoming a source of major platform instability, post 5.5.2 Upgrade I am experiencing Redis OOM based platform stalls. This bug is still very present - Please re-open, if there is anything you need on my end please let me know / Ping me in Slack side note - The 5.5.2 workers are not processing tasks at nearly the same rate.
Like not tring to nitpick here, but this is a serious performance decrease. On my end resourcing has not changed Between this issue and OpenCTI-Platform/opencti#2603 - This is kinda a loss in cabin pressure scenario |
I'm keen to get this resolved, too. I'm experiencing the exact same issue. I keep increasing RAM but redis-server is consuming everything I give it. Currently at 54GB that I've allocated to the VM. Let me know if I can provide any information to help resolve this issue. The platform can't survive 24 hours without crash looping. |
I'm experiencing the exact same issue at the moment...total 32GB memory - redis 20GB |
Hi @llid3nlq and @Beskletochnii , If you have the trimming in place and the Redis memory is still not stable, we need to organize a call and deep dive into your connectors/data to identify who could be the culprit. |
@richard-julien |
Yes sure, you can find it here https://filigran.notion.site/Configuration-a568604c46d84f39a8beae141505572a#232c60ef21d9472298baf3c104ccd0c7 on the dependencies section |
Thank you for this information - it is very helpful. |
The stream is used in OpenCTI for multiple aspects, for example the "rule engine", the "notification system", the "live stream" etc. The only implication of the trimming is the time period where all features that use the stream will be able to recover if they are late of fail for some reason. So you need to setup a number according to your capacity of monitoring / resolving a platform problem if one of this component fail. I would say 1 month of retention should be sufficiant. If you want to be specific about the number of the number you want to put in the configuration you can connect to your stream on the /stream uri and check the number of messages your already have and do some maths to know the number that will represent 1 month of data. If you focus only for now on stabilize the size of your redis, you can for now put 1 million on the configuration, that will represents in average 4 Go of memory and un unknow period of time because its depends on the volume of data you absorb every day. |
Description
The MISP connector seems to be attempting to re-import events it has seen previous. This leads to major queue backlog in a matter of min.
The odd part, is this seems to be new behavior as of ~72 hours ago
Environment
Reproducible Steps
Not explicitly sure whats causing the events to be reimported, It seems marginally intermittent depending on latest MISP events
Steps to create the smallest reproducible scenario:
Expected Output
Import once MISP Event once
Actual Output
The MISP connector is dropping thousands of tasks to the queue every <10 Min.
Additional information
MISP Connector Logs
Screenshots (optional)
Example of the continuous dumps of the same data from the MISP Connector
The text was updated successfully, but these errors were encountered: