Skip to content
This repository has been archived by the owner on May 12, 2021. It is now read-only.

METRON-1467: Replace guava caches in places where the keyspace might be large #947

Closed
wants to merge 30 commits into from

Conversation

cestella
Copy link
Member

@cestella cestella commented Mar 2, 2018

Contributor Comments

Based on the performance tuning exercise as part of METRON-1460, guava has difficulties with cache sizes over 10k. We, unfortunately, are quite demanding of guava in this regard so we should transition a few uses of guava to Caffeine:

  • Stellar processor cache
  • The JoinBolt cache
  • The Enrichment Bolt Cache

Test plan:
First, spin up full-dev and ensure things continue to work.

For myself, I created a new type called dummy, added some stellar enrichments, stellar threat triage enrichments and a triage rule to ensure things continued to work with stellar (exercising the stellar processor cache and the enrichment bolt cache).

Pull Request Checklist

Thank you for submitting a contribution to Apache Metron.
Please refer to our Development Guidelines for the complete guide to follow for contributions.
Please refer also to our Build Verification Guidelines for complete smoke testing guides.

In order to streamline the review of the contribution we ask you follow these guidelines and ask you to double check the following:

For all changes:

  • Is there a JIRA ticket associated with this PR? If not one needs to be created at Metron Jira.
  • Does your PR title start with METRON-XXXX where XXXX is the JIRA number you are trying to resolve? Pay particular attention to the hyphen "-" character.
  • Has your PR been rebased against the latest commit within the target branch (typically master)?

For code changes:

  • Have you included steps to reproduce the behavior or problem that is being changed or addressed?

  • Have you included steps or a guide to how the change may be verified and tested manually?

  • Have you ensured that the full suite of tests and checks have been executed in the root metron folder via:

    mvn -q clean integration-test install && dev-utilities/build-utils/verify_licenses.sh 
    
  • Have you written or updated unit tests and or integration tests to verify your changes?

  • If adding new dependencies to the code, are these dependencies licensed in a way that is compatible for inclusion under ASF 2.0?

  • Have you verified the basic functionality of the build by building and running locally with Vagrant full-dev environment or the equivalent?

For documentation related changes:

  • Have you ensured that format looks appropriate for the output in which it is rendered by building and verifying the site-book? If not then run the following commands and the verify changes via site-book/target/site/index.html:

    cd site-book
    mvn site
    

Note:

Please ensure that once the PR is submitted, you check travis-ci for build issues and submit an update to your PR as soon as possible.
It is also recommended that travis-ci is set up for your personal repository such that your branches are built there before submitting a pull request.

@cestella cestella changed the title METRON-1467: Replace guava caches in places where the keyspace might be large METRON-1467: Replace guava caches in places where the keyspace might be large (NOTE: Review after METRON-1460) Mar 2, 2018
@cestella cestella closed this Mar 5, 2018
@cestella cestella reopened this Mar 5, 2018
@cestella
Copy link
Member Author

cestella commented Mar 5, 2018

I ran this up with vagrant and ensured:

  • Normal stellar works still in field transformations as well as enrichments
  • swapped in and out new enrichments live
  • swapped in and out new threat intel live

@cestella cestella changed the title METRON-1467: Replace guava caches in places where the keyspace might be large (NOTE: Review after METRON-1460) METRON-1467: Replace guava caches in places where the keyspace might be large Mar 7, 2018
Copy link
Contributor

@nickwallen nickwallen left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Everything looks good. Just one comment on the "remove listener".

We know how important performance of these caches are, so it is important to sweat the small stuff here.

loader = s -> new HashMap<>();
cache = Caffeine.newBuilder().maximumSize(maxCacheSize)
.expireAfterWrite(maxTimeRetain, TimeUnit.MINUTES)
.removalListener(new JoinRemoveListener())
Copy link
Contributor

@nickwallen nickwallen Mar 7, 2018

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

It seems like we only want notified of a full cache when ERROR logging is set. Is that the case?

In the JoinRemoveListener we end up doing some work that we probably don't need to do unless ERROR logging is set. One easy fix would be to only add the "remove listener" if LOG.isDebugEnabled().

Copy link
Member Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

So, I believe this was intentionally done before this PR (I migrated this to the new caching strategy) and the idea is that if a removal is happening from the join cache under specific circumstances, we want to know about it because a message could be being dropped because the cache is being overwhelmed. @merrimanr Can you chime in here on the rationale?

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Yes, it is pre-existing. We can address at a later time.

I remember now, maxing out this cache causes the Split/Join to fail, which is a major problem for the Split/Join topology. And this cache here is only for the Split/Join, not the Unified topology.

We should probably look at adding similar logging (only when ERROR enabled) for the other places where we use the cache. Or just some mechanism to periodically log cache stats. Anywho, down the road.

@nickwallen
Copy link
Contributor

+1 LGTM

Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
None yet
Projects
None yet
2 participants