Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Documentation issue: spark, heuristics and hdfs readers #103

Closed
kockockockoc opened this issue Feb 6, 2019 · 2 comments
Closed

Documentation issue: spark, heuristics and hdfs readers #103

kockockockoc opened this issue Feb 6, 2019 · 2 comments

Comments

@kockockockoc
Copy link
Contributor

hi!

Could you give some tips/hints/examples how install/config/use:

  • garmadon-frameworks-spark
  • garmadon-readers-hdfs
  • garmadon-readers-heuristics

And how to use (what is it for?):

  • garmadon-schema.jar
  • garmadon-jvm-statistics.jar (garmadon-jvm-statistics-*)

thanks in advance for your help

@ashangit
Copy link
Contributor

ashangit commented Feb 7, 2019

Hi @kockockockoc ,

Only garmadon-agent, garmadon-forwarder and garmadon-readers-* are usefull to deploy garmadon
Other jars are dependencies to agent/forwarder/readers jars to share protobuf schema, some mechanism to trace metrics...

garmadon-frameworks-spark is loaded by the garmadon-agent to retreive metrics from spark application using Spark Listener mechanism.
The listener is automatically attached to spark driver if you use StandaloneModule (-javaagent:path-to-garmadon-agent-jar=com.criteo.hadoop.garmadon.agent.modules.StandaloneModule) or ContainerModule (-javaagent:path-to-garmadon-agent-jar=com.criteo.hadoop.garmadon.agent.modules.ContainerModule)

executorService.submit(() -> {
    SparkListenerTracer.setup(ContainerHeader.getInstance().getHeader(),
            (timestamp, header, event) -> eventProcessor.offer(timestamp, header, event));
});
Properties props = System.getProperties();
        props.setProperty("spark.extraListeners", "com.criteo.hadoop.garmadon.spark.listener.GarmadonSparkListener");

So you don't have any specific things to do, only need to bind the agent to all spark JVMs (client and containers).
Will see to improve the documentation around garmadon and spark

For readers:

  • garmadon-readers-hdfs, will need to add some documentation about it, @gray-matter ?
  • garmadon-readers-heuristics, it is still bind to some specificities from criteo so not really available for now.

@kockockockoc
Copy link
Contributor Author

Thank you for explaining. I was a bit confused by the fact that there were separate jars after "mvn clean package".

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants