Skip to content

Move to Spark 1.6.2 and Java 8 (resolves #231)#233

Merged
hannes-ucsc merged 1 commit intoBD2KGenomics:masterfrom
fnothaft:issues/231-spark-162-java8
Nov 14, 2016
Merged

Move to Spark 1.6.2 and Java 8 (resolves #231)#233
hannes-ucsc merged 1 commit intoBD2KGenomics:masterfrom
fnothaft:issues/231-spark-162-java8

Conversation

@fnothaft
Copy link
Contributor

No description provided.

@fnothaft
Copy link
Contributor Author

Hold on merge, needs to be tested.

@jpdna
Copy link

jpdna commented Oct 16, 2016

This branch not working for me yet.

I tried testing fnothaft:issues/231-spark-162-java8
by installing using "Developer" instructions make develop sdist after cloning this branch
and to install cgcloud-spark running python setup.py install
that worked it seemed, and I was able to create a test cluster of one machine.

But logging into master to test spark I find:

sparkbox@ip-172-31-37-40:~$ java -version
java version "1.8.0_101"
Java(TM) SE Runtime Environment (build 1.8.0_101-b13)
Java HotSpot(TM) 64-Bit Server VM (build 25.101-b13, mixed mode)
sparkbox@ip-172-31-37-40:~$ spark-shell
/opt/sparkbox/spark/bin/spark-class: line 86: /usr/lib/jvm/java-7-oracle/bin/java: No such file or directory

I'm also suspicious of reference to java-7 here:
https://github.com/fnothaft/cgcloud/blob/issues/231-spark-162-java8/spark/src/cgcloud/spark/spark_box.py#L198

It's possible that I am not installing locally correctly from the branch - let me know if I should try again a different way if this is working on your end @fnothaft .

@fnothaft
Copy link
Contributor Author

Ah! That is my mistake, will fix in AM.

@jpdna
Copy link

jpdna commented Oct 18, 2016

Ping on this, later this week/weekend I'd like to make use of this PR.
If you think changes are just a few more s/java-7/java-8/ I will look into if you can't get to this, just let me know.

@jpdna
Copy link

jpdna commented Oct 21, 2016

@fnothaft - do you want me to look into this further if you are swamped?
Do you think it is just a matter of more s/java-7/java-8/
spark_box.py or any other pointers you would have?

@fnothaft
Copy link
Contributor Author

Sorry, just fell off my radar. Give me 5min.

@fnothaft fnothaft force-pushed the issues/231-spark-162-java8 branch from 2b042ec to 426b5ab Compare October 21, 2016 15:22
@fnothaft
Copy link
Contributor Author

@jpdna just fixed and force pushed an amended commit. LMK if this works for you!

@jpdna
Copy link

jpdna commented Oct 21, 2016

I was able to launch a spark cluster with cgcloud which was indeed now spark 1.6.2 and java8, so this PR appears to work fine now.

@fnothaft
Copy link
Contributor Author

Great! Thanks for the confirmation, @jpdna.

@hannes-ucsc
Copy link
Contributor

@hannes-ucsc
Copy link
Contributor

Jenkins, test this please.

@hannes-ucsc
Copy link
Contributor

hannes-ucsc commented Nov 9, 2016

I've just triggered another build to see if the test failure is deterministic. It looks like it is:

java.lang.IllegalArgumentException: System memory 64880640 must be at least 4.718592E8. Please use a larger heap size.
    at org.apache.spark.memory.UnifiedMemoryManager$.getMaxMemory(UnifiedMemoryManager.scala:198)
    at org.apache.spark.memory.UnifiedMemoryManager$.apply(UnifiedMemoryManager.scala:180)
    at org.apache.spark.SparkEnv$.create(SparkEnv.scala:354)
    at org.apache.spark.SparkEnv$.createDriverEnv(SparkEnv.scala:193)
    at org.apache.spark.SparkContext.createSparkEnv(SparkContext.scala:288)
    at org.apache.spark.SparkContext.<init>(SparkContext.scala:457)
    at org.apache.spark.api.java.JavaSparkContext.<init>(JavaSparkContext.scala:59)
    at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
    at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
    at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
    at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
    at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:234)
    at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:381)
    at py4j.Gateway.invoke(Gateway.java:214)
    at py4j.commands.ConstructorCommand.invokeConstructor(ConstructorCommand.java:79)
    at py4j.commands.ConstructorCommand.execute(ConstructorCommand.java:68)
    at py4j.GatewayConnection.run(GatewayConnection.java:209)
    at java.lang.Thread.run(Thread.java:745)

@hannes-ucsc
Copy link
Contributor

@jpdna
Copy link

jpdna commented Nov 11, 2016

just pinging on this as its a blocker for docs work

@hannes-ucsc
Copy link
Contributor

I would just remove the passing of --executor-memory from the word count test. The defaults should work fine. Looking at

https://github.com/apache/spark/blob/master/core/src/main/scala/org/apache/spark/memory/UnifiedMemoryManager.scala#L210

@fnothaft fnothaft force-pushed the issues/231-spark-162-java8 branch from 426b5ab to 0006937 Compare November 14, 2016 20:24
@fnothaft
Copy link
Contributor Author

@hannes-ucsc this passes now. Thanks for catching the low driver/executor memory settings!

@hannes-ucsc hannes-ucsc merged commit a564a63 into BD2KGenomics:master Nov 14, 2016
@hannes-ucsc
Copy link
Contributor

Thank you!

@hannes-ucsc
Copy link
Contributor

The master build of the merge commit failed due to a pip outage. I just triggered it again.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants