Skip to content
This repository has been archived by the owner on Dec 20, 2022. It is now read-only.

spark on yarn Compatibility between different versions? #24

Open
chenzhaohangbj opened this issue Feb 18, 2019 · 13 comments
Open

spark on yarn Compatibility between different versions? #24

chenzhaohangbj opened this issue Feb 18, 2019 · 13 comments

Comments

@chenzhaohangbj
Copy link

When I run Spark on yarn
Spark2.1.0
Hadoop2.7.3

nodemanger pull-in spark-2.1.0-yarn-shuffle.jar,but when spark version is not spark-2.1.0,container can not launch .

image

@petro-rudenko
Copy link
Member

Hi, do you use external shuffle service?

@chenzhaohangbj
Copy link
Author

my spark conf:
spark.driver.extraClassPath /home/bigdata/local/spark-rdma-3.1-for-spark-2.1.0-jar-with-dependencies.jar
spark.executor.extraClassPath /home/bigdata/local/spark-rdma-3.1-for-spark-2.1.0-jar-with-dependencies.jar
spark.shuffle.manager org.apache.spark.shuffle.rdma.RdmaShuffleManager
spark.shuffle.compress false
spark.shuffle.spill.compress false
spark.broadcast.compress false
spark.broadcast.checksum false
spark.locality.wait 0

@petro-rudenko
Copy link
Member

So in release tar - there's prebuilded jars for spark versions started from 2.0 to 2.4.

@petro-rudenko
Copy link
Member

Do you try to use with different spark version?

@chenzhaohangbj
Copy link
Author

yes,is not ok.

@petro-rudenko
Copy link
Member

@chenzhaohangbj which spark version and which SparkRDMA jar do you use?

@chenzhaohangbj
Copy link
Author

spark 2.1.0
spark 2.1.1
spark 2.3.0

@petro-rudenko
Copy link
Member

So you need to use:

spark 2.1.0 - spark-rdma-3.1-for-spark-2.1.0-jar-with-dependencies.jar
spark 2.1.1 - spark-rdma-3.1-for-spark-2.1.0-jar-with-dependencies.jar
spark 2.3.0 - spark-rdma-3.1-for-spark-2.3.0-jar-with-dependencies.jar

@chenzhaohangbj
Copy link
Author

nodemanager need which jar?

@petro-rudenko
Copy link
Member

petro-rudenko commented Feb 19, 2019

Nodemanager doesn't need any jar. We don't support external shuffle service yet.

@chenzhaohangbj
Copy link
Author

spark rdma shuffle and spark shuffle can compatible on nodemanager?

@petro-rudenko
Copy link
Member

If you don't use External Yarn Shuffle Service - then nodemanager is used only to launch Spark application. Spark itself will instantiate configured shuffle service. SparkRDMA is fully compatible with default Spark Shuffle.

@ilovesxl
Copy link

I notice that it is hardcoded in https://github.com/apache/spark/blob/master/common/network-shuffle/src/main/java/org/apache/spark/network/shuffle/ExternalShuffleBlockResolver.java:150 that spark external shuffle only support two shufflemanager types.So if I edit the code and compile spark,can the SparkRDMA works with external shuffle enabled?

Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants