Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[SPARK-22769][CORE] When driver stopping, there is errors: Could not find CoarseGrainedScheduler and RpcEnv already stopped #19968

Closed
wants to merge 3 commits into from

Conversation

KaiXinXiaoLei
Copy link

@KaiXinXiaoLei KaiXinXiaoLei commented Dec 13, 2017

What changes were proposed in this pull request?

When driver stopping, there is a error:

17/12/12 18:30:16 INFO MapOutputTrackerMasterEndpoint: MapOutputTrackerMasterEndpoint stopped!
17/12/12 18:30:16 ERROR TransportRequestHandler: Error while invoking RpcHandler#receive() for one-way message.
org.apache.spark.SparkException: Could not find CoarseGrainedScheduler.
at org.apache.spark.rpc.netty.Dispatcher.postMessage(Dispatcher.scala:154)
at org.apache.spark.rpc.netty.Dispatcher.postOneWayMessage(Dispatcher.scala:134)
at org.apache.spark.rpc.netty.NettyRpcHandler.receive(NettyRpcEnv.scala:570)
at org.apache.spark.network.server.TransportRequestHandler.processOneWayMessage(TransportRequestHandler.java:180)
at org.apache.spark.network.server.TransportRequestHandler.handle(TransportRequestHandler.java:109)
at org.apache.spark.network.server.TransportChannelHandler.channelRead0(TransportChannelHandler.java:119)
at org.apache.spark.network.server.TransportChannelHandler.channelRead0(TransportChannelHandler.java:51)

Because the endpoint of CoarseGrainedScheduler is stopped, this error is not problem, so i think this error should not be printed.

And there is another error:

17/12/12 18:20:44 INFO MemoryStore: MemoryStore cleared
17/12/12 18:20:44 INFO BlockManager: BlockManager stopped
17/12/12 18:20:44 INFO BlockManagerMaster: BlockManagerMaster stopped
17/12/12 18:20:44 ERROR TransportRequestHandler: Error while invoking RpcHandler#receive() for one-way message.
org.apache.spark.rpc.RpcEnvStoppedException: RpcEnv already stopped.
at org.apache.spark.rpc.netty.Dispatcher.postMessage(Dispatcher.scala:152)
at org.apache.spark.rpc.netty.Dispatcher.postOneWayMessage(Dispatcher.scala:134)
at org.apache.spark.rpc.netty.NettyRpcHandler.receive(NettyRpcEnv.scala:570)

I think the log level should be warning, not error.

How was this patch tested?

(Please explain how this patch was tested. E.g. unit tests, integration tests, manual tests)
(If this patch involves UI changes, please attach a screenshot; otherwise, remove this)

Please review http://spark.apache.org/contributing.html before opening a pull request.

@srowen
Copy link
Member

srowen commented Dec 13, 2017

You already had another PR open for the same issue, which was an exact duplicate of another PR, for another issue that is a duplicate. Please clean it up into one issue -- SPAKR-22769

@KaiXinXiaoLei KaiXinXiaoLei changed the title [SPARK-22770][CORE] When driver stopping, there is error: Could not find CoarseGrainedScheduler [SPARK-22770][CORE] When driver stopping, there is error: Could not find CoarseGrainedScheduler and RpcEnv already stopped Dec 13, 2017
@KaiXinXiaoLei
Copy link
Author

@srowen i close #19965, and update description

@KaiXinXiaoLei KaiXinXiaoLei changed the title [SPARK-22770][CORE] When driver stopping, there is error: Could not find CoarseGrainedScheduler and RpcEnv already stopped [SPARK-22769][CORE] When driver stopping, there is errors: Could not find CoarseGrainedScheduler and RpcEnv already stopped Dec 13, 2017
@SparkQA
Copy link

SparkQA commented Dec 13, 2017

Test build #84855 has finished for PR 19968 at commit 8a52dc5.

  • This patch fails Spark unit tests.
  • This patch merges cleanly.
  • This patch adds no public classes.

@SparkQA
Copy link

SparkQA commented Dec 13, 2017

Test build #84856 has finished for PR 19968 at commit af7a021.

  • This patch fails Spark unit tests.
  • This patch merges cleanly.
  • This patch adds no public classes.

@SparkQA
Copy link

SparkQA commented Dec 13, 2017

Test build #84867 has finished for PR 19968 at commit 39bc683.

  • This patch fails Spark unit tests.
  • This patch merges cleanly.
  • This patch adds no public classes.

@@ -100,6 +102,7 @@ private[netty] class Dispatcher(nettyEnv: NettyRpcEnv, numUsableCores: Int) exte
return
}
unregisterRpcEndpoint(rpcEndpointRef.name)
endpointsIsStopped.putIfAbsent(rpcEndpointRef.name, true)
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

this will be a memory leak. Spark may create some temp RpcEndpoints. You can make postMessage return Some(RpcEndpointNotFoundException) and ignore it.

Copy link
Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

ok thanks

@srowen
Copy link
Member

srowen commented Dec 31, 2017

Ping @KaiXinXiaoLei to update or close

@KaiXinXiaoLei
Copy link
Author

@srowen ok ,i will update, thanks

@jiangxb1987
Copy link
Contributor

ping @KaiXinXiaoLei are you still working on this?

@KaiXinXiaoLei
Copy link
Author

Now this problem, i don't work. Now i close it .

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
6 participants