You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I got the error below while executing the "Run your app" step in the .NET for Apache Spark Tutorial. I am running this on
Operating System: windows 10
VSCode : 1.52.1
Apache Spark: spark-3.0.1-bin-hadoop3.2
Code snippet:
PS C:\Users\Theo\Documents\mySparkApp> spark-submit --class org.apache.spark.deploy.dotnet.DotnetRunner --master local bin\Debug\netcoreapp3.1\microsoft-spark-3-0_2.12-1.0.0.jar dotnet bin\Debug\netcoreapp3.1\mySparkApp.dll
spark-submit : The term 'spark-submit' is not recognized as the name of a cmdlet, function, script file, or operable program. Check the spelling of the name, or if a path was included, verify that the path is correct and try again.
At line:1 char:1
+ spark-submit --class org.apache.spark.deploy.dotnet.DotnetRunner --ma ...
+ ~~~~~~~~~~~~
+ CategoryInfo : ObjectNotFound: (spark-submit:String) [], CommandNotFoundException
+ FullyQualifiedErrorId : CommandNotFoundException
I've set up all the environment variables correctly as directed (i.e. I included the additional env variable DOTNET_ASSEMBLY_SEARCH_PATHS mentioned at 6:13-7:14 of the video).
So I'm really not sure what's causing the error now. I've seen issue#532, issue#276 and issue#268 but still no joy.
Could it be that I'm running spark-3.0.1-bin-hadoop3.2 instead of spark-3.0.1-bin-hadoop2.7 as in the tutorial?
The text was updated successfully, but these errors were encountered:
Verified that the jar exists (Indeed it does exist -> see attached screenshot)
Checked the NuGet version installed (i.e. I installed Microsoft.Spark 1.0.0)
I've also tried running the app by prepending %SPARK_HOME%\bin\ before the given command as below:
C:\Windows\System32\mySparkApp>%SPARK_HOME%\bin\spark-submit --class org.apache.spark.deploy.dotnet.DotnetRunner --master local bin\Debug\netcoreapp3.1\microsoft-spark-3-0_2.12-1.0.0.jar dotnet bin\Debug\netcoreapp3.1\mySparkApp.dll
21/06/08 14:36:35 WARN NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
Using Spark's default log4j profile: org/apache/spark/log4j-defaults.properties
21/06/08 14:36:35 WARN DependencyUtils: Local jar C:\Windows\System32\mySparkApp\bin\Debug\netcoreapp3.1\microsoft-spark-3-0_2.12-1.0.0.jar does not exist, skipping.
Error: Failed to load class org.apache.spark.deploy.dotnet.DotnetRunner.
21/06/08 14:36:35 INFO ShutdownHookManager: Shutdown hook called
21/06/08 14:36:35 INFO ShutdownHookManager: Deleting directory C:\Users\A241124\AppData\Local\Temp\spark-eb737269-81b9-49aa-aede-1807d48d843e
Problem encountered on https://dotnet.microsoft.com/learn/data/spark-tutorial/run
Hi
I got the error below while executing the "Run your app" step in the .NET for Apache Spark Tutorial. I am running this on
Operating System: windows 10
VSCode : 1.52.1
Apache Spark: spark-3.0.1-bin-hadoop3.2
Code snippet:
PS C:\Users\Theo\Documents\mySparkApp> spark-submit --class org.apache.spark.deploy.dotnet.DotnetRunner --master local bin\Debug\netcoreapp3.1\microsoft-spark-3-0_2.12-1.0.0.jar dotnet bin\Debug\netcoreapp3.1\mySparkApp.dll
I've set up all the environment variables correctly as directed (i.e. I included the additional env variable DOTNET_ASSEMBLY_SEARCH_PATHS mentioned at 6:13-7:14 of the video).
So I'm really not sure what's causing the error now. I've seen issue#532, issue#276 and issue#268 but still no joy.
Could it be that I'm running spark-3.0.1-bin-hadoop3.2 instead of spark-3.0.1-bin-hadoop2.7 as in the tutorial?
The text was updated successfully, but these errors were encountered: