You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: docs/deploy-worker-udf-binaries.md
+9-10Lines changed: 9 additions & 10 deletions
Original file line number
Diff line number
Diff line change
@@ -16,7 +16,7 @@ When deploying workers and writing UDFs, there are a few commonly used environme
16
16
</tr>
17
17
<tr>
18
18
<td><b>DOTNET_WORKER_DIR</b></td>
19
-
<td>Path where the <code>Microsoft.Spark.Worker</code> binary has been generated.</br>It's used by the Spark driver and will be passed to Spark executors. If this variable is not set up, the Spark executors will search the path specified in the <code>PATH</code> environment variable.</br><i>e.g. "C:\bin\Microsoft.Spark.Worker-0.6.0"</i></td>
19
+
<td>Path where the <code>Microsoft.Spark.Worker</code> binary has been generated.</br>It's used by the Spark driver and will be passed to Spark executors. If this variable is not set up, the Spark executors will search the path specified in the <code>PATH</code> environment variable.</br><i>e.g. "C:\bin\Microsoft.Spark.Worker"</i></td>
20
20
</tr>
21
21
<tr>
22
22
<td><b>DOTNET_ASSEMBLY_SEARCH_PATHS</b></td>
@@ -50,7 +50,7 @@ Once the Spark application is [bundled](https://spark.apache.org/docs/latest/sub
50
50
</tr>
51
51
<tr>
52
52
<td><b>--conf</b></td>
53
-
<td>Arbitrary Spark configuration property in <code>key=value</code> format.</br><i>e.g. spark.yarn.appMasterEnv.DOTNET_WORKER_DIR=.\worker\Microsoft.Spark.Worker-0.6.0</i></td>
53
+
<td>Arbitrary Spark configuration property in <code>key=value</code> format.</br><i>e.g. spark.yarn.appMasterEnv.DOTNET_WORKER_DIR=.\worker\Microsoft.Spark.Worker</i></td>
54
54
</tr>
55
55
<tr>
56
56
<td><b>--files</b></td>
@@ -68,12 +68,12 @@ Once the Spark application is [bundled](https://spark.apache.org/docs/latest/sub
68
68
<li>Please note that this option is only applicable for yarn mode.</li>
69
69
<li>It supports specifying file names with # similar to Hadoop.</br>
70
70
</ul>
71
-
<i>e.g. <code>hdfs://<path to your worker file>/Microsoft.Spark.Worker.net461.win-x64-0.6.0.zip#worker</code>. This will copy and extract the zip file to <code>worker</code> folder.</i></li></td>
71
+
<i>e.g. <code>hdfs://<path to your worker file>/Microsoft.Spark.Worker.zip#worker</code>. This will copy and extract the zip file to <code>worker</code> folder.</i></li></td>
72
72
</tr>
73
73
<tr>
74
74
<td><b>application-jar</b></td>
75
75
<td>Path to a bundled jar including your application and all dependencies.</br>
76
-
<i>e.g. hdfs://<path to your jar>/microsoft-spark-2.4.x-0.6.0.jar</i></td>
76
+
<i>e.g. hdfs://<path to your jar>/microsoft-spark-<version>.jar</i></td>
77
77
</tr>
78
78
<tr>
79
79
<td><b>application-arguments</b></td>
@@ -95,10 +95,9 @@ Once the Spark application is [bundled](https://spark.apache.org/docs/latest/sub
95
95
**Answer:** Please try restarting your PowerShell window (or other command windows) first so that it can take the latest environment variable values. Then start your program.
96
96
97
97
#### 3. Question: After submitting my Spark application, I get the error `System.TypeLoadException: Could not load type 'System.Runtime.Remoting.Contexts.Context'`.
98
-
> **Command:** %SPARK_HOME%\bin\spark-submit --class org.apache.spark.deploy.dotnet.DotnetRunner --master local microsoft-spark-2.4.x-0.6.0.jar mySparkApp.exe</br>
99
-
**Error:**[][][Error][TaskRunner][0] ProcessStream() failed with exception: System.TypeLoadException: Could not load type 'System.Runtime.Remoting.Contexts.Context' from assembly 'mscorlib, Version=4.0.0.0, Culture=neutral, PublicKeyToken=...'.
98
+
> **Error:**[][][Error][TaskRunner][0] ProcessStream() failed with exception: System.TypeLoadException: Could not load type 'System.Runtime.Remoting.Contexts.Context' from assembly 'mscorlib, Version=4.0.0.0, Culture=neutral, PublicKeyToken=...'.
100
99
101
-
**Answer:** Please check the `Microsoft.Spark.Worker` version you are using. We currently provide two versions: **.NET Framework 4.6.1** and **.NET Core 2.1.x**. In this case, `Microsoft.Spark.Worker.net461.win-x64-0.6.0` (which you can download [here](https://github.com/dotnet/spark/releases)) should be used since `System.Runtime.Remoting.Contexts.Context` is only for .NET Framework.
100
+
**Answer:** Please check the `Microsoft.Spark.Worker` version you are using. We currently provide two versions: **.NET Framework 4.6.1** and **.NET Core 2.1.x**. In this case, `Microsoft.Spark.Worker.net461.win-x64-<version>` (which you can download [here](https://github.com/dotnet/spark/releases)) should be used since `System.Runtime.Remoting.Contexts.Context` is only for .NET Framework.
102
101
103
102
#### 4. Question: How to run my spark application with UDFs on YARN? Which environment variables and parameters should I use?
* The following APIs have been removed due to the thread-local variable dependency (see [#332](https://github.com/dotnet/spark/pull/332) and [#333](https://github.com/dotnet/spark/issues/333) for more detail):
17
+
*`SparkSession.ClearActiveSession()`
18
+
*`SparkSession.GetActiveSession()`
19
+
*`SparkSession.SetActiveSession()`
20
+
21
+
### Supported Spark Versions
22
+
23
+
The following table outlines the supported Spark versions along with the microsoft-spark JAR to use with:
0 commit comments