105105USER_SCRIPTS_PATH = os .path .join (SPARK_HOME , "sbin" )
106106DATA_PATH = os .path .join (SPARK_HOME , "data" )
107107LICENSES_PATH = os .path .join (SPARK_HOME , "licenses" )
108+ DOCKER_PATH = os .path .join (SPARK_HOME , "resource-managers/kubernetes/docker/src/main/dockerfiles/spark" )
108109
109110SCRIPTS_TARGET = os .path .join (TEMP_PATH , "bin" )
110111USER_SCRIPTS_TARGET = os .path .join (TEMP_PATH , "sbin" )
111112JARS_TARGET = os .path .join (TEMP_PATH , "jars" )
112113EXAMPLES_TARGET = os .path .join (TEMP_PATH , "examples" )
113114DATA_TARGET = os .path .join (TEMP_PATH , "data" )
114115LICENSES_TARGET = os .path .join (TEMP_PATH , "licenses" )
116+ DOCKER_TARGET = os .path .join (TEMP_PATH , "k8s" )
115117
116118# Check and see if we are under the spark path in which case we need to build the symlink farm.
117119# This is important because we only want to build the symlink farm while under Spark otherwise we
@@ -221,6 +223,7 @@ def run(self):
221223 os .symlink (EXAMPLES_PATH , EXAMPLES_TARGET )
222224 os .symlink (DATA_PATH , DATA_TARGET )
223225 os .symlink (LICENSES_PATH , LICENSES_TARGET )
226+ os .symlink (DOCKER_PATH , DOCKER_TARGET )
224227 else :
225228 # For windows fall back to the slower copytree
226229 copytree (JARS_PATH , JARS_TARGET )
@@ -229,6 +232,7 @@ def run(self):
229232 copytree (EXAMPLES_PATH , EXAMPLES_TARGET )
230233 copytree (DATA_PATH , DATA_TARGET )
231234 copytree (LICENSES_PATH , LICENSES_TARGET )
235+ copytree (DOCKER_PATH , DOCKER_TARGET )
232236 else :
233237 # If we are not inside of SPARK_HOME verify we have the required symlink farm
234238 if not os .path .exists (JARS_TARGET ):
@@ -296,6 +300,7 @@ def run(self):
296300 "pyspark.streaming" ,
297301 "pyspark.bin" ,
298302 "pyspark.sbin" ,
303+ "pyspark.k8s" ,
299304 "pyspark.jars" ,
300305 "pyspark.pandas" ,
301306 "pyspark.pandas.data_type_ops" ,
@@ -321,6 +326,7 @@ def run(self):
321326 "pyspark.jars" : "deps/jars" ,
322327 "pyspark.bin" : "deps/bin" ,
323328 "pyspark.sbin" : "deps/sbin" ,
329+ "pyspark.k8s" : "deps/k8s" ,
324330 "pyspark.python.lib" : "lib" ,
325331 "pyspark.data" : "deps/data" ,
326332 "pyspark.licenses" : "deps/licenses" ,
@@ -329,6 +335,7 @@ def run(self):
329335 package_data = {
330336 "pyspark.jars" : ["*.jar" ],
331337 "pyspark.bin" : ["*" ],
338+ "pyspark.k8s" : ["*" ],
332339 "pyspark.sbin" : [
333340 "spark-config.sh" ,
334341 "spark-daemon.sh" ,
@@ -398,11 +405,13 @@ def run(self):
398405 os .remove (os .path .join (TEMP_PATH , "examples" ))
399406 os .remove (os .path .join (TEMP_PATH , "data" ))
400407 os .remove (os .path .join (TEMP_PATH , "licenses" ))
408+ os .remove (os .path .join (TEMP_PATH , "k8s" ))
401409 else :
402410 rmtree (os .path .join (TEMP_PATH , "jars" ))
403411 rmtree (os .path .join (TEMP_PATH , "bin" ))
404412 rmtree (os .path .join (TEMP_PATH , "sbin" ))
405413 rmtree (os .path .join (TEMP_PATH , "examples" ))
406414 rmtree (os .path .join (TEMP_PATH , "data" ))
407415 rmtree (os .path .join (TEMP_PATH , "licenses" ))
416+ rmtree (os .path .join (TEMP_PATH , "k8s" ))
408417 os .rmdir (TEMP_PATH )
0 commit comments