|
75 | 75 | "%env DB_Name = \"\" # Database|Schema Name\n", |
76 | 76 | "%env DB_TABLE = \"\" # Table Name\n", |
77 | 77 | "%env DB_USER = \"\" # Database User Name\n", |
78 | | - "%env DB_PASSWORD = \"\" # Database User's Password" |
| 78 | + "%env DB_PASSWORD = \"\" # Database User's Password\n", |
| 79 | + "\n", |
| 80 | + "os.environ[\"PYSPARK_SUBMIT_ARGS\"] = \"--packages mysql:mysql-connector-java:5.1.39 pyspark-shell\"" |
79 | 81 | ] |
80 | 82 | }, |
81 | 83 | { |
|
156 | 158 | "text": [ |
157 | 159 | "[('spark.sql.catalogImplementation', 'in-memory'),\n", |
158 | 160 | " ('spark.driver.extraLibraryPath', '/hadoop/etc/hadoop'),\n", |
159 | | - " ('spark.jars',\n", |
160 | | - " 'file:///spark/v3io-libs/v3io-hcfs_2.11.jar,file:///spark/v3io-libs/v3io-spark2-object-dataframe_2.11.jar,file:///spark/v3io-libs/v3io-spark2-streaming_2.11.jar'),\n", |
| 161 | + " ('spark.app.id', 'app-20190704070308-0001'),\n", |
161 | 162 | " ('spark.executor.memory', '2G'),\n", |
162 | | - " ('spark.repl.local.jars',\n", |
163 | | - " 'file:///spark/v3io-libs/v3io-hcfs_2.11.jar,file:///spark/v3io-libs/v3io-spark2-object-dataframe_2.11.jar,file:///spark/v3io-libs/v3io-spark2-streaming_2.11.jar'),\n", |
164 | 163 | " ('spark.executor.id', 'driver'),\n", |
165 | | - " ('spark.driver.port', '41461'),\n", |
| 164 | + " ('spark.jars',\n", |
| 165 | + " 'file:///spark/v3io-libs/v3io-hcfs_2.11.jar,file:///spark/v3io-libs/v3io-spark2-object-dataframe_2.11.jar,file:///spark/v3io-libs/v3io-spark2-streaming_2.11.jar,file:///igz/.ivy2/jars/mysql_mysql-connector-java-5.1.39.jar'),\n", |
166 | 166 | " ('spark.cores.max', '4'),\n", |
167 | | - " ('spark.master', 'spark://spark-9nv9ola1rl-3qgje-master:7077'),\n", |
| 167 | + " ('spark.executorEnv.V3IO_ACCESS_KEY', 'bb79fffa-7582-4fd2-9347-a350335801fc'),\n", |
168 | 168 | " ('spark.driver.extraClassPath',\n", |
169 | 169 | " '/spark/3rd_party/mysql-connector-java-8.0.13.jar'),\n", |
170 | 170 | " ('spark.executor.extraJavaOptions', '\"-Dsun.zip.disableMemoryMapping=true\"'),\n", |
| 171 | + " ('spark.driver.port', '33751'),\n", |
| 172 | + " ('spark.driver.host', '10.233.92.91'),\n", |
171 | 173 | " ('spark.executor.extraLibraryPath', '/hadoop/etc/hadoop'),\n", |
| 174 | + " ('spark.submit.pyFiles',\n", |
| 175 | + " '/igz/.ivy2/jars/mysql_mysql-connector-java-5.1.39.jar'),\n", |
172 | 176 | " ('spark.app.name', 'Spark JDBC to Databases - ipynb'),\n", |
173 | | - " ('spark.driver.host', '10.233.92.90'),\n", |
| 177 | + " ('spark.repl.local.jars',\n", |
| 178 | + " 'file:///spark/v3io-libs/v3io-hcfs_2.11.jar,file:///spark/v3io-libs/v3io-spark2-object-dataframe_2.11.jar,file:///spark/v3io-libs/v3io-spark2-streaming_2.11.jar,file:///igz/.ivy2/jars/mysql_mysql-connector-java-5.1.39.jar'),\n", |
174 | 179 | " ('spark.rdd.compress', 'True'),\n", |
175 | | - " ('spark.executorEnv.V3IO_ACCESS_KEY', '07934877-3b89-4f88-b08e-a6ea1fb1092b'),\n", |
176 | | - " ('spark.app.id', 'app-20190430214617-0002'),\n", |
177 | 180 | " ('spark.serializer.objectStreamReset', '100'),\n", |
| 181 | + " ('spark.files',\n", |
| 182 | + " 'file:///igz/.ivy2/jars/mysql_mysql-connector-java-5.1.39.jar'),\n", |
178 | 183 | " ('spark.executor.cores', '1'),\n", |
179 | 184 | " ('spark.executor.extraClassPath',\n", |
180 | 185 | " '/spark/3rd_party/mysql-connector-java-8.0.13.jar'),\n", |
181 | 186 | " ('spark.submit.deployMode', 'client'),\n", |
182 | | - " ('spark.submit.pyFiles', '/igz/java/libs/v3io-py.zip'),\n", |
183 | 187 | " ('spark.driver.extraJavaOptions', '\"-Dsun.zip.disableMemoryMapping=true\"'),\n", |
184 | 188 | " ('spark.ui.showConsoleProgress', 'true'),\n", |
185 | | - " ('spark.executorEnv.V3IO_USERNAME', 'iguazio')]\n" |
| 189 | + " ('spark.executorEnv.V3IO_USERNAME', 'iguazio'),\n", |
| 190 | + " ('spark.master', 'spark://spark-jddcm4iwas-qxw13-master:7077')]\n" |
186 | 191 | ] |
187 | 192 | } |
188 | 193 | ], |
|
279 | 284 | }, |
280 | 285 | { |
281 | 286 | "cell_type": "code", |
282 | | - "execution_count": 5, |
| 287 | + "execution_count": null, |
283 | 288 | "metadata": {}, |
284 | | - "outputs": [ |
285 | | - { |
286 | | - "name": "stdout", |
287 | | - "output_type": "stream", |
288 | | - "text": [ |
289 | | - "+--------+------+-------------------+\n", |
290 | | - "|fruit_id| name| variety|\n", |
291 | | - "+--------+------+-------------------+\n", |
292 | | - "| 1| Apple| Red Delicious|\n", |
293 | | - "| 2| Pear| Comice|\n", |
294 | | - "| 3|Orange| Navel|\n", |
295 | | - "| 4| Pear| Bartlett|\n", |
296 | | - "| 5|Orange| Blood|\n", |
297 | | - "| 6| Apple|Cox's Orange Pippin|\n", |
298 | | - "| 7| Apple| Granny Smith|\n", |
299 | | - "| 8| Pear| Anjou|\n", |
300 | | - "| 9|Orange| Valencia|\n", |
301 | | - "| 10|Banana| Plantain|\n", |
302 | | - "| 11|Banana| Burro|\n", |
303 | | - "| 12|Banana| Cavendish|\n", |
304 | | - "+--------+------+-------------------+\n", |
305 | | - "\n" |
306 | | - ] |
307 | | - } |
308 | | - ], |
| 289 | + "outputs": [], |
309 | 290 | "source": [ |
310 | 291 | "dfMySQL = spark.read \\\n", |
311 | 292 | " .format(\"jdbc\") \\\n", |
|
0 commit comments