Replies: 1 comment 3 replies
-
|
I don't think that'll work given how Hadoop works. The Hadoop master is typically configured with a list of all the worker nodes its allowed to schedule jobs on. Magpie would (hypothetically) have to dynamically update that Hadoop master's configuration whenever it spins up (and tears down) a new allocation on an HPC cluster. That is likely impossible [1] (ssh off cluster and have privileged access to re-configure Hadoop) and minimally not reasonable to do. Edit: That said, any job you start on an HPC cluster should be able to access the HDFS networked storage, since that is likely independent of the Hadoop master / Yarn scheduler. [1] - clarification, I should say impossible given the average HPC cluster environment. I can't speak for all environments. |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
If I have an existing Hadoop cluster and HDFS networked storage, can I point to it instead and have all nodes and cores for the job that are running hadoop point to the existing hadoop master node and be worker nodes of it ?
Beta Was this translation helpful? Give feedback.
All reactions