-
Notifications
You must be signed in to change notification settings - Fork 118
Ship Hadoop configuration files to the driver and add to its classpath #130
Comments
@kimoonkim as you've been testing the HDFS locality changes recently, are you passing the |
@ash211 Good question. For the HDFS node-level locality tests so far, Spark only needed the namenode address. I passed it as But we plan to work on the rack locality part soon, which involves more. Spark driver needs several config keys for the rack topology plugin. It also needs to access a script or text file that topology plugins refer to. (There are multiple topology plugin choices) Those files are usually in the same hadoop conf dir. So It would be better to pass |
We can build a ConfigMap instance, or allow the user to specify an existing one, which contains the Depending on whether or not we expect |
In what cases would we see |
One case where the XML files might have sensitive data is when configuring Spark to communicate with S3. In those cases these XML files might contain AWS credentials. |
If the user isn't specifying an existing ConfigMap how should the user be expected to specify the file locations which the ConfigMap will use in the creation step via the Submission Client. |
The submission client can set it for the user and just set |
#373 Should handle this |
Currently we do not do this, so the only way Hadoop configuration options can be set is by setting
spark.hadoop.*
parameters on the Spark configuration.The text was updated successfully, but these errors were encountered: