-
When using remote machine/s for workers in a stand-alone setup, the RAPIDS jar file is downloaded to the $SPARK_HOME/work/app-xxx folder and when the client disconnects, the jar file is not deleted. For a machine with a multi GPU setup, this builds up to a lot of disk space wasted. Steps/Code to reproduce bug I would have expected the temporary binaries to be cleaned upon disconnection. `[eyal.h@kubegpu00139 work]$ ls -lR ./app-20220524185015-0036: ./app-20220524185015-0036/0: ./app-20220524185015-0036/1: ./app-20220524185015-0036/2: ./app-20220524185313-0037: ./app-20220524185313-0037/0: ./app-20220524185313-0037/1: ./app-20220524185313-0037/2: |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment
-
this is not a Spark Rapids issue, this is just an apache Spark and cluster configuration issue. All Apache Spark standalone jobs will download necessary jars to the workers. See the documentation: For instance it has configurations like:
|
Beta Was this translation helpful? Give feedback.
this is not a Spark Rapids issue, this is just an apache Spark and cluster configuration issue. All Apache Spark standalone jobs will download necessary jars to the workers.
See the documentation:
https://spark.apache.org/docs/latest/spark-standalone.html
For instance it has configurations like:
spark.worker.cleanup.enabled