From be1517e9a1cde0c6d69071b2a64d68e4e5fa6a1a Mon Sep 17 00:00:00 2001 From: Chao Sun Date: Fri, 9 Oct 2020 11:02:09 -0700 Subject: [PATCH] Fix a silly mistake --- .../main/scala/org/apache/spark/util/HadoopFSUtils.scala | 7 ++++--- 1 file changed, 4 insertions(+), 3 deletions(-) diff --git a/core/src/main/scala/org/apache/spark/util/HadoopFSUtils.scala b/core/src/main/scala/org/apache/spark/util/HadoopFSUtils.scala index 7c3c2112f1ab7..a3a528cddee37 100644 --- a/core/src/main/scala/org/apache/spark/util/HadoopFSUtils.scala +++ b/core/src/main/scala/org/apache/spark/util/HadoopFSUtils.scala @@ -65,8 +65,8 @@ private[spark] object HadoopFSUtils extends Logging { ignoreLocality: Boolean, parallelismThreshold: Int, parallelismMax: Int): Seq[(Path, Seq[FileStatus])] = { - parallelListLeafFilesInternal(sc, paths, hadoopConf, filter, true, ignoreMissingFiles, - ignoreLocality, parallelismThreshold, parallelismMax) + parallelListLeafFilesInternal(sc, paths, hadoopConf, filter, isRootLevel = true, + ignoreMissingFiles, ignoreLocality, parallelismThreshold, parallelismMax) } private def parallelListLeafFilesInternal( @@ -212,11 +212,12 @@ private[spark] object HadoopFSUtils extends Logging { val (dirs, topLevelFiles) = statuses.partition(_.isDirectory) val nestedFiles: Seq[FileStatus] = contextOpt match { case Some(context) if dirs.size > parallelismThreshold => - parallelListLeafFiles( + parallelListLeafFilesInternal( context, dirs.map(_.getPath), hadoopConf = hadoopConf, filter = filter, + isRootLevel = false, ignoreMissingFiles = ignoreMissingFiles, ignoreLocality = ignoreLocality, parallelismThreshold = parallelismThreshold,