Add any user credentials to the job conf which are necessary for running on a secure Hadoop cluster.
Add any user credentials to the job conf which are necessary for running on a secure Hadoop cluster.
Appends S3-specific, spark.hadoop.*, and spark.buffer.size configurations to a Hadoop configuration.
Appends S3-specific, spark.hadoop.*, and spark.buffer.size configurations to a Hadoop configuration.
Lists all the files in a directory with the specified prefix, and does not end with the given suffix.
Lists all the files in a directory with the specified prefix, and does not end with the given suffix. The returned {{FileStatus}} instances are sorted by the modification times of the respective files.
Get FileStatus objects for all leaf children (files) under the given base path.
Get FileStatus objects for all leaf children (files) under the given base path. If the given path points to a file, return a single-element collection containing FileStatus of that file.
Get FileStatus objects for all leaf children (files) under the given base path.
Get FileStatus objects for all leaf children (files) under the given base path. If the given path points to a file, return a single-element collection containing FileStatus of that file.
Return an appropriate (subclass) of Configuration.
Return an appropriate (subclass) of Configuration. Creating config can initializes some Hadoop subsystems.
Runs the given function with a Hadoop UserGroupInformation as a thread local variable (distributed to child threads), used for authenticating HDFS and YARN calls.
Runs the given function with a Hadoop UserGroupInformation as a thread local variable (distributed to child threads), used for authenticating HDFS and YARN calls.
IMPORTANT NOTE: If this function is going to be called repeated in the same process you need to look https://issues.apache.org/jira/browse/HDFS-3545 and possibly do a FileSystem.closeAllForUGI in order to avoid leaking Filesystems
Substitute variables by looking them up in Hadoop configs.
Substitute variables by looking them up in Hadoop configs. Only variables that match the ${hadoopconf- .. } pattern are substituted.
Contains util methods to interact with Hadoop from spark.