/dports/devel/hadoop2/hadoop-2.7.2-src/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/test/java/org/apache/hadoop/lib/service/hadoop/ |
H A D | TestFileSystemAccessService.java | 50 private void createHadoopConf(Configuration hadoopConf) throws Exception { in createHadoopConf() 60 Configuration hadoopConf = new Configuration(false); in createHadoopConf() local 184 Configuration hadoopConf = new Configuration(false); in serviceHadoopConfCustomDir() local 259 Configuration hadoopConf = new Configuration(false); in createFileSystem() local 293 Configuration hadoopConf = new Configuration(false); in fileSystemExecutor() local 334 Configuration hadoopConf = new Configuration(false); in fileSystemExecutorNoNameNode() local 364 Configuration hadoopConf = new Configuration(false); in fileSystemExecutorException() local 411 Configuration hadoopConf = new Configuration(false); in fileSystemCache() local
|
/dports/devel/spark/spark-2.1.1/yarn/src/test/scala/org/apache/spark/deploy/yarn/security/ |
H A D | HDFSCredentialProviderSuite.scala | 56 val hadoopConf = new Configuration() constant 64 val hadoopConf = new Configuration() constant
|
H A D | ConfigurableCredentialManagerSuite.scala | 100 val hadoopConf = new Configuration() constant 113 val hadoopConf = new Configuration() constant
|
/dports/devel/hadoop/hadoop-1.2.1/src/contrib/failmon/src/java/org/apache/hadoop/contrib/failmon/ |
H A D | HDFSMerger.java | 48 Configuration hadoopConf; field in HDFSMerger 74 Configuration hadoopConf = new Configuration(); in HDFSMerger() local
|
H A D | LocalStore.java | 239 Configuration hadoopConf = new Configuration(); in copyToHDFS() local
|
/dports/devel/spark/spark-2.1.1/sql/core/src/main/scala/org/apache/spark/sql/internal/ |
H A D | SessionState.scala | 53 val hadoopConf = new Configuration(sparkSession.sparkContext.hadoopConfiguration) constant 59 val hadoopConf = newHadoopConf() constant
|
/dports/devel/spark/spark-2.1.1/sql/hive/src/test/scala/org/apache/spark/sql/hive/client/ |
H A D | HiveClientSuite.scala | 45 val hadoopConf = new Configuration() constant
|
H A D | VersionsSuite.scala | 62 val hadoopConf = new Configuration() constant 99 val hadoopConf = new Configuration() constant
|
/dports/devel/spark/spark-2.1.1/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/ |
H A D | InsertIntoHadoopFsRelationCommand.scala | 73 val hadoopConf = sparkSession.sessionState.newHadoopConfWithOptions(options) constant
|
H A D | DataSource.scala | 126 val hadoopConf = sparkSession.sessionState.newHadoopConf() constant 369 val hadoopConf = sparkSession.sessionState.newHadoopConf() constant
|
/dports/devel/spark/spark-2.1.1/core/src/main/scala/org/apache/spark/rdd/ |
H A D | PairRDDFunctions.scala | 998 val hadoopConf = conf constant 1039 val hadoopConf = conf constant 1087 val hadoopConf = conf constant 1170 val hadoopConf = conf constant
|
/dports/devel/spark/spark-2.1.1/sql/core/src/test/scala/org/apache/spark/sql/execution/streaming/state/ |
H A D | StateStoreSuite.scala | 342 val hadoopConf = new Configuration() constant 391 val hadoopConf = new Configuration() constant 488 val hadoopConf = new Configuration() constant
|
/dports/devel/spark/spark-2.1.1/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/parquet/ |
H A D | ParquetIOSuite.scala | 283 val hadoopConf = spark.sessionState.newHadoopConf() constant 368 val hadoopConf = spark.sessionState.newHadoopConf() constant 529 val hadoopConf = spark.sessionState.newHadoopConfWithOptions(extraOptions) constant
|
H A D | ParquetCompatibilityTest.scala | 41 val hadoopConf = spark.sessionState.newHadoopConf() constant
|
/dports/devel/spark/spark-2.1.1/streaming/src/main/scala/org/apache/spark/streaming/rdd/ |
H A D | WriteAheadLogBackedBlockRDD.scala | 116 val hadoopConf = broadcastedHadoopConf.value constant
|
/dports/devel/spark/spark-2.1.1/streaming/src/test/scala/org/apache/spark/streaming/rdd/ |
H A D | WriteAheadLogBackedBlockRDDSuite.scala | 40 val hadoopConf = new Configuration() constant
|
/dports/devel/spark/spark-2.1.1/mllib/src/test/scala/org/apache/spark/mllib/impl/ |
H A D | PeriodicRDDCheckpointerSuite.scala | 130 val hadoopConf = rdd.sparkContext.hadoopConfiguration constant
|
H A D | PeriodicGraphCheckpointerSuite.scala | 143 val hadoopConf = graph.vertices.sparkContext.hadoopConfiguration constant
|
/dports/devel/spark/spark-2.1.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/ |
H A D | MetastoreRelation.scala | 135 val hadoopConf = sparkSession.sessionState.newHadoopConf() constant
|
/dports/devel/spark/spark-2.1.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/client/ |
H A D | IsolatedClientLoader.scala | 157 val hadoopConf: Configuration, constant
|
/dports/devel/spark/spark-2.1.1/sql/core/src/main/scala/org/apache/spark/sql/execution/command/ |
H A D | ddl.scala | 509 val hadoopConf = spark.sparkContext.hadoopConfiguration constant 587 val hadoopConf = spark.sparkContext.hadoopConfiguration constant
|
/dports/devel/spark/spark-2.1.1/sql/core/src/main/scala/org/apache/spark/sql/execution/streaming/ |
H A D | HDFSMetadataLog.scala | 301 val hadoopConf = sparkSession.sessionState.newHadoopConf() constant
|
H A D | StreamExecution.scala | 104 val hadoopConf = sparkSession.sessionState.newHadoopConf() constant
|
/dports/devel/spark/spark-2.1.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/execution/ |
H A D | InsertIntoHiveTable.scala | 88 val hadoopConf = sessionState.newHadoopConf() constant
|
/dports/devel/spark/spark-2.1.1/streaming/src/test/scala/org/apache/spark/streaming/ |
H A D | ReceivedBlockTrackerSuite.scala | 44 val hadoopConf = new Configuration() constant
|