Home
last modified time | relevance | path

Searched defs:hadoopConf (Results 1 – 25 of 37) sorted by relevance

12

/dports/devel/hadoop2/hadoop-2.7.2-src/hadoop-hdfs-project/hadoop-hdfs-httpfs/src/test/java/org/apache/hadoop/lib/service/hadoop/
H A DTestFileSystemAccessService.java50 private void createHadoopConf(Configuration hadoopConf) throws Exception { in createHadoopConf()
60 Configuration hadoopConf = new Configuration(false); in createHadoopConf() local
184 Configuration hadoopConf = new Configuration(false); in serviceHadoopConfCustomDir() local
259 Configuration hadoopConf = new Configuration(false); in createFileSystem() local
293 Configuration hadoopConf = new Configuration(false); in fileSystemExecutor() local
334 Configuration hadoopConf = new Configuration(false); in fileSystemExecutorNoNameNode() local
364 Configuration hadoopConf = new Configuration(false); in fileSystemExecutorException() local
411 Configuration hadoopConf = new Configuration(false); in fileSystemCache() local
/dports/devel/spark/spark-2.1.1/yarn/src/test/scala/org/apache/spark/deploy/yarn/security/
H A DHDFSCredentialProviderSuite.scala56 val hadoopConf = new Configuration() constant
64 val hadoopConf = new Configuration() constant
H A DConfigurableCredentialManagerSuite.scala100 val hadoopConf = new Configuration() constant
113 val hadoopConf = new Configuration() constant
/dports/devel/hadoop/hadoop-1.2.1/src/contrib/failmon/src/java/org/apache/hadoop/contrib/failmon/
H A DHDFSMerger.java48 Configuration hadoopConf; field in HDFSMerger
74 Configuration hadoopConf = new Configuration(); in HDFSMerger() local
H A DLocalStore.java239 Configuration hadoopConf = new Configuration(); in copyToHDFS() local
/dports/devel/spark/spark-2.1.1/sql/core/src/main/scala/org/apache/spark/sql/internal/
H A DSessionState.scala53 val hadoopConf = new Configuration(sparkSession.sparkContext.hadoopConfiguration) constant
59 val hadoopConf = newHadoopConf() constant
/dports/devel/spark/spark-2.1.1/sql/hive/src/test/scala/org/apache/spark/sql/hive/client/
H A DHiveClientSuite.scala45 val hadoopConf = new Configuration() constant
H A DVersionsSuite.scala62 val hadoopConf = new Configuration() constant
99 val hadoopConf = new Configuration() constant
/dports/devel/spark/spark-2.1.1/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/
H A DInsertIntoHadoopFsRelationCommand.scala73 val hadoopConf = sparkSession.sessionState.newHadoopConfWithOptions(options) constant
H A DDataSource.scala126 val hadoopConf = sparkSession.sessionState.newHadoopConf() constant
369 val hadoopConf = sparkSession.sessionState.newHadoopConf() constant
/dports/devel/spark/spark-2.1.1/core/src/main/scala/org/apache/spark/rdd/
H A DPairRDDFunctions.scala998 val hadoopConf = conf constant
1039 val hadoopConf = conf constant
1087 val hadoopConf = conf constant
1170 val hadoopConf = conf constant
/dports/devel/spark/spark-2.1.1/sql/core/src/test/scala/org/apache/spark/sql/execution/streaming/state/
H A DStateStoreSuite.scala342 val hadoopConf = new Configuration() constant
391 val hadoopConf = new Configuration() constant
488 val hadoopConf = new Configuration() constant
/dports/devel/spark/spark-2.1.1/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/parquet/
H A DParquetIOSuite.scala283 val hadoopConf = spark.sessionState.newHadoopConf() constant
368 val hadoopConf = spark.sessionState.newHadoopConf() constant
529 val hadoopConf = spark.sessionState.newHadoopConfWithOptions(extraOptions) constant
H A DParquetCompatibilityTest.scala41 val hadoopConf = spark.sessionState.newHadoopConf() constant
/dports/devel/spark/spark-2.1.1/streaming/src/main/scala/org/apache/spark/streaming/rdd/
H A DWriteAheadLogBackedBlockRDD.scala116 val hadoopConf = broadcastedHadoopConf.value constant
/dports/devel/spark/spark-2.1.1/streaming/src/test/scala/org/apache/spark/streaming/rdd/
H A DWriteAheadLogBackedBlockRDDSuite.scala40 val hadoopConf = new Configuration() constant
/dports/devel/spark/spark-2.1.1/mllib/src/test/scala/org/apache/spark/mllib/impl/
H A DPeriodicRDDCheckpointerSuite.scala130 val hadoopConf = rdd.sparkContext.hadoopConfiguration constant
H A DPeriodicGraphCheckpointerSuite.scala143 val hadoopConf = graph.vertices.sparkContext.hadoopConfiguration constant
/dports/devel/spark/spark-2.1.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/
H A DMetastoreRelation.scala135 val hadoopConf = sparkSession.sessionState.newHadoopConf() constant
/dports/devel/spark/spark-2.1.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/client/
H A DIsolatedClientLoader.scala157 val hadoopConf: Configuration, constant
/dports/devel/spark/spark-2.1.1/sql/core/src/main/scala/org/apache/spark/sql/execution/command/
H A Dddl.scala509 val hadoopConf = spark.sparkContext.hadoopConfiguration constant
587 val hadoopConf = spark.sparkContext.hadoopConfiguration constant
/dports/devel/spark/spark-2.1.1/sql/core/src/main/scala/org/apache/spark/sql/execution/streaming/
H A DHDFSMetadataLog.scala301 val hadoopConf = sparkSession.sessionState.newHadoopConf() constant
H A DStreamExecution.scala104 val hadoopConf = sparkSession.sessionState.newHadoopConf() constant
/dports/devel/spark/spark-2.1.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/execution/
H A DInsertIntoHiveTable.scala88 val hadoopConf = sessionState.newHadoopConf() constant
/dports/devel/spark/spark-2.1.1/streaming/src/test/scala/org/apache/spark/streaming/
H A DReceivedBlockTrackerSuite.scala44 val hadoopConf = new Configuration() constant

12