/dports/devel/spark/spark-2.1.1/sql/core/src/test/scala/org/apache/spark/sql/ |
H A D | SparkSessionBuilderSuite.scala | 27 private var initialSession: SparkSession = _ 30 initialSession = SparkSession.builder() 43 SparkSession.clearDefaultSession() 47 val session = SparkSession.builder().getOrCreate() 49 SparkSession.clearDefaultSession() 58 SparkSession.clearDefaultSession() 64 SparkSession.setActiveSession(activeSession) 70 SparkSession.clearActiveSession() 73 SparkSession.clearDefaultSession() 78 SparkSession.setDefaultSession(defaultSession) [all …]
|
H A D | LocalSparkSession.scala | 28 @transient var spark: SparkSession = _ 51 def stop(spark: SparkSession) { 60 def withSparkSession[T](sc: SparkSession)(f: SparkSession => T): T = {
|
/dports/devel/spark/spark-2.1.1/sql/core/src/main/scala/org/apache/spark/sql/execution/command/ |
H A D | SetCommand.scala | 21 import org.apache.spark.sql.{Row, SparkSession} 44 private val (_output, runFunc): (Seq[Attribute], SparkSession => Seq[Row]) = kv match { 47 val runFunc = (sparkSession: SparkSession) => { 64 val runFunc = (sparkSession: SparkSession) => { 72 val runFunc = (sparkSession: SparkSession) => { 81 val runFunc = (sparkSession: SparkSession) => { 89 val runFunc = (sparkSession: SparkSession) => { 102 val runFunc = (sparkSession: SparkSession) => { 114 val runFunc = (sparkSession: SparkSession) => { 123 override def run(sparkSession: SparkSession): Seq[Row] = runFunc(sparkSession) [all …]
|
H A D | ddl.scala | 63 override def run(sparkSession: SparkSession): Seq[Row] = { 99 override def run(sparkSession: SparkSession): Seq[Row] = { 119 override def run(sparkSession: SparkSession): Seq[Row] = { 143 override def run(sparkSession: SparkSession): Seq[Row] = { 185 override def run(sparkSession: SparkSession): Seq[Row] = { 228 override def run(sparkSession: SparkSession): Seq[Row] = { 256 override def run(sparkSession: SparkSession): Seq[Row] = { 489 override def run(spark: SparkSession): Seq[Row] = { 536 spark: SparkSession, 581 spark: SparkSession, [all …]
|
H A D | resources.scala | 25 import org.apache.spark.sql.{Row, SparkSession} 39 override def run(sparkSession: SparkSession): Seq[Row] = { 49 override def run(sparkSession: SparkSession): Seq[Row] = { 63 override def run(sparkSession: SparkSession): Seq[Row] = { 90 override def run(sparkSession: SparkSession): Seq[Row] = {
|
H A D | cache.scala | 20 import org.apache.spark.sql.{Dataset, Row, SparkSession} 37 override def run(sparkSession: SparkSession): Seq[Row] = { 57 override def run(sparkSession: SparkSession): Seq[Row] = { 73 override def run(sparkSession: SparkSession): Seq[Row] = {
|
/dports/devel/spark/spark-2.1.1/examples/src/main/scala/org/apache/spark/examples/sql/ |
H A D | SQLDataSourceExample.scala | 21 import org.apache.spark.sql.SparkSession 28 val spark = SparkSession 43 private def runBasicDataSourceExample(spark: SparkSession): Unit = { 57 private def runBasicParquetExample(spark: SparkSession): Unit = { 84 private def runParquetSchemaMergingExample(spark: SparkSession): Unit = { 112 private def runJsonDatasetExample(spark: SparkSession): Unit = { 151 private def runJdbcDatasetExample(spark: SparkSession): Unit = {
|
H A D | SparkSQLExample.scala | 25 import org.apache.spark.sql.SparkSession 43 val spark = SparkSession 61 private def runBasicDataFrameExample(spark: SparkSession): Unit = { 165 private def runDatasetCreationExample(spark: SparkSession): Unit = { 195 private def runInferSchemaExample(spark: SparkSession): Unit = { 239 private def runProgrammaticSchemaExample(spark: SparkSession): Unit = {
|
/dports/devel/spark/spark-2.1.1/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/ |
H A D | FileFormat.scala | 43 sparkSession: SparkSession, 53 sparkSession: SparkSession, 63 def supportBatch(sparkSession: SparkSession, dataSchema: StructType): Boolean = { 71 sparkSession: SparkSession, 94 sparkSession: SparkSession, 112 sparkSession: SparkSession, 154 sparkSession: SparkSession,
|
/dports/devel/spark/spark-2.1.1/sql/core/src/main/scala/org/apache/spark/sql/ |
H A D | SparkSession.scala | 73 class SparkSession private( class 110 SparkSession.reflect[SessionState, SparkSession]( 212 def newSession(): SparkSession = { 213 new SparkSession(sparkContext, Some(sharedState)) 253 SparkSession.setActiveSession(this) 267 SparkSession.setActiveSession(this) 718 object SparkSession { object 836 def getOrCreate(): SparkSession = synchronized { 848 SparkSession.synchronized { 877 session = new SparkSession(sparkContext) [all …]
|
/dports/devel/spark/spark-2.1.1/sql/core/src/test/scala/org/apache/spark/sql/execution/ |
H A D | ExchangeCoordinatorSuite.scala | 30 private var originalActiveSparkSession: Option[SparkSession] = _ 34 originalActiveSparkSession = SparkSession.getActiveSession 35 originalInstantiatedSparkSession = SparkSession.getDefaultSession 37 SparkSession.clearActiveSession() 38 SparkSession.clearDefaultSession() 253 f: SparkSession => Unit, 275 val spark = SparkSession.builder 288 val test = { spark: SparkSession => 329 val test = { spark: SparkSession => 380 val test = { spark: SparkSession => [all …]
|
/dports/devel/spark/spark-2.1.1/examples/src/main/java/org/apache/spark/examples/sql/ |
H A D | JavaSQLDataSourceExample.java | 39 import org.apache.spark.sql.SparkSession; 96 SparkSession spark = SparkSession in main() 111 private static void runBasicDataSourceExample(SparkSession spark) { in runBasicDataSourceExample() 127 private static void runBasicParquetExample(SparkSession spark) { in runBasicParquetExample() 156 private static void runParquetSchemaMergingExample(SparkSession spark) { in runParquetSchemaMergingExample() 197 private static void runJsonDatasetExample(SparkSession spark) { in runJsonDatasetExample() 237 private static void runJdbcDatasetExample(SparkSession spark) { in runJdbcDatasetExample()
|
H A D | JavaSparkSQLExample.java | 50 import org.apache.spark.sql.SparkSession; 90 SparkSession spark = SparkSession in main() 105 private static void runBasicDataFrameExample(SparkSession spark) throws AnalysisException { in runBasicDataFrameExample() 207 private static void runDatasetCreationExample(SparkSession spark) { in runDatasetCreationExample() 252 private static void runInferSchemaExample(SparkSession spark) { in runInferSchemaExample() 308 private static void runProgrammaticSchemaExample(SparkSession spark) { in runProgrammaticSchemaExample()
|
/dports/devel/spark/spark-2.1.1/sql/core/src/main/scala/org/apache/spark/sql/api/r/ |
H A D | SQLUtils.scala | 49 enableHiveSupport: Boolean): SparkSession = { 50 val spark = if (SparkSession.hiveClassesArePresent && enableHiveSupport 52 SparkSession.builder().sparkContext(withHiveExternalCatalog(jsc.sc)).getOrCreate() 59 SparkSession.builder().sparkContext(jsc.sc).getOrCreate() 66 spark: SparkSession, 76 def getSessionConf(spark: SparkSession): JMap[String, String] = { 80 def getJavaSparkContext(spark: SparkSession): JavaSparkContext = { 234 sparkSession: SparkSession, 241 sparkSession: SparkSession, 271 def getTables(sparkSession: SparkSession, databaseName: String): DataFrame = { [all …]
|
/dports/devel/spark/spark-2.1.1/mllib/src/test/scala/org/apache/spark/mllib/util/ |
H A D | MLlibTestSparkContext.scala | 26 import org.apache.spark.sql.{SparkSession, SQLContext, SQLImplicits} 30 @transient var spark: SparkSession = _ 36 spark = SparkSession.builder 49 SparkSession.clearActiveSession()
|
/dports/devel/spark/spark-2.1.1/repl/scala-2.11/src/main/scala/org/apache/spark/repl/ |
H A D | Main.scala | 26 import org.apache.spark.sql.SparkSession 40 var sparkSession: SparkSession = _ 74 def createSparkSession(): SparkSession = { 90 val builder = SparkSession.builder.config(conf) 92 if (SparkSession.hiveClassesArePresent) {
|
/dports/devel/spark/spark-2.1.1/python/pyspark/ |
H A D | shell.py | 32 from pyspark.sql import SparkSession, SQLContext 43 spark = SparkSession.builder\ 47 spark = SparkSession.builder.getOrCreate() 49 spark = SparkSession.builder.getOrCreate()
|
/dports/devel/spark/spark-2.1.1/examples/src/main/scala/org/apache/spark/examples/streaming/ |
H A D | SqlNetworkWordCount.scala | 23 import org.apache.spark.sql.SparkSession 92 @transient private var instance: SparkSession = _ 94 def getInstance(sparkConf: SparkConf): SparkSession = { 96 instance = SparkSession
|
/dports/devel/spark/spark-2.1.1/mllib/src/test/scala/org/apache/spark/ml/util/ |
H A D | MLTestingUtils.scala | 29 import org.apache.spark.sql.{DataFrame, Row, SparkSession} 43 spark: SparkSession, 66 spark: SparkSession, 88 def checkNumericTypes[T <: Evaluator](evaluator: T, spark: SparkSession): Unit = { 105 spark: SparkSession, 125 spark: SparkSession, 147 spark: SparkSession, 167 spark: SparkSession, 186 spark: SparkSession,
|
/dports/devel/spark/spark-2.1.1/examples/src/main/java/org/apache/spark/examples/streaming/ |
H A D | JavaSqlNetworkWordCount.java | 31 import org.apache.spark.sql.SparkSession; 84 SparkSession spark = JavaSparkSessionSingleton.getInstance(rdd.context().getConf()); in main() 115 private static transient SparkSession instance = null; 116 public static SparkSession getInstance(SparkConf sparkConf) { in getInstance() 118 instance = SparkSession in getInstance()
|
/dports/devel/spark/spark-2.1.1/sql/core/src/test/scala/org/apache/spark/sql/execution/streaming/state/ |
H A D | StateStoreRDDSuite.scala | 34 import org.apache.spark.sql.SparkSession 58 withSparkSession(SparkSession.builder.config(sparkConf).getOrCreate()) { spark => 82 spark: SparkSession, 91 withSparkSession(SparkSession.builder.config(sparkConf).getOrCreate()) { spark => 98 withSparkSession(SparkSession.builder.config(sparkConf).getOrCreate()) { spark => 104 withSparkSession(SparkSession.builder.config(sparkConf).getOrCreate()) { spark => 152 withSparkSession(SparkSession.builder.config(sparkConf).getOrCreate()) { spark => 183 SparkSession.builder
|
/dports/devel/spark/spark-2.1.1/mllib/src/test/java/org/apache/spark/ |
H A D | SharedSparkSession.java | 27 import org.apache.spark.sql.SparkSession; 31 protected transient SparkSession spark; 36 spark = SparkSession.builder() in setUp()
|
/dports/devel/spark/spark-2.1.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/ |
H A D | HiveContext.scala | 23 import org.apache.spark.sql.{SparkSession, SQLContext} 31 class HiveContext private[hive](_sparkSession: SparkSession) 37 this(SparkSession.builder().sparkContext(HiveUtils.withHiveExternalCatalog(sc)).getOrCreate())
|
/dports/devel/spark/spark-2.1.1/examples/src/main/java/org/apache/spark/examples/ml/ |
H A D | JavaMulticlassLogisticRegressionWithElasticNetExample.java | 25 import org.apache.spark.sql.SparkSession; 30 SparkSession spark = SparkSession in main()
|
H A D | JavaStandardScalerExample.java | 20 import org.apache.spark.sql.SparkSession; 31 SparkSession spark = SparkSession in main()
|