Home
last modified time | relevance | path

Searched refs:SparkSession (Results 1 – 25 of 486) sorted by relevance

12345678910>>...20

/dports/devel/spark/spark-2.1.1/sql/core/src/test/scala/org/apache/spark/sql/
H A DSparkSessionBuilderSuite.scala27 private var initialSession: SparkSession = _
30 initialSession = SparkSession.builder()
43 SparkSession.clearDefaultSession()
47 val session = SparkSession.builder().getOrCreate()
49 SparkSession.clearDefaultSession()
58 SparkSession.clearDefaultSession()
64 SparkSession.setActiveSession(activeSession)
70 SparkSession.clearActiveSession()
73 SparkSession.clearDefaultSession()
78 SparkSession.setDefaultSession(defaultSession)
[all …]
H A DLocalSparkSession.scala28 @transient var spark: SparkSession = _
51 def stop(spark: SparkSession) {
60 def withSparkSession[T](sc: SparkSession)(f: SparkSession => T): T = {
/dports/devel/spark/spark-2.1.1/sql/core/src/main/scala/org/apache/spark/sql/execution/command/
H A DSetCommand.scala21 import org.apache.spark.sql.{Row, SparkSession}
44 private val (_output, runFunc): (Seq[Attribute], SparkSession => Seq[Row]) = kv match {
47 val runFunc = (sparkSession: SparkSession) => {
64 val runFunc = (sparkSession: SparkSession) => {
72 val runFunc = (sparkSession: SparkSession) => {
81 val runFunc = (sparkSession: SparkSession) => {
89 val runFunc = (sparkSession: SparkSession) => {
102 val runFunc = (sparkSession: SparkSession) => {
114 val runFunc = (sparkSession: SparkSession) => {
123 override def run(sparkSession: SparkSession): Seq[Row] = runFunc(sparkSession)
[all …]
H A Dddl.scala63 override def run(sparkSession: SparkSession): Seq[Row] = {
99 override def run(sparkSession: SparkSession): Seq[Row] = {
119 override def run(sparkSession: SparkSession): Seq[Row] = {
143 override def run(sparkSession: SparkSession): Seq[Row] = {
185 override def run(sparkSession: SparkSession): Seq[Row] = {
228 override def run(sparkSession: SparkSession): Seq[Row] = {
256 override def run(sparkSession: SparkSession): Seq[Row] = {
489 override def run(spark: SparkSession): Seq[Row] = {
536 spark: SparkSession,
581 spark: SparkSession,
[all …]
H A Dresources.scala25 import org.apache.spark.sql.{Row, SparkSession}
39 override def run(sparkSession: SparkSession): Seq[Row] = {
49 override def run(sparkSession: SparkSession): Seq[Row] = {
63 override def run(sparkSession: SparkSession): Seq[Row] = {
90 override def run(sparkSession: SparkSession): Seq[Row] = {
H A Dcache.scala20 import org.apache.spark.sql.{Dataset, Row, SparkSession}
37 override def run(sparkSession: SparkSession): Seq[Row] = {
57 override def run(sparkSession: SparkSession): Seq[Row] = {
73 override def run(sparkSession: SparkSession): Seq[Row] = {
/dports/devel/spark/spark-2.1.1/examples/src/main/scala/org/apache/spark/examples/sql/
H A DSQLDataSourceExample.scala21 import org.apache.spark.sql.SparkSession
28 val spark = SparkSession
43 private def runBasicDataSourceExample(spark: SparkSession): Unit = {
57 private def runBasicParquetExample(spark: SparkSession): Unit = {
84 private def runParquetSchemaMergingExample(spark: SparkSession): Unit = {
112 private def runJsonDatasetExample(spark: SparkSession): Unit = {
151 private def runJdbcDatasetExample(spark: SparkSession): Unit = {
H A DSparkSQLExample.scala25 import org.apache.spark.sql.SparkSession
43 val spark = SparkSession
61 private def runBasicDataFrameExample(spark: SparkSession): Unit = {
165 private def runDatasetCreationExample(spark: SparkSession): Unit = {
195 private def runInferSchemaExample(spark: SparkSession): Unit = {
239 private def runProgrammaticSchemaExample(spark: SparkSession): Unit = {
/dports/devel/spark/spark-2.1.1/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/
H A DFileFormat.scala43 sparkSession: SparkSession,
53 sparkSession: SparkSession,
63 def supportBatch(sparkSession: SparkSession, dataSchema: StructType): Boolean = {
71 sparkSession: SparkSession,
94 sparkSession: SparkSession,
112 sparkSession: SparkSession,
154 sparkSession: SparkSession,
/dports/devel/spark/spark-2.1.1/sql/core/src/main/scala/org/apache/spark/sql/
H A DSparkSession.scala73 class SparkSession private( class
110 SparkSession.reflect[SessionState, SparkSession](
212 def newSession(): SparkSession = {
213 new SparkSession(sparkContext, Some(sharedState))
253 SparkSession.setActiveSession(this)
267 SparkSession.setActiveSession(this)
718 object SparkSession { object
836 def getOrCreate(): SparkSession = synchronized {
848 SparkSession.synchronized {
877 session = new SparkSession(sparkContext)
[all …]
/dports/devel/spark/spark-2.1.1/sql/core/src/test/scala/org/apache/spark/sql/execution/
H A DExchangeCoordinatorSuite.scala30 private var originalActiveSparkSession: Option[SparkSession] = _
34 originalActiveSparkSession = SparkSession.getActiveSession
35 originalInstantiatedSparkSession = SparkSession.getDefaultSession
37 SparkSession.clearActiveSession()
38 SparkSession.clearDefaultSession()
253 f: SparkSession => Unit,
275 val spark = SparkSession.builder
288 val test = { spark: SparkSession =>
329 val test = { spark: SparkSession =>
380 val test = { spark: SparkSession =>
[all …]
/dports/devel/spark/spark-2.1.1/examples/src/main/java/org/apache/spark/examples/sql/
H A DJavaSQLDataSourceExample.java39 import org.apache.spark.sql.SparkSession;
96 SparkSession spark = SparkSession in main()
111 private static void runBasicDataSourceExample(SparkSession spark) { in runBasicDataSourceExample()
127 private static void runBasicParquetExample(SparkSession spark) { in runBasicParquetExample()
156 private static void runParquetSchemaMergingExample(SparkSession spark) { in runParquetSchemaMergingExample()
197 private static void runJsonDatasetExample(SparkSession spark) { in runJsonDatasetExample()
237 private static void runJdbcDatasetExample(SparkSession spark) { in runJdbcDatasetExample()
H A DJavaSparkSQLExample.java50 import org.apache.spark.sql.SparkSession;
90 SparkSession spark = SparkSession in main()
105 private static void runBasicDataFrameExample(SparkSession spark) throws AnalysisException { in runBasicDataFrameExample()
207 private static void runDatasetCreationExample(SparkSession spark) { in runDatasetCreationExample()
252 private static void runInferSchemaExample(SparkSession spark) { in runInferSchemaExample()
308 private static void runProgrammaticSchemaExample(SparkSession spark) { in runProgrammaticSchemaExample()
/dports/devel/spark/spark-2.1.1/sql/core/src/main/scala/org/apache/spark/sql/api/r/
H A DSQLUtils.scala49 enableHiveSupport: Boolean): SparkSession = {
50 val spark = if (SparkSession.hiveClassesArePresent && enableHiveSupport
52 SparkSession.builder().sparkContext(withHiveExternalCatalog(jsc.sc)).getOrCreate()
59 SparkSession.builder().sparkContext(jsc.sc).getOrCreate()
66 spark: SparkSession,
76 def getSessionConf(spark: SparkSession): JMap[String, String] = {
80 def getJavaSparkContext(spark: SparkSession): JavaSparkContext = {
234 sparkSession: SparkSession,
241 sparkSession: SparkSession,
271 def getTables(sparkSession: SparkSession, databaseName: String): DataFrame = {
[all …]
/dports/devel/spark/spark-2.1.1/mllib/src/test/scala/org/apache/spark/mllib/util/
H A DMLlibTestSparkContext.scala26 import org.apache.spark.sql.{SparkSession, SQLContext, SQLImplicits}
30 @transient var spark: SparkSession = _
36 spark = SparkSession.builder
49 SparkSession.clearActiveSession()
/dports/devel/spark/spark-2.1.1/repl/scala-2.11/src/main/scala/org/apache/spark/repl/
H A DMain.scala26 import org.apache.spark.sql.SparkSession
40 var sparkSession: SparkSession = _
74 def createSparkSession(): SparkSession = {
90 val builder = SparkSession.builder.config(conf)
92 if (SparkSession.hiveClassesArePresent) {
/dports/devel/spark/spark-2.1.1/python/pyspark/
H A Dshell.py32 from pyspark.sql import SparkSession, SQLContext
43 spark = SparkSession.builder\
47 spark = SparkSession.builder.getOrCreate()
49 spark = SparkSession.builder.getOrCreate()
/dports/devel/spark/spark-2.1.1/examples/src/main/scala/org/apache/spark/examples/streaming/
H A DSqlNetworkWordCount.scala23 import org.apache.spark.sql.SparkSession
92 @transient private var instance: SparkSession = _
94 def getInstance(sparkConf: SparkConf): SparkSession = {
96 instance = SparkSession
/dports/devel/spark/spark-2.1.1/mllib/src/test/scala/org/apache/spark/ml/util/
H A DMLTestingUtils.scala29 import org.apache.spark.sql.{DataFrame, Row, SparkSession}
43 spark: SparkSession,
66 spark: SparkSession,
88 def checkNumericTypes[T <: Evaluator](evaluator: T, spark: SparkSession): Unit = {
105 spark: SparkSession,
125 spark: SparkSession,
147 spark: SparkSession,
167 spark: SparkSession,
186 spark: SparkSession,
/dports/devel/spark/spark-2.1.1/examples/src/main/java/org/apache/spark/examples/streaming/
H A DJavaSqlNetworkWordCount.java31 import org.apache.spark.sql.SparkSession;
84 SparkSession spark = JavaSparkSessionSingleton.getInstance(rdd.context().getConf()); in main()
115 private static transient SparkSession instance = null;
116 public static SparkSession getInstance(SparkConf sparkConf) { in getInstance()
118 instance = SparkSession in getInstance()
/dports/devel/spark/spark-2.1.1/sql/core/src/test/scala/org/apache/spark/sql/execution/streaming/state/
H A DStateStoreRDDSuite.scala34 import org.apache.spark.sql.SparkSession
58 withSparkSession(SparkSession.builder.config(sparkConf).getOrCreate()) { spark =>
82 spark: SparkSession,
91 withSparkSession(SparkSession.builder.config(sparkConf).getOrCreate()) { spark =>
98 withSparkSession(SparkSession.builder.config(sparkConf).getOrCreate()) { spark =>
104 withSparkSession(SparkSession.builder.config(sparkConf).getOrCreate()) { spark =>
152 withSparkSession(SparkSession.builder.config(sparkConf).getOrCreate()) { spark =>
183 SparkSession.builder
/dports/devel/spark/spark-2.1.1/mllib/src/test/java/org/apache/spark/
H A DSharedSparkSession.java27 import org.apache.spark.sql.SparkSession;
31 protected transient SparkSession spark;
36 spark = SparkSession.builder() in setUp()
/dports/devel/spark/spark-2.1.1/sql/hive/src/main/scala/org/apache/spark/sql/hive/
H A DHiveContext.scala23 import org.apache.spark.sql.{SparkSession, SQLContext}
31 class HiveContext private[hive](_sparkSession: SparkSession)
37 this(SparkSession.builder().sparkContext(HiveUtils.withHiveExternalCatalog(sc)).getOrCreate())
/dports/devel/spark/spark-2.1.1/examples/src/main/java/org/apache/spark/examples/ml/
H A DJavaMulticlassLogisticRegressionWithElasticNetExample.java25 import org.apache.spark.sql.SparkSession;
30 SparkSession spark = SparkSession in main()
H A DJavaStandardScalerExample.java20 import org.apache.spark.sql.SparkSession;
31 SparkSession spark = SparkSession in main()

12345678910>>...20