org.apache.spark

sql

package sql

Allows the execution of relational queries, including those expressed in SQL using Spark.

Linear Supertypes
AnyRef, Any
Ordering
  1. Alphabetic
  2. By inheritance
Inherited
  1. sql
  2. AnyRef
  3. Any
  1. Hide All
  2. Show all
Learn more about member selection
Visibility
  1. Public
  2. All

Type Members

  1. class Column extends Logging

    :: Experimental :: A column that will be computed based on the data in a DataFrame.

  2. class ColumnName extends Column

    :: Experimental :: A convenient class used for constructing schema.

  3. trait ContinuousQuery extends AnyRef

    :: Experimental :: A handle to a query that is executing continuously in the background as new data arrives.

  4. class ContinuousQueryException extends Exception

    :: Experimental :: Exception that stopped a ContinuousQuery.

  5. class ContinuousQueryManager extends AnyRef

    :: Experimental :: A class to manage all the ContinuousQueries active on a SparkSession.

  6. type DataFrame = Dataset[Row]

  7. final class DataFrameNaFunctions extends AnyRef

    :: Experimental :: Functionality for working with missing data in DataFrames.

  8. class DataFrameReader extends Logging

    Interface used to load a Dataset from external storage systems (e.

  9. final class DataFrameStatFunctions extends AnyRef

    :: Experimental :: Statistic functions for DataFrames.

  10. final class DataFrameWriter extends AnyRef

    Interface used to write a Dataset to external storage systems (e.

  11. class Dataset[T] extends Serializable

    A Dataset is a strongly typed collection of domain-specific objects that can be transformed in parallel using functional or relational operations.

  12. case class DatasetHolder[T] extends Product with Serializable

    A container for a Dataset, used for implicit conversions in Scala.

  13. class ExperimentalMethods extends AnyRef

    :: Experimental :: Holder for experimental methods for the bravest.

  14. class KeyValueGroupedDataset[K, V] extends Serializable

    :: Experimental :: A Dataset has been logically grouped by a user specified grouping key.

  15. case class ProcessingTime(intervalMs: Long) extends Trigger with Product with Serializable

    :: Experimental :: A trigger that runs a query periodically based on the processing time.

  16. class RelationalGroupedDataset extends AnyRef

    A set of methods for aggregations on a DataFrame, created by Dataset.groupBy.

  17. class RuntimeConfig extends AnyRef

    Runtime configuration interface for Spark.

  18. class SQLContext extends Logging with Serializable

    The entry point for working with structured data (rows and columns) in Spark, in Spark 1.

  19. abstract class SQLImplicits extends AnyRef

    A collection of implicit methods for converting common Scala objects into DataFrames.

  20. class SaveMode extends Enum[SaveMode]

  21. class SinkStatus extends AnyRef

    :: Experimental :: Status and metrics of a streaming Sink.

  22. class SourceStatus extends AnyRef

    :: Experimental :: Status and metrics of a streaming Source.

  23. class SparkSession extends Serializable with Logging

    The entry point to programming Spark with the Dataset and DataFrame API.

  24. type Strategy = GenericStrategy[SparkPlan]

    Converts a logical plan into zero or more SparkPlans.

    Converts a logical plan into zero or more SparkPlans. This API is exposed for experimenting with the query planner and is not designed to be stable across spark releases. Developers writing libraries should instead consider using the stable APIs provided in org.apache.spark.sql.sources

    Annotations
    @DeveloperApi()
  25. sealed trait Trigger extends AnyRef

    :: Experimental :: Used to indicate how often results should be produced by a ContinuousQuery.

  26. class TypedColumn[-T, U] extends Column

    A Column where an Encoder has been given for the expected input and return type.

  27. class UDFRegistration extends Logging

    Functions for registering user-defined functions.

Value Members

  1. object ProcessingTime extends Serializable

    :: Experimental :: Used to create ProcessingTime triggers for ContinuousQuerys.

  2. object SQLContext extends Serializable

    This SQLContext object contains utility functions to create a singleton SQLContext instance, or to get the created SQLContext instance.

  3. object SparkSession extends Serializable

  4. package api

    Contains API classes that are specific to a single language (i.

  5. package catalog

  6. package catalyst

  7. package execution

    The physical execution component of Spark SQL.

  8. package expressions

  9. object functions

    :: Experimental :: Functions available for DataFrame.

  10. package internal

    All classes in this package are considered an internal API to Spark and are subject to change between minor releases.

  11. package jdbc

  12. package sources

    A set of APIs for adding data sources to Spark SQL.

  13. package util

Inherited from AnyRef

Inherited from Any

Ungrouped