package sql

Linear Supertypes
AnyRef, Any
Ordering
  1. Alphabetic
  2. By Inheritance
Inherited
  1. sql
  2. AnyRef
  3. Any
  1. Hide All
  2. Show All
Visibility
  1. Public
  2. All

Type Members

  1. type DataFrame = Dataset[Row]
  2. case class DateFieldStatAccumulator(nulls: LongAccumulator, nonNulls: LongAccumulator, mins: CollectionAccumulator[Long], maxs: CollectionAccumulator[Long], runningStats: StatsAccumulator) extends FieldStatAccumulator with Product with Serializable
  3. case class DoubleFieldStatAccumulator(nulls: LongAccumulator, nonNulls: LongAccumulator, mins: CollectionAccumulator[Double], maxs: CollectionAccumulator[Double], runningStats: StatsAccumulator) extends FieldStatAccumulator with Product with Serializable
  4. sealed trait FieldStatAccumulator extends Serializable
  5. case class Interim(outputFunction: InterimOutput, subgraph: String, component: String, port: String, subPath: String, numRows: Int, child: LogicalPlan, detailedStats: Boolean = false) extends LogicalPlan with UnaryLike[LogicalPlan] with IgnoreCachedData with Product with Serializable

    IgnoreCachedData:- To ignore caching in planning phase.

    IgnoreCachedData:- To ignore caching in planning phase. In planning phase we re-compute each and every node/step in tree repeatedly. and try to replace it with computed version of the node. this helps to not re-compute things of the tree which are already computed.

    Add this trait if you want to force the re-computation of the plan.

  6. case class InterimExec(outputFunction: InterimOutput, subgraph: String, component: String, port: String, subPath: String, child: SparkPlan, numRows: Int, detailedStats: Boolean, sparkSession: SparkSession) extends SparkPlan with InterimExec2 with Product with Serializable
  7. trait InterimExec2 extends SparkPlan with InterimExecBase with UnaryExecNode
  8. trait InterimExecBase extends SparkPlan
  9. case class InterimIteratorVersionWrapper(name: String, numRows: Int, rowAccumulator: CollectionAccumulator[InterimRow], collectStats: Boolean, statAccumulator: Array[FieldStatAccumulator], countAccumulator: LongAccumulator, partitionCounter: LongAccumulator, partitionId: Int)(schema: StructType, underlying: Iterator[InternalRow]) extends InterimIteratorWrapperTrait[Row] with Product with Serializable
  10. trait InterimIteratorWrapperTrait[T] extends Iterator[InternalRow] with Serializable
  11. trait InterimOutput extends Serializable
  12. case class InterimOutputHive2(session: String, workflowID: String, workflowVersion: String, jobid: String) extends Serializable with InterimOutput with Product
  13. case class InterimStat(dataFrameSchema: StructType, basicStats: Boolean, statsAccs: Array[FieldStatAccumulator], rowAccumulator: CollectionAccumulator[InterimRow], counterAccumulator: LongAccumulator, partitionCounter: LongAccumulator, rowAccSchema: StructType, schema: StructType, numPartitions: Int, numRows: Int, schemaJson: Seq[InternalRow], schemaSchema: StructType) extends Serializable with Product
  14. class InterimStore extends LazyLogging
  15. case class InterimStrategy(spark: SparkSession) extends SparkStrategy with Product with Serializable

    Spark strategy converts logical plan to sequence of spark plan.

    Spark strategy converts logical plan to sequence of spark plan. This class represents custom interim strategy. Logical operators other than Interim are converted using the regular SparkStrategy which is applied because of planLater method

    * GenericStrategy.planLater scala doc: Returns a placeholder for a physical plan that executes plan. This placeholder will be filled in automatically by the QueryPlanner using the other execution strategies that are available.

  16. case class LongFieldStatAccumulator(nulls: LongAccumulator, nonNulls: LongAccumulator, mins: CollectionAccumulator[Long], maxs: CollectionAccumulator[Long], runningStats: StatsAccumulator) extends FieldStatAccumulator with Product with Serializable
  17. case class OtherFieldStatAccumulator(nulls: LongAccumulator, nonNulls: LongAccumulator) extends FieldStatAccumulator with Product with Serializable
  18. class ProphecyDataFrame extends AnyRef
  19. case class ProphecySparkSession(spark: SparkSession) extends Product with Serializable
  20. type Strategy = SparkStrategy
    Annotations
    @DeveloperApi() @Unstable()
  21. case class StreamQueryListener(query: io.prophecy.libs.StreamingTarget, maxEmptyTicks: Int = 3) extends StreamingQueryListener with Product with Serializable
  22. case class StringFieldStatAccumulator(nulls: LongAccumulator, nonNulls: LongAccumulator, mins: CollectionAccumulator[String], maxs: CollectionAccumulator[String]) extends FieldStatAccumulator with Product with Serializable
  23. case class TimestampFieldStatAccumulator(nulls: LongAccumulator, nonNulls: LongAccumulator, mins: CollectionAccumulator[Long], maxs: CollectionAccumulator[Long], runningStats: StatsAccumulator) extends FieldStatAccumulator with Product with Serializable

Value Members

  1. object FieldStatAccumulator extends Serializable
  2. object InterimExecBase extends Serializable
  3. object InterimIteratorVersionWrapper extends Serializable
  4. object InterimOutputHive
  5. object InterimOutputHive2 extends Serializable
  6. object InterimStore extends Serializable
  7. object MetricsCollector extends Serializable with LazyLogging
  8. object ProphecyDataFrame
  9. object ProphecyDebugger
  10. object ProphecyExtensions
  11. object ProphecySparkSession extends Serializable
  12. object ProphecyTimer extends Serializable

Inherited from AnyRef

Inherited from Any

Ungrouped