package sql
- Alphabetic
- By Inheritance
- sql
- AnyRef
- Any
- Hide All
- Show All
- Public
- All
Type Members
- type DataFrame = Dataset[Row]
- case class DateFieldStatAccumulator(nulls: LongAccumulator, nonNulls: LongAccumulator, mins: CollectionAccumulator[Long], maxs: CollectionAccumulator[Long], runningStats: StatsAccumulator) extends FieldStatAccumulator with Product with Serializable
- case class DoubleFieldStatAccumulator(nulls: LongAccumulator, nonNulls: LongAccumulator, mins: CollectionAccumulator[Double], maxs: CollectionAccumulator[Double], runningStats: StatsAccumulator) extends FieldStatAccumulator with Product with Serializable
- sealed trait FieldStatAccumulator extends Serializable
-
case class
Interim(outputFunction: InterimOutput, subgraph: String, component: String, port: String, subPath: String, numRows: Int, child: LogicalPlan, detailedStats: Boolean = false) extends LogicalPlan with UnaryLike[LogicalPlan] with IgnoreCachedData with Product with Serializable
IgnoreCachedData:- To ignore caching in planning phase.
IgnoreCachedData:- To ignore caching in planning phase. In planning phase we re-compute each and every node/step in tree repeatedly. and try to replace it with computed version of the node. this helps to not re-compute things of the tree which are already computed.
Add this trait if you want to force the re-computation of the plan.
- case class InterimExec(outputFunction: InterimOutput, subgraph: String, component: String, port: String, subPath: String, child: SparkPlan, numRows: Int, detailedStats: Boolean, sparkSession: SparkSession) extends SparkPlan with InterimExec2 with Product with Serializable
- trait InterimExec2 extends SparkPlan with InterimExecBase with UnaryExecNode
- trait InterimExecBase extends SparkPlan
- case class InterimIteratorVersionWrapper(name: String, numRows: Int, rowAccumulator: CollectionAccumulator[InterimRow], collectStats: Boolean, statAccumulator: Array[FieldStatAccumulator], countAccumulator: LongAccumulator, partitionCounter: LongAccumulator, partitionId: Int)(schema: StructType, underlying: Iterator[InternalRow]) extends InterimIteratorWrapperTrait[Row] with Product with Serializable
- trait InterimIteratorWrapperTrait[T] extends Iterator[InternalRow] with Serializable
- trait InterimOutput extends Serializable
- case class InterimOutputHive2(session: String, workflowID: String, workflowVersion: String, jobid: String) extends Serializable with InterimOutput with Product
- case class InterimStat(dataFrameSchema: StructType, basicStats: Boolean, statsAccs: Array[FieldStatAccumulator], rowAccumulator: CollectionAccumulator[InterimRow], counterAccumulator: LongAccumulator, partitionCounter: LongAccumulator, rowAccSchema: StructType, schema: StructType, numPartitions: Int, numRows: Int, schemaJson: Seq[InternalRow], schemaSchema: StructType) extends Serializable with Product
- class InterimStore extends LazyLogging
-
case class
InterimStrategy(spark: SparkSession) extends SparkStrategy with Product with Serializable
Spark strategy converts logical plan to sequence of spark plan.
Spark strategy converts logical plan to sequence of spark plan. This class represents custom interim strategy. Logical operators other than
Interim
are converted using the regular SparkStrategy which is applied because of planLater method* GenericStrategy.planLater scala doc: Returns a placeholder for a physical plan that executes
plan
. This placeholder will be filled in automatically by the QueryPlanner using the other execution strategies that are available. - case class LongFieldStatAccumulator(nulls: LongAccumulator, nonNulls: LongAccumulator, mins: CollectionAccumulator[Long], maxs: CollectionAccumulator[Long], runningStats: StatsAccumulator) extends FieldStatAccumulator with Product with Serializable
- case class OtherFieldStatAccumulator(nulls: LongAccumulator, nonNulls: LongAccumulator) extends FieldStatAccumulator with Product with Serializable
- class ProphecyDataFrame extends AnyRef
- case class ProphecySparkSession(spark: SparkSession) extends Product with Serializable
-
type
Strategy = SparkStrategy
- Annotations
- @DeveloperApi() @Unstable()
- case class StreamQueryListener(query: io.prophecy.libs.StreamingTarget, maxEmptyTicks: Int = 3) extends StreamingQueryListener with Product with Serializable
- case class StringFieldStatAccumulator(nulls: LongAccumulator, nonNulls: LongAccumulator, mins: CollectionAccumulator[String], maxs: CollectionAccumulator[String]) extends FieldStatAccumulator with Product with Serializable
- case class TimestampFieldStatAccumulator(nulls: LongAccumulator, nonNulls: LongAccumulator, mins: CollectionAccumulator[Long], maxs: CollectionAccumulator[Long], runningStats: StatsAccumulator) extends FieldStatAccumulator with Product with Serializable
Value Members
- object FieldStatAccumulator extends Serializable
- object InterimExecBase extends Serializable
- object InterimIteratorVersionWrapper extends Serializable
- object InterimOutputHive
- object InterimOutputHive2 extends Serializable
- object InterimStore extends Serializable
- object MetricsCollector extends Serializable with LazyLogging
- object ProphecyDataFrame
- object ProphecyDebugger
- object ProphecyExtensions
- object ProphecySparkSession extends Serializable
- object ProphecyTimer extends Serializable