com.memsql.spark.connector.rdd

MemSQLRDD

case class MemSQLRDD[T](sc: SparkContext, cluster: MemSQLCluster, sql: String, sqlParams: Seq[Any] = immutable.this.Nil, databaseName: Option[String] = scala.None, mapRow: (ResultSet) ⇒ T = ...)(implicit evidence$1: ClassTag[T]) extends RDD[T] with Logging with Product with Serializable

An org.apache.spark.rdd.RDD that can read data from a MemSQL database based on a SQL query.

If the given query supports it, this RDD will read data directly from the MemSQL cluster's leaf nodes rather than from the master aggregator, which typically results in much faster reads. However, if the given query does not support this (e.g. queries involving joins or GROUP BY operations), the results will be returned in a single partition.

cluster

A connected MemSQLCluster instance.

sql

The text of the query. Can be a prepared statement template, in which case parameters from sqlParams are substituted.

sqlParams

The parameters of the query if sql is a template.

databaseName

Optionally provide a database name for this RDD. This is required for Partition Pushdown

mapRow

A function from a ResultSet to a single row of the desired result type(s). This should only call getInt, getString, etc; the RDD takes care of calling next. The default maps a ResultSet to an array of Any.

Linear Supertypes
Ordering
  1. Alphabetic
  2. By inheritance
Inherited
  1. MemSQLRDD
  2. Product
  3. Equals
  4. RDD
  5. Logging
  6. Serializable
  7. Serializable
  8. AnyRef
  9. Any
  1. Hide All
  2. Show all
Learn more about member selection
Visibility
  1. Public
  2. All

Instance Constructors

  1. new MemSQLRDD(sc: SparkContext, cluster: MemSQLCluster, sql: String, sqlParams: Seq[Any] = immutable.this.Nil, databaseName: Option[String] = scala.None, mapRow: (ResultSet) ⇒ T = ...)(implicit arg0: ClassTag[T])

    cluster

    A connected MemSQLCluster instance.

    sql

    The text of the query. Can be a prepared statement template, in which case parameters from sqlParams are substituted.

    sqlParams

    The parameters of the query if sql is a template.

    databaseName

    Optionally provide a database name for this RDD. This is required for Partition Pushdown

    mapRow

    A function from a ResultSet to a single row of the desired result type(s). This should only call getInt, getString, etc; the RDD takes care of calling next. The default maps a ResultSet to an array of Any.

Value Members

  1. final def !=(arg0: AnyRef): Boolean

    Definition Classes
    AnyRef
  2. final def !=(arg0: Any): Boolean

    Definition Classes
    Any
  3. final def ##(): Int

    Definition Classes
    AnyRef → Any
  4. def ++(other: RDD[T]): RDD[T]

    Definition Classes
    RDD
  5. final def ==(arg0: AnyRef): Boolean

    Definition Classes
    AnyRef
  6. final def ==(arg0: Any): Boolean

    Definition Classes
    Any
  7. def aggregate[U](zeroValue: U)(seqOp: (U, T) ⇒ U, combOp: (U, U) ⇒ U)(implicit arg0: ClassTag[U]): U

    Definition Classes
    RDD
  8. final def asInstanceOf[T0]: T0

    Definition Classes
    Any
  9. def cache(): MemSQLRDD.this.type

    Definition Classes
    RDD
  10. def cartesian[U](other: RDD[U])(implicit arg0: ClassTag[U]): RDD[(T, U)]

    Definition Classes
    RDD
  11. def checkpoint(): Unit

    Definition Classes
    RDD
  12. def clearDependencies(): Unit

    Attributes
    protected
    Definition Classes
    RDD
  13. def clone(): AnyRef

    Attributes
    protected[java.lang]
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  14. val cluster: MemSQLCluster

    A connected MemSQLCluster instance.

  15. def coalesce(numPartitions: Int, shuffle: Boolean)(implicit ord: Ordering[T]): RDD[T]

    Definition Classes
    RDD
  16. def collect[U](f: PartialFunction[T, U])(implicit arg0: ClassTag[U]): RDD[U]

    Definition Classes
    RDD
  17. def collect(): Array[T]

    Definition Classes
    RDD
  18. def compute(sparkPartition: Partition, context: TaskContext): Iterator[T]

    Definition Classes
    MemSQLRDD → RDD
  19. def context: SparkContext

    Definition Classes
    RDD
  20. def count(): Long

    Definition Classes
    RDD
  21. def countApprox(timeout: Long, confidence: Double): PartialResult[BoundedDouble]

    Definition Classes
    RDD
    Annotations
    @Experimental()
  22. def countApproxDistinct(relativeSD: Double): Long

    Definition Classes
    RDD
  23. def countApproxDistinct(p: Int, sp: Int): Long

    Definition Classes
    RDD
    Annotations
    @Experimental()
  24. def countByValue()(implicit ord: Ordering[T]): Map[T, Long]

    Definition Classes
    RDD
  25. def countByValueApprox(timeout: Long, confidence: Double)(implicit ord: Ordering[T]): PartialResult[Map[T, BoundedDouble]]

    Definition Classes
    RDD
    Annotations
    @Experimental()
  26. val databaseName: Option[String]

    Optionally provide a database name for this RDD.

    Optionally provide a database name for this RDD. This is required for Partition Pushdown

  27. final def dependencies: Seq[Dependency[_]]

    Definition Classes
    RDD
  28. def distinct(): RDD[T]

    Definition Classes
    RDD
  29. def distinct(numPartitions: Int)(implicit ord: Ordering[T]): RDD[T]

    Definition Classes
    RDD
  30. final def eq(arg0: AnyRef): Boolean

    Definition Classes
    AnyRef
  31. def filter(f: (T) ⇒ Boolean): RDD[T]

    Definition Classes
    RDD
  32. def finalize(): Unit

    Attributes
    protected[java.lang]
    Definition Classes
    AnyRef
    Annotations
    @throws( classOf[java.lang.Throwable] )
  33. def first(): T

    Definition Classes
    RDD
  34. def firstParent[U](implicit arg0: ClassTag[U]): RDD[U]

    Attributes
    protected[org.apache.spark]
    Definition Classes
    RDD
  35. def flatMap[U](f: (T) ⇒ TraversableOnce[U])(implicit arg0: ClassTag[U]): RDD[U]

    Definition Classes
    RDD
  36. def fold(zeroValue: T)(op: (T, T) ⇒ T): T

    Definition Classes
    RDD
  37. def foreach(f: (T) ⇒ Unit): Unit

    Definition Classes
    RDD
  38. def foreachPartition(f: (Iterator[T]) ⇒ Unit): Unit

    Definition Classes
    RDD
  39. def getCheckpointFile: Option[String]

    Definition Classes
    RDD
  40. final def getClass(): Class[_]

    Definition Classes
    AnyRef → Any
  41. def getDependencies: Seq[Dependency[_]]

    Attributes
    protected
    Definition Classes
    RDD
  42. def getPartitions: Array[Partition]

    Definition Classes
    MemSQLRDD → RDD
  43. def getPreferredLocations(sparkPartition: Partition): Seq[String]

    Definition Classes
    MemSQLRDD → RDD
  44. def getStorageLevel: StorageLevel

    Definition Classes
    RDD
  45. def glom(): RDD[Array[T]]

    Definition Classes
    RDD
  46. def groupBy[K](f: (T) ⇒ K, p: Partitioner)(implicit kt: ClassTag[K], ord: Ordering[K]): RDD[(K, Iterable[T])]

    Definition Classes
    RDD
  47. def groupBy[K](f: (T) ⇒ K, numPartitions: Int)(implicit kt: ClassTag[K]): RDD[(K, Iterable[T])]

    Definition Classes
    RDD
  48. def groupBy[K](f: (T) ⇒ K)(implicit kt: ClassTag[K]): RDD[(K, Iterable[T])]

    Definition Classes
    RDD
  49. val id: Int

    Definition Classes
    RDD
  50. def intersection(other: RDD[T], numPartitions: Int): RDD[T]

    Definition Classes
    RDD
  51. def intersection(other: RDD[T], partitioner: Partitioner)(implicit ord: Ordering[T]): RDD[T]

    Definition Classes
    RDD
  52. def intersection(other: RDD[T]): RDD[T]

    Definition Classes
    RDD
  53. def isCheckpointed: Boolean

    Definition Classes
    RDD
  54. def isEmpty(): Boolean

    Definition Classes
    RDD
  55. final def isInstanceOf[T0]: Boolean

    Definition Classes
    Any
  56. def isTraceEnabled(): Boolean

    Attributes
    protected
    Definition Classes
    Logging
  57. final def iterator(split: Partition, context: TaskContext): Iterator[T]

    Definition Classes
    RDD
  58. def keyBy[K](f: (T) ⇒ K): RDD[(K, T)]

    Definition Classes
    RDD
  59. def localCheckpoint(): MemSQLRDD.this.type

    Definition Classes
    RDD
  60. def log: Logger

    Attributes
    protected
    Definition Classes
    Logging
  61. def logDebug(msg: ⇒ String, throwable: Throwable): Unit

    Attributes
    protected
    Definition Classes
    Logging
  62. def logDebug(msg: ⇒ String): Unit

    Attributes
    protected
    Definition Classes
    Logging
  63. def logError(msg: ⇒ String, throwable: Throwable): Unit

    Attributes
    protected
    Definition Classes
    Logging
  64. def logError(msg: ⇒ String): Unit

    Attributes
    protected
    Definition Classes
    Logging
  65. def logInfo(msg: ⇒ String, throwable: Throwable): Unit

    Attributes
    protected
    Definition Classes
    Logging
  66. def logInfo(msg: ⇒ String): Unit

    Attributes
    protected
    Definition Classes
    Logging
  67. def logName: String

    Attributes
    protected
    Definition Classes
    Logging
  68. def logTrace(msg: ⇒ String, throwable: Throwable): Unit

    Attributes
    protected
    Definition Classes
    Logging
  69. def logTrace(msg: ⇒ String): Unit

    Attributes
    protected
    Definition Classes
    Logging
  70. def logWarning(msg: ⇒ String, throwable: Throwable): Unit

    Attributes
    protected
    Definition Classes
    Logging
  71. def logWarning(msg: ⇒ String): Unit

    Attributes
    protected
    Definition Classes
    Logging
  72. def map[U](f: (T) ⇒ U)(implicit arg0: ClassTag[U]): RDD[U]

    Definition Classes
    RDD
  73. def mapPartitions[U](f: (Iterator[T]) ⇒ Iterator[U], preservesPartitioning: Boolean)(implicit arg0: ClassTag[U]): RDD[U]

    Definition Classes
    RDD
  74. def mapPartitionsWithIndex[U](f: (Int, Iterator[T]) ⇒ Iterator[U], preservesPartitioning: Boolean)(implicit arg0: ClassTag[U]): RDD[U]

    Definition Classes
    RDD
  75. val mapRow: (ResultSet) ⇒ T

    A function from a ResultSet to a single row of the desired result type(s).

    A function from a ResultSet to a single row of the desired result type(s). This should only call getInt, getString, etc; the RDD takes care of calling next. The default maps a ResultSet to an array of Any.

  76. def max()(implicit ord: Ordering[T]): T

    Definition Classes
    RDD
  77. def min()(implicit ord: Ordering[T]): T

    Definition Classes
    RDD
  78. var name: String

    Definition Classes
    RDD
  79. final def ne(arg0: AnyRef): Boolean

    Definition Classes
    AnyRef
  80. final def notify(): Unit

    Definition Classes
    AnyRef
  81. final def notifyAll(): Unit

    Definition Classes
    AnyRef
  82. def parent[U](j: Int)(implicit arg0: ClassTag[U]): RDD[U]

    Attributes
    protected[org.apache.spark]
    Definition Classes
    RDD
  83. val partitioner: Option[Partitioner]

    Definition Classes
    RDD
  84. final def partitions: Array[Partition]

    Definition Classes
    RDD
  85. def persist(): MemSQLRDD.this.type

    Definition Classes
    RDD
  86. def persist(newLevel: StorageLevel): MemSQLRDD.this.type

    Definition Classes
    RDD
  87. def pipe(command: Seq[String], env: Map[String, String], printPipeContext: ((String) ⇒ Unit) ⇒ Unit, printRDDElement: (T, (String) ⇒ Unit) ⇒ Unit, separateWorkingDir: Boolean): RDD[String]

    Definition Classes
    RDD
  88. def pipe(command: String, env: Map[String, String]): RDD[String]

    Definition Classes
    RDD
  89. def pipe(command: String): RDD[String]

    Definition Classes
    RDD
  90. final def preferredLocations(split: Partition): Seq[String]

    Definition Classes
    RDD
  91. def randomSplit(weights: Array[Double], seed: Long): Array[RDD[T]]

    Definition Classes
    RDD
  92. def reduce(f: (T, T) ⇒ T): T

    Definition Classes
    RDD
  93. def repartition(numPartitions: Int)(implicit ord: Ordering[T]): RDD[T]

    Definition Classes
    RDD
  94. def sample(withReplacement: Boolean, fraction: Double, seed: Long): RDD[T]

    Definition Classes
    RDD
  95. def saveAsObjectFile(path: String): Unit

    Definition Classes
    RDD
  96. def saveAsTextFile(path: String, codec: Class[_ <: CompressionCodec]): Unit

    Definition Classes
    RDD
  97. def saveAsTextFile(path: String): Unit

    Definition Classes
    RDD
  98. val sc: SparkContext

  99. def setName(_name: String): MemSQLRDD.this.type

    Definition Classes
    RDD
  100. def sortBy[K](f: (T) ⇒ K, ascending: Boolean, numPartitions: Int)(implicit ord: Ordering[K], ctag: ClassTag[K]): RDD[T]

    Definition Classes
    RDD
  101. def sparkContext: SparkContext

    Definition Classes
    RDD
  102. val sql: String

    The text of the query.

    The text of the query. Can be a prepared statement template, in which case parameters from sqlParams are substituted.

  103. val sqlParams: Seq[Any]

    The parameters of the query if sql is a template.

  104. def subtract(other: RDD[T], p: Partitioner)(implicit ord: Ordering[T]): RDD[T]

    Definition Classes
    RDD
  105. def subtract(other: RDD[T], numPartitions: Int): RDD[T]

    Definition Classes
    RDD
  106. def subtract(other: RDD[T]): RDD[T]

    Definition Classes
    RDD
  107. final def synchronized[T0](arg0: ⇒ T0): T0

    Definition Classes
    AnyRef
  108. def take(num: Int): Array[T]

    Definition Classes
    RDD
  109. def takeOrdered(num: Int)(implicit ord: Ordering[T]): Array[T]

    Definition Classes
    RDD
  110. def takeSample(withReplacement: Boolean, num: Int, seed: Long): Array[T]

    Definition Classes
    RDD
  111. def toDebugString: String

    Definition Classes
    RDD
  112. def toJavaRDD(): JavaRDD[T]

    Definition Classes
    RDD
  113. def toLocalIterator: Iterator[T]

    Definition Classes
    RDD
  114. def toString(): String

    Definition Classes
    RDD → AnyRef → Any
  115. def top(num: Int)(implicit ord: Ordering[T]): Array[T]

    Definition Classes
    RDD
  116. def treeAggregate[U](zeroValue: U)(seqOp: (U, T) ⇒ U, combOp: (U, U) ⇒ U, depth: Int)(implicit arg0: ClassTag[U]): U

    Definition Classes
    RDD
  117. def treeReduce(f: (T, T) ⇒ T, depth: Int): T

    Definition Classes
    RDD
  118. def union(other: RDD[T]): RDD[T]

    Definition Classes
    RDD
  119. def unpersist(blocking: Boolean): MemSQLRDD.this.type

    Definition Classes
    RDD
  120. final def wait(): Unit

    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  121. final def wait(arg0: Long, arg1: Int): Unit

    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  122. final def wait(arg0: Long): Unit

    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  123. def zip[U](other: RDD[U])(implicit arg0: ClassTag[U]): RDD[(T, U)]

    Definition Classes
    RDD
  124. def zipPartitions[B, C, D, V](rdd2: RDD[B], rdd3: RDD[C], rdd4: RDD[D])(f: (Iterator[T], Iterator[B], Iterator[C], Iterator[D]) ⇒ Iterator[V])(implicit arg0: ClassTag[B], arg1: ClassTag[C], arg2: ClassTag[D], arg3: ClassTag[V]): RDD[V]

    Definition Classes
    RDD
  125. def zipPartitions[B, C, D, V](rdd2: RDD[B], rdd3: RDD[C], rdd4: RDD[D], preservesPartitioning: Boolean)(f: (Iterator[T], Iterator[B], Iterator[C], Iterator[D]) ⇒ Iterator[V])(implicit arg0: ClassTag[B], arg1: ClassTag[C], arg2: ClassTag[D], arg3: ClassTag[V]): RDD[V]

    Definition Classes
    RDD
  126. def zipPartitions[B, C, V](rdd2: RDD[B], rdd3: RDD[C])(f: (Iterator[T], Iterator[B], Iterator[C]) ⇒ Iterator[V])(implicit arg0: ClassTag[B], arg1: ClassTag[C], arg2: ClassTag[V]): RDD[V]

    Definition Classes
    RDD
  127. def zipPartitions[B, C, V](rdd2: RDD[B], rdd3: RDD[C], preservesPartitioning: Boolean)(f: (Iterator[T], Iterator[B], Iterator[C]) ⇒ Iterator[V])(implicit arg0: ClassTag[B], arg1: ClassTag[C], arg2: ClassTag[V]): RDD[V]

    Definition Classes
    RDD
  128. def zipPartitions[B, V](rdd2: RDD[B])(f: (Iterator[T], Iterator[B]) ⇒ Iterator[V])(implicit arg0: ClassTag[B], arg1: ClassTag[V]): RDD[V]

    Definition Classes
    RDD
  129. def zipPartitions[B, V](rdd2: RDD[B], preservesPartitioning: Boolean)(f: (Iterator[T], Iterator[B]) ⇒ Iterator[V])(implicit arg0: ClassTag[B], arg1: ClassTag[V]): RDD[V]

    Definition Classes
    RDD
  130. def zipWithIndex(): RDD[(T, Long)]

    Definition Classes
    RDD
  131. def zipWithUniqueId(): RDD[(T, Long)]

    Definition Classes
    RDD

Deprecated Value Members

  1. def filterWith[A](constructA: (Int) ⇒ A)(p: (T, A) ⇒ Boolean): RDD[T]

    Definition Classes
    RDD
    Annotations
    @deprecated
    Deprecated

    (Since version 1.0.0) use mapPartitionsWithIndex and filter

  2. def flatMapWith[A, U](constructA: (Int) ⇒ A, preservesPartitioning: Boolean)(f: (T, A) ⇒ Seq[U])(implicit arg0: ClassTag[U]): RDD[U]

    Definition Classes
    RDD
    Annotations
    @deprecated
    Deprecated

    (Since version 1.0.0) use mapPartitionsWithIndex and flatMap

  3. def foreachWith[A](constructA: (Int) ⇒ A)(f: (T, A) ⇒ Unit): Unit

    Definition Classes
    RDD
    Annotations
    @deprecated
    Deprecated

    (Since version 1.0.0) use mapPartitionsWithIndex and foreach

  4. def mapPartitionsWithContext[U](f: (TaskContext, Iterator[T]) ⇒ Iterator[U], preservesPartitioning: Boolean)(implicit arg0: ClassTag[U]): RDD[U]

    Definition Classes
    RDD
    Annotations
    @DeveloperApi() @deprecated
    Deprecated

    (Since version 1.2.0) use TaskContext.get

  5. def mapPartitionsWithSplit[U](f: (Int, Iterator[T]) ⇒ Iterator[U], preservesPartitioning: Boolean)(implicit arg0: ClassTag[U]): RDD[U]

    Definition Classes
    RDD
    Annotations
    @deprecated
    Deprecated

    (Since version 0.7.0) use mapPartitionsWithIndex

  6. def mapWith[A, U](constructA: (Int) ⇒ A, preservesPartitioning: Boolean)(f: (T, A) ⇒ U)(implicit arg0: ClassTag[U]): RDD[U]

    Definition Classes
    RDD
    Annotations
    @deprecated
    Deprecated

    (Since version 1.0.0) use mapPartitionsWithIndex

  7. def toArray(): Array[T]

    Definition Classes
    RDD
    Annotations
    @deprecated
    Deprecated

    (Since version 1.0.0) use collect

Inherited from Product

Inherited from Equals

Inherited from RDD[T]

Inherited from Logging

Inherited from Serializable

Inherited from Serializable

Inherited from AnyRef

Inherited from Any

Ungrouped