Package

doric

Permalink

package doric

Linear Supertypes
Ordering
  1. Grouped
  2. Alphabetic
  3. By Inheritance
Inherited
  1. doric
  2. All
  3. SortingOps
  4. CollectOps
  5. JoinOps
  6. TransformOps
  7. AggregationOps
  8. RelationalGroupedDatasetDoricInterface
  9. All
  10. NumericColumns2_31
  11. StringColumn24
  12. BinaryColumns30_31
  13. Interpolators
  14. BinaryColumns
  15. CNameOps
  16. AggregationColumns
  17. ControlStructures
  18. StringColumns
  19. BooleanColumns
  20. TimestampColumns
  21. DateColumns
  22. NumericColumns
  23. MapColumns
  24. LiteralConversions
  25. DStructs
  26. CommonColumns
  27. ColGetters
  28. TypeMatcher
  29. ArrayColumns
  30. AnyRef
  31. Any
  1. Hide All
  2. Show All
Visibility
  1. Public
  2. All

Type Members

  1. implicit class ArrayArrayColumnSyntax[G[_], F[_], T] extends AnyRef

    Permalink
    Definition Classes
    ArrayColumns
  2. type ArrayColumn[A] = DoricColumn[Array[A]]

    Permalink
  3. implicit class ArrayColumnSyntax[T, F[_]] extends AnyRef

    Permalink

    Extension methods for arrays

    Extension methods for arrays

    Definition Classes
    ArrayColumns
  4. implicit class ArrayColumnTupleSyntax[K, V, F[_]] extends AnyRef

    Permalink

    Extension methods for arrays

    Extension methods for arrays

    Definition Classes
    ArrayColumns
  5. implicit class BasicCol[T] extends AnyRef

    Permalink

    Extension methods for any kind of column

    Extension methods for any kind of column

    Definition Classes
    CommonColumns
  6. type BinaryColumn = DoricColumn[Array[Byte]]

    Permalink
  7. implicit class BinaryOperationsSyntax[T] extends AnyRef

    Permalink
    Definition Classes
    BinaryColumns
  8. implicit class BinaryOperationsSyntax30_31[T] extends AnyRef

    Permalink
    Definition Classes
    BinaryColumns30_31
  9. type BooleanColumn = DoricColumn[Boolean]

    Permalink
  10. implicit class BooleanOperationsSyntax extends AnyRef

    Permalink
    Definition Classes
    BooleanColumns
  11. type ByteColumn = DoricColumn[Byte]

    Permalink
  12. case class CName(value: String) extends Product with Serializable

    Permalink
  13. implicit class CNameOps extends AnyRef

    Permalink
    Definition Classes
    CNameOps
  14. case class CNameOrd(name: CName, order: Order) extends Product with Serializable

    Permalink
  15. implicit class CastingImpl[T] extends AnyRef

    Permalink

    Casting methods

    Casting methods

    Definition Classes
    CommonColumns
  16. implicit class CollectSyntax[A] extends AnyRef

    Permalink
    Definition Classes
    CollectOps
  17. implicit class ControlStructuresImpl[O] extends AnyRef

    Permalink
    Definition Classes
    ControlStructures
  18. implicit class DStructOps[T] extends AnyRef

    Permalink
    Definition Classes
    DStructs
  19. implicit class DataframeAggSyntax extends AnyRef

    Permalink
    Definition Classes
    AggregationOps
  20. implicit class DataframeJoinSyntax[A] extends AnyRef

    Permalink
    Definition Classes
    JoinOps
  21. implicit class DataframeSortSyntax extends AnyRef

    Permalink
    Definition Classes
    SortingOps
  22. implicit class DataframeTransformationSyntax[A] extends AnyRef

    Permalink
    Definition Classes
    TransformOps
  23. type DateColumn = DoricColumn[Date]

    Permalink
  24. implicit class DateColumnLikeSyntax[T] extends AnyRef

    Permalink
    Definition Classes
    DateColumns
  25. type Doric[T] = Kleisli[DoricValidated, Dataset[_], T]

    Permalink
  26. implicit class DoricColLiteralGetter[T] extends AnyRef

    Permalink
    Definition Classes
    LiteralConversions
  27. sealed trait DoricColumn[T] extends AnyRef

    Permalink
  28. type DoricJoin[T] = Kleisli[DoricValidated, (Dataset[_], Dataset[_]), T]

    Permalink
  29. case class DoricJoinColumn(elem: DoricJoin[Column]) extends Product with Serializable

    Permalink
  30. type DoricValidated[T] = Validated[NonEmptyChain[DoricSingleError], T]

    Permalink
  31. type DoubleColumn = DoricColumn[Double]

    Permalink
  32. class DynamicFieldAccessor[T] extends Dynamic

    Permalink
    Definition Classes
    DStructs
  33. type FloatColumn = DoricColumn[Float]

    Permalink
  34. type InstantColumn = DoricColumn[Instant]

    Permalink
  35. type IntegerColumn = DoricColumn[Int]

    Permalink
  36. implicit class IntegralOperationsSyntax[T] extends AnyRef

    Permalink

    INTEGRAL OPERATIONS

    INTEGRAL OPERATIONS

    Definition Classes
    NumericColumns
  37. implicit class IntegralOperationsSyntax2_31[T] extends AnyRef

    Permalink

    INTEGRAL OPERATIONS

    INTEGRAL OPERATIONS

    Definition Classes
    NumericColumns2_31
  38. sealed abstract class JoinSideDoricColumn[T] extends AnyRef

    Permalink
  39. case class LeftDoricColumn[T](elem: Doric[Column]) extends JoinSideDoricColumn[T] with Product with Serializable

    Permalink
  40. case class LiteralDoricColumn[T] extends DoricColumn[T] with Product with Serializable

    Permalink
  41. implicit class LiteralOps[L] extends AnyRef

    Permalink
    Definition Classes
    LiteralConversions
  42. type LocalDateColumn = DoricColumn[LocalDate]

    Permalink
  43. type LongColumn = DoricColumn[Long]

    Permalink
  44. implicit class LongOperationsSyntax extends AnyRef

    Permalink

    LONG OPERATIONS

    LONG OPERATIONS

    Definition Classes
    NumericColumns
  45. type MapColumn[K, V] = DoricColumn[Map[K, V]]

    Permalink
  46. implicit class MapColumnOps[K, V] extends AnyRef

    Permalink

    Extension methods for Map Columns

    Extension methods for Map Columns

    Definition Classes
    MapColumns
  47. case class NamedDoricColumn[T] extends DoricColumn[T] with Product with Serializable

    Permalink
  48. type NullColumn = DoricColumn[Null]

    Permalink
  49. implicit class NumWithDecimalsOperationsSyntax[T] extends AnyRef

    Permalink

    NUM WITH DECIMALS OPERATIONS

    NUM WITH DECIMALS OPERATIONS

    Definition Classes
    NumericColumns
  50. implicit class NumericOperationsSyntax[T] extends AnyRef

    Permalink

    GENERIC NUMERIC OPERATIONS

    GENERIC NUMERIC OPERATIONS

    Definition Classes
    NumericColumns
  51. sealed trait Order extends AnyRef

    Permalink
  52. implicit class RelationalGroupedDatasetSem extends AnyRef

    Permalink
    Definition Classes
    AggregationOps
  53. case class RightDoricColumn[T](elem: Doric[Column]) extends JoinSideDoricColumn[T] with Product with Serializable

    Permalink
  54. type RowColumn = DoricColumn[Row]

    Permalink
  55. trait SelectorLPI extends AnyRef

    Permalink
    Definition Classes
    DStructs
  56. trait SelectorWithSparkType[L <: HList, K <: Symbol] extends AnyRef

    Permalink
    Definition Classes
    DStructs
    Annotations
    @implicitNotFound( "No field ${K} in record ${L}" )
  57. implicit class SparkCol extends AnyRef

    Permalink
    Definition Classes
    CommonColumns
  58. implicit class StringCNameOps extends AnyRef

    Permalink
    Definition Classes
    CNameOps
  59. type StringColumn = DoricColumn[String]

    Permalink
  60. implicit final class StringIntCNameOps extends AnyVal

    Permalink
  61. implicit class StringOperationsSyntax extends AnyRef

    Permalink

    Unique column operations

    Unique column operations

    Definition Classes
    StringColumns
  62. implicit class StringOperationsSyntax24 extends AnyRef

    Permalink
    Definition Classes
    StringColumn24
  63. implicit class StructOps[T, L <: HList] extends AnyRef

    Permalink
    Definition Classes
    DStructs
  64. type TimestampColumn = DoricColumn[Timestamp]

    Permalink
  65. implicit class TimestampColumnLikeSyntax[T] extends AnyRef

    Permalink
    Definition Classes
    TimestampColumns
  66. case class TransformationDoricColumn[T] extends DoricColumn[T] with Product with Serializable

    Permalink
  67. implicit class doricStringInterpolator extends AnyRef

    Permalink
    Definition Classes
    Interpolators

Abstract Value Members

  1. abstract def constructSide[T](column: Doric[Column], colName: String): F[T]

    Permalink
    Attributes
    protected
    Definition Classes
    ColGetters
    Annotations
    @inline()

Concrete Value Members

  1. object Asc extends Order

    Permalink
  2. object AscNullsFirst extends Order

    Permalink
  3. object AscNullsLast extends Order

    Permalink
  4. object CName extends Serializable

    Permalink
  5. object CNameOrd extends Serializable

    Permalink
  6. object Desc extends Order

    Permalink
  7. object DescNullsFirst extends Order

    Permalink
  8. object DescNullsLast extends Order

    Permalink
  9. object Doric extends Serializable

    Permalink
  10. object DoricColumn extends ColGetters[NamedDoricColumn]

    Permalink
  11. object LeftDF extends ColGetters[LeftDoricColumn]

    Permalink
  12. object LiteralDoricColumn extends Serializable

    Permalink
  13. object NamedDoricColumn extends Serializable

    Permalink
  14. object RightDF extends ColGetters[RightDoricColumn]

    Permalink
  15. object SelectorWithSparkType extends SelectorLPI

    Permalink
    Definition Classes
    DStructs
  16. def andAgg(col: BooleanColumn): BooleanColumn

    Permalink

    Aggregate function: returns the AND value for a boolean column

    Aggregate function: returns the AND value for a boolean column

    Definition Classes
    AggregationColumns
  17. def aproxCountDistinct(colName: String): LongColumn

    Permalink

    Aggregate function: returns the approximate number of distinct items in a group.

    Aggregate function: returns the approximate number of distinct items in a group.

    Definition Classes
    AggregationColumns
    See also

    org.apache.spark.sql.functions.approx_count_distinct

  18. def aproxCountDistinct(colName: String, rsd: Double): LongColumn

    Permalink

    Aggregate function: returns the approximate number of distinct items in a group.

    Aggregate function: returns the approximate number of distinct items in a group.

    Definition Classes
    AggregationColumns
    See also

    org.apache.spark.sql.functions.approx_count_distinct

  19. def aproxCountDistinct(col: DoricColumn[_]): LongColumn

    Permalink

    Aggregate function: returns the approximate number of distinct items in a group.

    Aggregate function: returns the approximate number of distinct items in a group.

    Definition Classes
    AggregationColumns
    See also

    org.apache.spark.sql.functions.approx_count_distinct

  20. def aproxCountDistinct(col: DoricColumn[_], rsd: Double): LongColumn

    Permalink

    Aggregate function: returns the approximate number of distinct items in a group.

    Aggregate function: returns the approximate number of distinct items in a group.

    Definition Classes
    AggregationColumns
    See also

    org.apache.spark.sql.functions.approx_count_distinct

  21. def array[T](cols: DoricColumn[T]*)(implicit arg0: SparkType[T], arg1: ClassTag[T], lt: LiteralSparkType[Array[T]]): ArrayColumn[T]

    Permalink

    Creates a new array column.

    Creates a new array column. The input columns must all have the same data type.

    Definition Classes
    ArrayColumns
    To do

    scaladoc link (issue #135)

    See also

    org.apache.spark.sql.functions.array

  22. def avg[T](col: DoricColumn[T])(implicit arg0: NumericType[T]): DoubleColumn

    Permalink

    Aggregate function: returns the average of the values in a group.

    Aggregate function: returns the average of the values in a group.

    Definition Classes
    AggregationColumns
    See also

    org.apache.spark.sql.functions.avg

  23. def coalesce[T](cols: DoricColumn[T]*): DoricColumn[T]

    Permalink

    Returns the first column that is not null, or null if all inputs are null.

    Returns the first column that is not null, or null if all inputs are null.

    For example, coalesce(a, b, c) will return a if a is not null, or b if a is null and b is not null, or c if both a and b are null but c is not null.

    cols

    the DoricColumns to coalesce

    returns

    the first column that is not null, or null if all inputs are null.

    Definition Classes
    CommonColumns
    See also

    org.apache.spark.sql.functions.coalesce

  24. def col[T](colName: String)(implicit arg0: SparkType[T], location: Location): F[T]

    Permalink

    Retrieves a column with the provided name and the provided type.

    Retrieves a column with the provided name and the provided type.

    T

    the expected type of the column

    colName

    the name of the column to find.

    location

    error location.

    returns

    the column reference

    Definition Classes
    ColGetters
  25. def colArray[T](colName: String)(implicit arg0: ClassTag[T], location: Location, st: SparkType[Array[T]]): F[Array[T]]

    Permalink

    Retrieves a column with the provided name expecting it to be of array of T type.

    Retrieves a column with the provided name expecting it to be of array of T type.

    T

    the type of the elements of the array.

    colName

    the name of the column to find.

    location

    error location.

    returns

    the array of T column reference.

    Definition Classes
    ColGetters
  26. def colArrayInt(colName: String)(implicit location: Location): F[Array[Int]]

    Permalink

    Retrieves a column with the provided name expecting it to be of array of integers type.

    Retrieves a column with the provided name expecting it to be of array of integers type.

    colName

    the name of the column to find.

    location

    error location.

    returns

    the array of integers column reference.

    Definition Classes
    ColGetters
  27. def colArrayString(colName: String)(implicit location: Location): F[Array[String]]

    Permalink

    Retrieves a column with the provided name expecting it to be of array of string type.

    Retrieves a column with the provided name expecting it to be of array of string type.

    colName

    the name of the column to find.

    location

    error location.

    returns

    the array of string column reference.

    Definition Classes
    ColGetters
  28. def colBinary(colName: String)(implicit location: Location): F[Array[Byte]]

    Permalink

    Retrieves a column with the provided name expecting it to be of array of bytes type.

    Retrieves a column with the provided name expecting it to be of array of bytes type.

    colName

    the name of the column to find.

    location

    error location.

    returns

    the binary column reference.

    Definition Classes
    ColGetters
  29. def colBoolean(colName: String)(implicit location: Location): F[Boolean]

    Permalink

    Retrieves a column with the provided name expecting it to be of double type.

    Retrieves a column with the provided name expecting it to be of double type.

    colName

    the name of the column to find.

    location

    error location.

    returns

    the long column reference

    Definition Classes
    ColGetters
  30. def colDate(colName: String)(implicit location: Location): F[Date]

    Permalink

    Retrieves a column with the provided name expecting it to be of Date type.

    Retrieves a column with the provided name expecting it to be of Date type.

    colName

    the name of the column to find.

    location

    error location.

    returns

    the Date column reference

    Definition Classes
    ColGetters
  31. def colDouble(colName: String)(implicit location: Location): F[Double]

    Permalink

    Retrieves a column with the provided name expecting it to be of double type.

    Retrieves a column with the provided name expecting it to be of double type.

    colName

    the name of the column to find.

    location

    error location.

    returns

    the double column reference

    Definition Classes
    ColGetters
  32. def colFloat(colName: String)(implicit location: Location): F[Float]

    Permalink

    Retrieves a column with the provided name expecting it to be of float type.

    Retrieves a column with the provided name expecting it to be of float type.

    colName

    the name of the column to find.

    location

    error location.

    returns

    the float column reference

    Definition Classes
    ColGetters
  33. def colFromDF[T](colName: String, originDF: Dataset[_])(implicit arg0: SparkType[T], location: Location): F[T]

    Permalink

    Retrieves a column of the provided dataframe.

    Retrieves a column of the provided dataframe. Useful to prevent column ambiguity errors.

    T

    the type of the doric column.

    colName

    the name of the column to find.

    originDF

    the dataframe to force the column.

    location

    error location.

    returns

    the column of type T column reference.

    Definition Classes
    ColGetters
  34. def colInstant(colName: String)(implicit location: Location): F[Instant]

    Permalink

    Retrieves a column with the provided name expecting it to be of instant type.

    Retrieves a column with the provided name expecting it to be of instant type.

    colName

    the name of the column to find.

    location

    error location.

    returns

    the instant column reference

    Definition Classes
    ColGetters
  35. def colInt(colName: String)(implicit location: Location): F[Int]

    Permalink

    Retrieves a column with the provided name expecting it to be of integer type.

    Retrieves a column with the provided name expecting it to be of integer type.

    colName

    the name of the column to find.

    location

    error location.

    returns

    the integer column reference

    Definition Classes
    ColGetters
  36. def colLocalDate(colName: String)(implicit location: Location): F[LocalDate]

    Permalink

    Retrieves a column with the provided name expecting it to be of LocalDate type.

    Retrieves a column with the provided name expecting it to be of LocalDate type.

    colName

    the name of the column to find.

    location

    error location.

    returns

    the LocalDate column reference

    Definition Classes
    ColGetters
  37. def colLong(colName: String)(implicit location: Location): F[Long]

    Permalink

    Retrieves a column with the provided name expecting it to be of long type.

    Retrieves a column with the provided name expecting it to be of long type.

    colName

    the name of the column to find.

    location

    error location.

    returns

    the long column reference

    Definition Classes
    ColGetters
  38. def colMap[K, V](colName: String)(implicit arg0: SparkType[K], arg1: SparkType[V], location: Location): F[Map[K, V]]

    Permalink

    Retrieves a column with the provided name expecting it to be of map type.

    Retrieves a column with the provided name expecting it to be of map type.

    K

    the type of the keys of the map.

    V

    the type of the values of the map.

    colName

    the name of the column to find.

    location

    error location.

    returns

    the map column reference.

    Definition Classes
    ColGetters
  39. def colMapString[V](colName: String)(implicit arg0: SparkType[V], location: Location): F[Map[String, V]]

    Permalink

    Retrieves a column with the provided name expecting it to be of map type.

    Retrieves a column with the provided name expecting it to be of map type.

    V

    the type of the values of the map.

    colName

    the name of the column to find.

    location

    error location.

    returns

    the map column reference.

    Definition Classes
    ColGetters
  40. def colNull(colName: String)(implicit location: Location): F[Null]

    Permalink

    Retrieves a column with the provided name expecting it to be of null type.

    Retrieves a column with the provided name expecting it to be of null type.

    colName

    the name of the column to find.

    location

    error location.

    returns

    the null column reference

    Definition Classes
    ColGetters
  41. def colString(colName: String)(implicit location: Location): F[String]

    Permalink

    Retrieves a column with the provided name expecting it to be of string type.

    Retrieves a column with the provided name expecting it to be of string type.

    colName

    the name of the column to find.

    location

    error location.

    returns

    the string column reference

    Definition Classes
    ColGetters
  42. def colStruct(colName: String)(implicit location: Location): F[Row]

    Permalink

    Retrieves a column with the provided name expecting it to be of struct type.

    Retrieves a column with the provided name expecting it to be of struct type.

    colName

    the name of the column to find.

    location

    error location.

    returns

    the struct column reference.

    Definition Classes
    ColGetters
  43. def colTimestamp(colName: String)(implicit location: Location): F[Timestamp]

    Permalink

    Retrieves a column with the provided name expecting it to be of Timestamp type.

    Retrieves a column with the provided name expecting it to be of Timestamp type.

    colName

    the name of the column to find.

    location

    error location.

    returns

    the Timestamp column reference

    Definition Classes
    ColGetters
  44. def collectList[T](col: DoricColumn[T]): ArrayColumn[T]

    Permalink

    Aggregate function: returns a list of objects with duplicates.

    Aggregate function: returns a list of objects with duplicates.

    Definition Classes
    AggregationColumns
    Note

    The function is non-deterministic because the order of collected results depends on the order of the rows which may be non-deterministic after a shuffle.

    See also

    org.apache.spark.sql.functions.collect_list

  45. def collectSet[T](col: DoricColumn[T]): ArrayColumn[T]

    Permalink

    Aggregate function: returns a set of objects with duplicate elements eliminated.

    Aggregate function: returns a set of objects with duplicate elements eliminated.

    Definition Classes
    AggregationColumns
    Note

    The function is non-deterministic because the order of collected results depends on the order of the rows which may be non-deterministic after a shuffle.

    See also

    org.apache.spark.sql.functions.collect_set

  46. def concat(cols: StringColumn*): StringColumn

    Permalink

    Concatenate string columns to form a single one

    Concatenate string columns to form a single one

    cols

    the String DoricColumns to concatenate

    returns

    a reference of a single DoricColumn with all strings concatenated. If at least one is null will return null.

    Definition Classes
    StringColumns
    See also

    org.apache.spark.sql.functions.concat

  47. def concatArrays[T, F[_]](cols: DoricColumn[F[T]]*)(implicit arg0: CollectionType[F]): DoricColumn[F[T]]

    Permalink

    Concatenates multiple array columns together into a single column.

    Concatenates multiple array columns together into a single column.

    T

    The type of the elements of the arrays.

    cols

    the array columns, must be Arrays of the same type.

    returns

    Doric Column with the concatenation.

    Definition Classes
    ArrayColumns
    See also

    org.apache.spark.sql.functions.concat

  48. def concatBinary(col: BinaryColumn, cols: BinaryColumn*): BinaryColumn

    Permalink

    Concatenates multiple binary columns together into a single column.

    Concatenates multiple binary columns together into a single column.

    col

    the first binary column

    cols

    the binary columns

    returns

    Doric Column with the concatenation.

    Definition Classes
    BinaryColumns
    See also

    org.apache.spark.sql.functions.concat

  49. def concatMaps[K, V](col: MapColumn[K, V], cols: MapColumn[K, V]*): MapColumn[K, V]

    Permalink

    Returns the union of all the given maps.

    Returns the union of all the given maps.

    Definition Classes
    MapColumns
    See also

    org.apache.spark.sql.functions.map_concat

  50. def concatWs(sep: StringColumn, cols: StringColumn*): StringColumn

    Permalink

    Concatenates multiple input string columns together into a single string column, using the given separator.

    Concatenates multiple input string columns together into a single string column, using the given separator.

    Definition Classes
    StringColumns
    Example:
    1. df.withColumn("res", concatWs("-".lit, col("col1"), col("col2")))
        .show(false)
          +----+----+----+
          |col1|col2| res|
          +----+----+----+
          |   1|   1| 1-1|
          |null|   2|   2|
          |   3|null|   3|
          |null|null|    |
          +----+----+----+
    Note

    even if cols contain null columns, it prints remaining string columns (or empty string).

    See also

    org.apache.spark.sql.functions.concat_ws

  51. def correlation(col1: DoubleColumn, col2: DoubleColumn): DoubleColumn

    Permalink

    Aggregate function: returns the Pearson Correlation Coefficient for two columns.

    Aggregate function: returns the Pearson Correlation Coefficient for two columns.

    Definition Classes
    AggregationColumns
    See also

    org.apache.spark.sql.functions.corr

  52. def count(colName: CName): LongColumn

    Permalink

    Aggregate function: returns the number of items in a group.

    Aggregate function: returns the number of items in a group.

    Definition Classes
    AggregationColumns
    See also

    org.apache.spark.sql.functions.count

  53. def count(col: DoricColumn[_]): LongColumn

    Permalink

    Aggregate function: returns the number of items in a group.

    Aggregate function: returns the number of items in a group.

    Definition Classes
    AggregationColumns
    See also

    org.apache.spark.sql.functions.count

  54. def countDistinct(columnName: CName, columnNames: CName*): LongColumn

    Permalink

    Aggregate function: returns the number of distinct items in a group.

    Aggregate function: returns the number of distinct items in a group.

    Definition Classes
    AggregationColumns
    See also

    org.apache.spark.sql.functions.countDistinct

  55. def countDistinct(expr: DoricColumn[_], exprs: DoricColumn[_]*): LongColumn

    Permalink

    Aggregate function: returns the number of distinct items in a group.

    Aggregate function: returns the number of distinct items in a group.

    Definition Classes
    AggregationColumns
    See also

    org.apache.spark.sql.functions.countDistinct

  56. def covarPop(col1: DoubleColumn, col2: DoubleColumn): DoubleColumn

    Permalink

    Aggregate function: returns the population covariance for two columns.

    Aggregate function: returns the population covariance for two columns.

    Definition Classes
    AggregationColumns
    See also

    org.apache.spark.sql.functions.covar_pop

  57. def covarSamp(col1: DoubleColumn, col2: DoubleColumn): DoubleColumn

    Permalink

    Aggregate function: returns the sample covariance for two columns.

    Aggregate function: returns the sample covariance for two columns.

    Definition Classes
    AggregationColumns
    See also

    org.apache.spark.sql.functions.covar_samp

  58. def currentDate(): DateColumn

    Permalink

    Returns the current date at the start of query evaluation as a date column.

    Returns the current date at the start of query evaluation as a date column. All calls of current_date within the same query return the same value.

    Definition Classes
    DateColumns
    See also

    org.apache.spark.sql.functions.current_date

  59. def currentDateT[T]()(implicit arg0: DateType[T], arg1: SparkType[T]): DoricColumn[T]

    Permalink

    Returns the current date at the start of query evaluation as a date column typed with the provided T.

    Returns the current date at the start of query evaluation as a date column typed with the provided T. All calls of current_date within the same query return the same value.

    Definition Classes
    DateColumns
    See also

    org.apache.spark.sql.functions.current_date

  60. def currentTimestamp(): TimestampColumn

    Permalink

    Returns the current timestamp at the start of query evaluation as a timestamp column.

    Returns the current timestamp at the start of query evaluation as a timestamp column. All calls of current_timestamp within the same query return the same value.

    Definition Classes
    TimestampColumns
    See also

    org.apache.spark.sql.functions.current_timestamp

  61. def currentTimestampT[T]()(implicit arg0: TimestampType[T], arg1: SparkType[T]): DoricColumn[T]

    Permalink

    Returns the current timestamp at the start of query evaluation as a timestamp column.

    Returns the current timestamp at the start of query evaluation as a timestamp column. All calls of current_timestamp within the same query return the same value.

    Definition Classes
    TimestampColumns
    See also

    org.apache.spark.sql.functions.current_timestamp

  62. def first[T](col: DoricColumn[T], ignoreNulls: Boolean): DoricColumn[T]

    Permalink

    Aggregate function: returns the first value in a group.

    Aggregate function: returns the first value in a group.

    The function by default returns the first values it sees. It will return the first non-null value it sees when ignoreNulls is set to true. If all values are null, then null is returned.

    Definition Classes
    AggregationColumns
    Note

    The function is non-deterministic because its results depends on the order of the rows which may be non-deterministic after a shuffle.

    See also

    org.apache.spark.sql.functions.first

  63. def first[T](col: DoricColumn[T]): DoricColumn[T]

    Permalink

    Aggregate function: returns the first value in a group.

    Aggregate function: returns the first value in a group.

    The function by default returns the first values it sees. It will return the first non-null value it sees when ignoreNulls is set to true. If all values are null, then null is returned.

    Definition Classes
    AggregationColumns
    Note

    The function is non-deterministic because its results depends on the order of the rows which may be non-deterministic after a shuffle.

    See also

    org.apache.spark.sql.functions.first

  64. def formatString(format: StringColumn, arguments: DoricColumn[_]*): StringColumn

    Permalink

    Formats the arguments in printf-style and returns the result as a string column.

    Formats the arguments in printf-style and returns the result as a string column.

    format

    Printf format

    arguments

    the String DoricColumns to format

    returns

    Formats the arguments in printf-style and returns the result as a string column.

    Definition Classes
    StringColumns
    See also

    org.apache.spark.sql.functions.format_string

  65. def greatest[T](col: DoricColumn[T], cols: DoricColumn[T]*): DoricColumn[T]

    Permalink

    Returns the greatest value of the list of values, skipping null values.

    Returns the greatest value of the list of values, skipping null values. This function takes at least 2 parameters. It will return null iff all parameters are null.

    Definition Classes
    CommonColumns
    Note

    skips null values

    See also

    org.apache.spark.sql.functions.greatest

  66. def grouping(columnName: CName): ByteColumn

    Permalink

    Aggregate function: indicates whether a specified column in a GROUP BY list is aggregated or not, returns 1 for aggregated or 0 for not aggregated in the result set.

    Aggregate function: indicates whether a specified column in a GROUP BY list is aggregated or not, returns 1 for aggregated or 0 for not aggregated in the result set.

    Definition Classes
    AggregationColumns
    See also

    org.apache.spark.sql.functions.grouping

  67. def grouping(col: DoricColumn[_]): ByteColumn

    Permalink

    Aggregate function: indicates whether a specified column in a GROUP BY list is aggregated or not, returns 1 for aggregated or 0 for not aggregated in the result set.

    Aggregate function: indicates whether a specified column in a GROUP BY list is aggregated or not, returns 1 for aggregated or 0 for not aggregated in the result set.

    Definition Classes
    AggregationColumns
    See also

    org.apache.spark.sql.functions.grouping

  68. def groupingId(colName: CName, colNames: CName*): LongColumn

    Permalink

    Aggregate function: returns the level of grouping, equals to

    Aggregate function: returns the level of grouping, equals to

    Definition Classes
    AggregationColumns
    Example:
    1. (grouping(c1) <<; (n-1)) + (grouping(c2) <<; (n-2)) + ... + grouping(cn)
    Note

    The list of columns should match with grouping columns exactly, or empty (means all the grouping columns).

    See also

    org.apache.spark.sql.functions.grouping_id

  69. def groupingId(col: DoricColumn[_], cols: DoricColumn[_]*): LongColumn

    Permalink

    Aggregate function: returns the level of grouping, equals to

    Aggregate function: returns the level of grouping, equals to

    Definition Classes
    AggregationColumns
    Example:
    1. (grouping(c1) <<; (n-1)) + (grouping(c2) <<; (n-2)) + ... + grouping(cn)
    Note

    The list of columns should match with grouping columns exactly, or empty (means all the grouping columns).

    See also

    org.apache.spark.sql.functions.grouping_id

  70. def hash(cols: DoricColumn[_]*): IntegerColumn

    Permalink

    Calculates the hash code of given columns, and returns the result as an integer column.

    Calculates the hash code of given columns, and returns the result as an integer column.

    Definition Classes
    CommonColumns
    See also

    org.apache.spark.sql.functions.hash

  71. def inputFileName(): StringColumn

    Permalink

    Creates a string column for the file name of the current Spark task.

    Creates a string column for the file name of the current Spark task.

    Definition Classes
    StringColumns
    See also

    org.apache.spark.sql.functions.input_file_name

  72. def kurtosis(col: DoubleColumn): DoubleColumn

    Permalink

    Aggregate function: returns the kurtosis of the values in a group.

    Aggregate function: returns the kurtosis of the values in a group.

    Definition Classes
    AggregationColumns
    See also

    org.apache.spark.sql.functions.kurtosis

  73. def last[T](col: DoricColumn[T], ignoreNulls: Boolean): DoricColumn[T]

    Permalink

    Aggregate function: returns the last value in a group.

    Aggregate function: returns the last value in a group.

    The function by default returns the last values it sees. It will return the last non-null value it sees when ignoreNulls is set to true. If all values are null, then null is returned.

    Definition Classes
    AggregationColumns
    Note

    The function is non-deterministic because its results depends on the order of the rows which may be non-deterministic after a shuffle.

    See also

    org.apache.spark.sql.functions.last

  74. def last[T](col: DoricColumn[T]): DoricColumn[T]

    Permalink

    Aggregate function: returns the last value in a group.

    Aggregate function: returns the last value in a group.

    The function by default returns the last values it sees. It will return the last non-null value it sees when ignoreNulls is set to true. If all values are null, then null is returned.

    Definition Classes
    AggregationColumns
    Note

    The function is non-deterministic because its results depends on the order of the rows which may be non-deterministic after a shuffle.

    See also

    org.apache.spark.sql.functions.last

  75. def least[T](col: DoricColumn[T], cols: DoricColumn[T]*): DoricColumn[T]

    Permalink

    Returns the least value of the list of values, skipping null values.

    Returns the least value of the list of values, skipping null values. This function takes at least 2 parameters. It will return null iff all parameters are null.

    Definition Classes
    CommonColumns
    Note

    skips null values

    See also

    org.apache.spark.sql.functions.least

  76. def list[T](cols: DoricColumn[T]*): DoricColumn[List[T]]

    Permalink

    Creates a new list column.

    Creates a new list column. The input columns must all have the same data type.

    Definition Classes
    ArrayColumns
    To do

    scaladoc link (issue #135)

    See also

    org.apache.spark.sql.functions.array

  77. def lit[L](litv: L)(implicit arg0: SparkType[L], arg1: LiteralSparkType[L], l: Location): LiteralDoricColumn[L]

    Permalink

    Creates a literal with the provided value.

    Creates a literal with the provided value.

    L

    The type of the literal.

    litv

    the element to create as a literal.

    returns

    A doric column that represent the literal value and the same type as the value.

    Definition Classes
    LiteralConversions
  78. def map[K, V](first: (DoricColumn[K], DoricColumn[V]), rest: (DoricColumn[K], DoricColumn[V])*): MapColumn[K, V]

    Permalink

    Creates a new map column.

    Creates a new map column. The input is formed by tuples of key and the corresponding value.

    K

    the type of the keys of the Map

    V

    the type of the values of the Map

    first

    a pair of key value DoricColumns

    rest

    the rest of pairs of key and corresponding Values.

    returns

    the DoricColumn of the corresponding Map type

    Definition Classes
    MapColumns
    See also

    org.apache.spark.sql.functions.map

  79. def mapFromArrays[K, V](keys: DoricColumn[Array[K]], values: DoricColumn[Array[V]]): MapColumn[K, V]

    Permalink

    Creates a new map column.

    Creates a new map column. The array in the first column is used for keys. The array in the second column is used for values. All elements in the array for key should not be null.

    K

    the type of the Array elements of the keys.

    V

    the type of the Array elements of the value.

    keys

    the array to create the keys.

    values

    the array to create the values.

    returns

    an DoricColumn of type Map of the keys and values.

    Definition Classes
    MapColumns
    See also

    org.apache.spark.sql.functions.map_from_arrays

  80. def matchToType[T](colName: String)(implicit arg0: SparkType[T]): EmptyTypeMatcher[T]

    Permalink
    Definition Classes
    TypeMatcher
  81. def max[T](col: DoricColumn[T]): DoricColumn[T]

    Permalink

    Aggregate function: returns the maximum value of the expression in a group.

    Aggregate function: returns the maximum value of the expression in a group.

    Definition Classes
    AggregationColumns
    See also

    org.apache.spark.sql.functions.max

  82. def mean[T](col: DoricColumn[T])(implicit arg0: NumericType[T]): DoubleColumn

    Permalink

    Aggregate function: returns the maximum value of the expression in a group.

    Aggregate function: returns the maximum value of the expression in a group.

    Definition Classes
    AggregationColumns
    See also

    org.apache.spark.sql.functions.mean

  83. def min[T](col: DoricColumn[T]): DoricColumn[T]

    Permalink

    Aggregate function: returns the maximum value of the expression in a group.

    Aggregate function: returns the maximum value of the expression in a group.

    Definition Classes
    AggregationColumns
    See also

    org.apache.spark.sql.functions.min

  84. lazy val minorScalaVersion: Int

    Permalink
  85. def monotonicallyIncreasingId(): LongColumn

    Permalink

    A column expression that generates monotonically increasing 64-bit integers.

    A column expression that generates monotonically increasing 64-bit integers.

    The generated ID is guaranteed to be monotonically increasing and unique, but not consecutive. The current implementation puts the partition ID in the upper 31 bits, and the record number within each partition in the lower 33 bits. The assumption is that the data frame has less than 1 billion partitions, and each partition has less than 8 billion records.

    Definition Classes
    NumericColumns
    Example:
    1. consider a DataFrame with two partitions, each with 3 records. This expression would return the following IDs:

      0, 1, 2, 8589934592 (1L << 33), 8589934593, 8589934594.
    See also

    org.apache.spark.sql.functions.monotonically_increasing_id

  86. def not(col: BooleanColumn): BooleanColumn

    Permalink

    Inversion of boolean expression, i.e.

    Inversion of boolean expression, i.e. NOT.

    Definition Classes
    BooleanColumns
    See also

    org.apache.spark.sql.functions.not

  87. def orAgg(col: BooleanColumn): BooleanColumn

    Permalink

    Aggregate function: returns the OR value for a boolean column

    Aggregate function: returns the OR value for a boolean column

    Definition Classes
    AggregationColumns
  88. def random(seed: LongColumn): DoubleColumn

    Permalink

    Generate a random column with independent and identically distributed (i.i.d.) samples uniformly distributed in [0.0, 1.0).

    Generate a random column with independent and identically distributed (i.i.d.) samples uniformly distributed in [0.0, 1.0).

    Definition Classes
    NumericColumns
    Note

    The function is non-deterministic in general case.

    See also

    org.apache.spark.sql.functions.rand

  89. def random(): DoubleColumn

    Permalink

    Generate a random column with independent and identically distributed (i.i.d.) samples uniformly distributed in [0.0, 1.0).

    Generate a random column with independent and identically distributed (i.i.d.) samples uniformly distributed in [0.0, 1.0).

    Definition Classes
    NumericColumns
    Note

    The function is non-deterministic in general case.

    See also

    org.apache.spark.sql.functions.rand

  90. def randomN(seed: LongColumn): DoubleColumn

    Permalink

    Generate a column with independent and identically distributed (i.i.d.) samples from the standard normal distribution.

    Generate a column with independent and identically distributed (i.i.d.) samples from the standard normal distribution.

    Definition Classes
    NumericColumns
    Note

    The function is non-deterministic in general case.

    See also

    org.apache.spark.sql.functions.randn

  91. def randomN(): DoubleColumn

    Permalink

    Generate a column with independent and identically distributed (i.i.d.) samples from the standard normal distribution.

    Generate a column with independent and identically distributed (i.i.d.) samples from the standard normal distribution.

    Definition Classes
    NumericColumns
    Note

    The function is non-deterministic in general case.

    See also

    org.apache.spark.sql.functions.randn

  92. object row extends Dynamic

    Permalink

    The object row stands for the top-level row of the DataFrame.

    The object row stands for the top-level row of the DataFrame.

    Definition Classes
    ColGetters
  93. package sem

    Permalink
  94. def skewness[T](col: DoricColumn[T])(implicit arg0: NumericType[T]): DoubleColumn

    Permalink

    Aggregate function: returns the skewness of the values in a group.

    Aggregate function: returns the skewness of the values in a group.

    Definition Classes
    AggregationColumns
    See also

    org.apache.spark.sql.functions.skewness

  95. def sparkAgg(relationalGroupedDataset: RelationalGroupedDataset, expr: DoricColumn[_], exprs: DoricColumn[_]*): DoricValidated[DataFrame]

    Permalink
    Definition Classes
    RelationalGroupedDatasetDoricInterface
  96. def sparkCube(df: DataFrame, cols: DoricColumn[_]*): DoricValidated[RelationalGroupedDataset]

    Permalink
    Attributes
    protected
    Definition Classes
    RelationalGroupedDatasetDoricInterface
  97. def sparkGroupBy(df: DataFrame, cols: DoricColumn[_]*): DoricValidated[RelationalGroupedDataset]

    Permalink
    Attributes
    protected
    Definition Classes
    RelationalGroupedDatasetDoricInterface
  98. def sparkPartitionId(): IntegerColumn

    Permalink

    Partition ID.

    Partition ID.

    Definition Classes
    NumericColumns
    Note

    This is non-deterministic because it depends on data partitioning and task scheduling.

    See also

    org.apache.spark.sql.functions.spark_partition_id

  99. def sparkPivot[T](relationalGroupedDataset: RelationalGroupedDataset, expr: DoricColumn[T], values: Seq[T]): DoricValidated[RelationalGroupedDataset]

    Permalink
    Definition Classes
    RelationalGroupedDatasetDoricInterface
  100. def sparkRollup(df: DataFrame, cols: DoricColumn[_]*): DoricValidated[RelationalGroupedDataset]

    Permalink
    Attributes
    protected
    Definition Classes
    RelationalGroupedDatasetDoricInterface
  101. def sparkTaskName(): StringColumn

    Permalink

    Creates a string column for the file name of the current Spark task.

    Creates a string column for the file name of the current Spark task.

    Definition Classes
    StringColumns
    Annotations
    @inline()
    See also

    inputFileName

  102. def stdDev[T](col: DoricColumn[T])(implicit arg0: NumericType[T]): DoubleColumn

    Permalink

    Aggregate function: alias for stddev_samp.

    Aggregate function: alias for stddev_samp.

    Definition Classes
    AggregationColumns
    See also

    org.apache.spark.sql.functions.stddev

  103. def stdDevPop[T](col: DoricColumn[T])(implicit arg0: NumericType[T]): DoubleColumn

    Permalink

    Aggregate function: returns the population standard deviation of the expression in a group.

    Aggregate function: returns the population standard deviation of the expression in a group.

    Definition Classes
    AggregationColumns
    See also

    org.apache.spark.sql.functions.stddev_pop

  104. def stdDevSamp[T](col: DoricColumn[T])(implicit arg0: NumericType[T]): DoubleColumn

    Permalink

    Aggregate function: returns the sample standard deviation of the expression in a group.

    Aggregate function: returns the sample standard deviation of the expression in a group.

    Definition Classes
    AggregationColumns
    See also

    org.apache.spark.sql.functions.stddev_samp

  105. def struct(cols: DoricColumn[_]*): RowColumn

    Permalink

    Creates a struct with the columns

    Creates a struct with the columns

    cols

    the columns that will form the struct

    returns

    A DStruct DoricColumn.

    Definition Classes
    DStructs
  106. def sum[T](col: DoricColumn[T])(implicit nt: NumericType[T]): DoricColumn[Sum]

    Permalink

    Aggregate function: returns the sum of all values in the expression.

    Aggregate function: returns the sum of all values in the expression.

    Definition Classes
    AggregationColumns
    See also

    org.apache.spark.sql.functions.sum

  107. def sumDistinct[T](col: DoricColumn[T])(implicit nt: NumericType[T]): DoricColumn[Sum]

    Permalink

    Aggregate function: returns the sum of distinct values in the expression.

    Aggregate function: returns the sum of distinct values in the expression.

    Definition Classes
    AggregationColumns
    See also

    org.apache.spark.sql.functions.sumDistinct

  108. package syntax

    Permalink
  109. package types

    Permalink
  110. def unixTimestamp(): LongColumn

    Permalink

    Returns the current Unix timestamp (in seconds) as a long.

    Returns the current Unix timestamp (in seconds) as a long.

    Definition Classes
    NumericColumns
    Note

    All calls of unix_timestamp within the same query return the same value (i.e. the current timestamp is calculated at the start of query evaluation).

    See also

    org.apache.spark.sql.functions.unix_timestamp

  111. def varPop[T](col: DoricColumn[T])(implicit arg0: NumericType[T]): DoubleColumn

    Permalink

    Aggregate function: returns the population variance of the values in a group.

    Aggregate function: returns the population variance of the values in a group.

    Definition Classes
    AggregationColumns
    See also

    org.apache.spark.sql.functions.var_pop

  112. def varSamp[T](col: DoricColumn[T])(implicit arg0: NumericType[T]): DoubleColumn

    Permalink

    Aggregate function: returns the unbiased variance of the values in a group.

    Aggregate function: returns the unbiased variance of the values in a group.

    Definition Classes
    AggregationColumns
    See also

    org.apache.spark.sql.functions.var_samp

  113. def variance[T](col: DoricColumn[T])(implicit arg0: NumericType[T]): DoubleColumn

    Permalink

    Aggregate function: alias for var_samp.

    Aggregate function: alias for var_samp.

    Definition Classes
    AggregationColumns
    See also

    org.apache.spark.sql.functions.variance

  114. def when[T]: WhenBuilder[T]

    Permalink

    Initialize a when builder

    Initialize a when builder

    T

    the type of the returnign DoricColumn

    returns

    WhenBuilder instance to add the required logic.

    Definition Classes
    ControlStructures

Inherited from All

Inherited from SortingOps

Inherited from CollectOps

Inherited from JoinOps

Inherited from TransformOps

Inherited from AggregationOps

Inherited from RelationalGroupedDatasetDoricInterface

Inherited from All

Inherited from NumericColumns2_31

Inherited from StringColumn24

Inherited from BinaryColumns30_31

Inherited from Interpolators

Inherited from BinaryColumns

Inherited from doric.syntax.CNameOps

Inherited from AggregationColumns

Inherited from ControlStructures

Inherited from StringColumns

Inherited from BooleanColumns

Inherited from TimestampColumns

Inherited from DateColumns

Inherited from NumericColumns

Inherited from MapColumns

Inherited from LiteralConversions

Inherited from DStructs

Inherited from CommonColumns

Inherited from ColGetters[NamedDoricColumn]

Inherited from TypeMatcher

Inherited from ArrayColumns

Inherited from AnyRef

Inherited from Any

Aggregation Any Type

Aggregation Boolean Type

Aggregation Double Type

Aggregation Numeric Type

All Types

Array Type

Binary Type

Boolean Type

Control structure

Date Type

Map Type

Numeric Type

String Type

Struct Type

Timestamp Type

Ungrouped