Packages

package libs

Linear Supertypes
FixedFileFormatImplicits, SparkFunctions, DataHelpers, Component, UDFUtils, LazyLogging, Serializable, Serializable, RestAPIUtils, ProphecyDataFrame, AnyRef, Any
Ordering
  1. Alphabetic
  2. By Inheritance
Inherited
  1. libs
  2. FixedFileFormatImplicits
  3. SparkFunctions
  4. DataHelpers
  5. Component
  6. UDFUtils
  7. LazyLogging
  8. Serializable
  9. Serializable
  10. RestAPIUtils
  11. ProphecyDataFrame
  12. AnyRef
  13. Any
  1. Hide All
  2. Show All
Visibility
  1. Public
  2. All

Type Members

  1. type Aggregate = Dataset[Row]
    Definition Classes
    Component
  2. class CDC extends AnyRef
  3. class CLIConf extends ScallopConf
  4. trait Component extends AnyRef
  5. case class UsesDataset(id: String, version: Int = -1) extends Annotation with StaticAnnotation with Product with Serializable
    Definition Classes
    Component
  6. case class UsesRuleset(id: String) extends Annotation with StaticAnnotation with Product with Serializable
    Definition Classes
    Component
  7. case class Visual(id: String = "ID", label: String = "Label", x: Long = 0, y: Long = 0, phase: Int = 0, mode: String = "batch", interimMode: String = "full", detailedStats: Boolean = false) extends Annotation with StaticAnnotation with Product with Serializable
    Definition Classes
    Component
  8. trait ConfigBase extends AnyRef
  9. abstract class ConfigurationFactory[C <: ConfigBase] extends AnyRef
  10. type CreateData = Dataset[Row]
    Definition Classes
    Component
  11. type DataFrame1 = Dataset[Row]
    Definition Classes
    Component
  12. type DataFrame10 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  13. type DataFrame11 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  14. type DataFrame12 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  15. type DataFrame13 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  16. type DataFrame14 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  17. type DataFrame15 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  18. type DataFrame16 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  19. type DataFrame17 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  20. type DataFrame18 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  21. type DataFrame19 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  22. type DataFrame2 = (DataFrame, DataFrame)
    Definition Classes
    Component
  23. type DataFrame20 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  24. type DataFrame21 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  25. type DataFrame22 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  26. type DataFrame3 = (DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  27. type DataFrame4 = (DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  28. type DataFrame5 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  29. type DataFrame6 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  30. type DataFrame7 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  31. type DataFrame8 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  32. type DataFrame9 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  33. trait DataHelpers extends LazyLogging

    Helper Utilities for reading/writing data from/to different data sources.

  34. type DataQualityTest = Dataset[Row]
    Definition Classes
    Component
  35. type DatabaseInput = Dataset[Row]
    Definition Classes
    Component
  36. type Deduplicate = Dataset[Row]
    Definition Classes
    Component
  37. case class Description(comment: String) extends Annotation with StaticAnnotation with Product with Serializable
  38. implicit class ExtendedDataFrameGlobal extends ExtendedDataFrame
  39. implicit class ExtendedStreamingTargetGlobal extends ExtendedStreamingTarget
  40. trait FFAST extends Positional
  41. case class FFCompoundSchemaRow(compound: FFCompoundType, rows: Seq[FFSchemaRow]) extends FFSchemaRow with Product with Serializable
  42. sealed trait FFCompoundType extends FFAST
  43. case class FFConditionalSchemaRow(condition: String, schemaRow: FFSchemaRow) extends FFSchemaRow with Product with Serializable
  44. sealed trait FFDataFormat extends FFAST
  45. case class FFDateFormat(name: FFTypeName, format: Option[String], miscProperties: Map[String, Any] = Map()) extends FFDataFormat with Product with Serializable
  46. case class FFDateTimeFormat(name: FFTypeName, format: Option[String], miscProperties: Map[String, Any] = Map()) extends FFDataFormat with Product with Serializable
  47. sealed trait FFDefaultVal extends FFAST
  48. case class FFDoubleDefaultVal(value: Double) extends FFDefaultVal with Product with Serializable
  49. case class FFExpressionDefaultVal(value: CustomExpression) extends FFDefaultVal with Product with Serializable
  50. case class FFIncludeFileRow(filePath: String) extends FFSchemaRow with Product with Serializable
  51. case class FFIntDefaultVal(value: Int) extends FFDefaultVal with Product with Serializable
  52. case class FFNoDefaultVal() extends FFDefaultVal with Product with Serializable
  53. case class FFNullDefaultVal(value: Option[Any] = None) extends FFDefaultVal with Product with Serializable
  54. case class FFNumberArrayFormat(name: FFTypeName, precision: Option[Int], scale: Option[Int], arraySizeInfo: Option[String], miscProperties: Map[String, Any] = ...) extends FFDataFormat with Product with Serializable
  55. case class FFNumberFormat(name: FFTypeName, precision: Option[Int], scale: Option[Int], miscProperties: Map[String, Any] = ...) extends FFDataFormat with Product with Serializable
  56. case class FFRecordType(startType: String) extends FFAST with Product with Serializable
  57. case class FFSchemaRecord(recordType: String, rows: Seq[FFSchemaRow]) extends FFAST with Product with Serializable
  58. sealed trait FFSchemaRow extends FFAST
  59. case class FFSimpleSchemaList(rows: Seq[FFSimpleSchemaRow]) extends FFSchemaRow with Product with Serializable
  60. case class FFSimpleSchemaRow(name: String, format: FFDataFormat, value: FFDefaultVal) extends FFSchemaRow with Product with Serializable
  61. case class FFStringArrayFormat(name: FFTypeName, precision: Option[Int], arraySizeInfo: Option[String]) extends FFDataFormat with Product with Serializable
  62. case class FFStringDefaultVal(value: String) extends FFDefaultVal with Product with Serializable
  63. case class FFStringFormat(name: FFTypeName, precision: Option[Int], props: Option[Map[String, String]] = None) extends FFDataFormat with Product with Serializable
  64. case class FFStructArrayType(name1: String, arraySizeInfo: Option[String]) extends FFCompoundType with Product with Serializable
  65. case class FFStructFormat(name: FFTypeName, precision: Option[Int]) extends FFDataFormat with Product with Serializable
  66. case class FFStructType(name1: String) extends FFCompoundType with Product with Serializable
  67. case class FFTypeName(name: String, delimiter: Option[String]) extends FFAST with Product with Serializable
  68. case class FFTypeNameWithProperties(name: String, delimiter: Option[String], miscProperties: Map[String, Any] = Map("packed" → false)) extends FFAST with Product with Serializable
  69. case class FFUnionType(name: Option[String] = None) extends FFCompoundType with Product with Serializable
  70. case class FFUnknownFormat(name: FFTypeName, arraySizeInfo: Option[String]) extends FFDataFormat with Product with Serializable
  71. case class FFVoidFormat(name: FFTypeName, size: Option[Int]) extends FFDataFormat with Product with Serializable
  72. type FileInput = Dataset[Row]
    Definition Classes
    Component
  73. type FileIntermediate = Dataset[Row]
    Definition Classes
    Component
  74. type FileOutput = Unit
    Definition Classes
    Component
  75. type Filter = Dataset[Row]
    Definition Classes
    Component
  76. class FixedFileFormat extends FileFormat with DataSourceRegister with Serializable
  77. implicit class FixedFileFormatDataFrameGlobal extends FixedFileFormatDataFrame
  78. trait FixedFileFormatImplicits extends AnyRef
  79. implicit class FixedFileFormatDataFrame extends AnyRef
    Definition Classes
    FixedFileFormatImplicits
  80. implicit class FixedFileFormatSpark extends AnyRef
    Definition Classes
    FixedFileFormatImplicits
  81. type FixedFileOutput = Unit
    Definition Classes
    Component
  82. class FixedFormatOutputWriter extends OutputWriter
  83. type FlattenSchema = Dataset[Row]
    Definition Classes
    Component
  84. type Generate = Dataset[Row]
    Definition Classes
    Component
  85. type HashPartition = Dataset[Row]
    Definition Classes
    Component
  86. type Join = Dataset[Row]
    Definition Classes
    Component
  87. type Limit = Dataset[Row]
    Definition Classes
    Component
  88. type Lookup = UserDefinedFunction
    Definition Classes
    Component
  89. case class LookupDataset(datasetId: String, columnName: String) extends Annotation with StaticAnnotation with Product with Serializable
  90. type LookupFileInput = UserDefinedFunction
    Definition Classes
    Component
  91. type LookupUnit = Unit
    Definition Classes
    Component
  92. trait LookupUtils extends AnyRef
  93. class MDumpReader extends AnyRef
  94. type MultiFileRead = Dataset[Row]
    Definition Classes
    Component
  95. type MultiFileWrite = Unit
    Definition Classes
    Component
  96. type MultiFileWriteUnit = Unit
    Definition Classes
    Component
  97. type MultiJoin = Dataset[Row]
    Definition Classes
    Component
  98. type Normalize = Dataset[Row]
    Definition Classes
    Component
  99. type OrderBy = Dataset[Row]
    Definition Classes
    Component
  100. type OrderByPartition = Dataset[Row]
    Definition Classes
    Component
  101. type Prepare = Dataset[Row]
    Definition Classes
    Component
  102. type ReadSV = Dataset[Row]
    Definition Classes
    Component
  103. type Reformat = Dataset[Row]
    Definition Classes
    Component
  104. type Repartition = Dataset[Row]
    Definition Classes
    Component
  105. trait RestAPIUtils extends AnyRef

    Spark utilities for handling rest api connections.

  106. type RoundRobinPartition = Dataset[Row]
    Definition Classes
    Component
  107. type RowDistributor = Dataset[Row]
    Definition Classes
    Component
  108. type RowDistributor1 = Dataset[Row]
    Definition Classes
    Component
  109. type RowDistributor10 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  110. type RowDistributor11 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  111. type RowDistributor12 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  112. type RowDistributor13 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  113. type RowDistributor14 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  114. type RowDistributor15 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  115. type RowDistributor16 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  116. type RowDistributor17 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  117. type RowDistributor18 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  118. type RowDistributor19 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  119. type RowDistributor2 = (DataFrame, DataFrame)
    Definition Classes
    Component
  120. type RowDistributor20 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  121. type RowDistributor21 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  122. type RowDistributor22 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  123. type RowDistributor3 = (DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  124. type RowDistributor4 = (DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  125. type RowDistributor5 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  126. type RowDistributor6 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  127. type RowDistributor7 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  128. type RowDistributor8 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  129. type RowDistributor9 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  130. type Ruleset = Dataset[Row]
    Definition Classes
    Component
  131. type SQLStatement = Dataset[Row]
    Definition Classes
    Component
  132. type SQLStatement1 = Dataset[Row]
    Definition Classes
    Component
  133. type SQLStatement10 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  134. type SQLStatement11 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  135. type SQLStatement12 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  136. type SQLStatement13 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  137. type SQLStatement14 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  138. type SQLStatement15 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  139. type SQLStatement16 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  140. type SQLStatement17 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  141. type SQLStatement18 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  142. type SQLStatement19 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  143. type SQLStatement2 = (DataFrame, DataFrame)
    Definition Classes
    Component
  144. type SQLStatement20 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  145. type SQLStatement21 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  146. type SQLStatement22 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  147. type SQLStatement3 = (DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  148. type SQLStatement4 = (DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  149. type SQLStatement5 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  150. type SQLStatement6 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  151. type SQLStatement7 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  152. type SQLStatement8 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  153. type SQLStatement9 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  154. type SQLStatementUnit = Unit
    Definition Classes
    Component
  155. type Scan = Dataset[Row]
    Definition Classes
    Component
  156. type SchemaTransformer = Dataset[Row]
    Definition Classes
    Component
  157. type Script = Dataset[Row]
    Definition Classes
    Component
  158. type Script1 = Dataset[Row]
    Definition Classes
    Component
  159. type Script10 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  160. type Script11 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  161. type Script12 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  162. type Script13 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  163. type Script14 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  164. type Script15 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  165. type Script16 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  166. type Script17 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  167. type Script18 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  168. type Script19 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  169. type Script2 = (DataFrame, DataFrame)
    Definition Classes
    Component
  170. type Script20 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  171. type Script21 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  172. type Script22 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  173. type Script3 = (DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  174. type Script4 = (DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  175. type Script5 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  176. type Script6 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  177. type Script7 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  178. type Script8 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  179. type Script9 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  180. type ScriptUnit = Unit
    Definition Classes
    Component
  181. type Select = Dataset[Row]
    Definition Classes
    Component
  182. type Sequence = Dataset[Row]
    Definition Classes
    Component
  183. type SetOperation = Dataset[Row]
    Definition Classes
    Component
  184. type Source = Dataset[Row]
    Definition Classes
    Component
  185. trait SparkFunctions extends AnyRef

    Library of all spark functions which implements different abinitio functions used in abinitio workflows.

  186. class StringAsStream extends Serializable
    Definition Classes
    SparkFunctions
  187. type StreamingTarget = StreamingQuery
    Definition Classes
    Component
  188. type SubGraph = Dataset[Row]
    Definition Classes
    Component
  189. type SubGraph1 = Dataset[Row]
    Definition Classes
    Component
  190. type SubGraph10 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  191. type SubGraph11 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  192. type SubGraph12 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  193. type SubGraph13 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  194. type SubGraph14 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  195. type SubGraph15 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  196. type SubGraph16 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  197. type SubGraph17 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  198. type SubGraph18 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  199. type SubGraph19 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  200. type SubGraph2 = (DataFrame, DataFrame)
    Definition Classes
    Component
  201. type SubGraph20 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  202. type SubGraph21 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  203. type SubGraph22 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  204. type SubGraph3 = (DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  205. type SubGraph4 = (DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  206. type SubGraph5 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  207. type SubGraph6 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  208. type SubGraph7 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  209. type SubGraph8 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  210. type SubGraph9 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  211. type SubGraphUnit = Unit
    Definition Classes
    Component
  212. type Subgraph = Dataset[Row]
    Definition Classes
    Component
  213. type Subgraph1 = Dataset[Row]
    Definition Classes
    Component
  214. type Subgraph10 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  215. type Subgraph11 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  216. type Subgraph12 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  217. type Subgraph13 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  218. type Subgraph14 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  219. type Subgraph15 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  220. type Subgraph16 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  221. type Subgraph17 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  222. type Subgraph18 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  223. type Subgraph19 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  224. type Subgraph2 = (DataFrame, DataFrame)
    Definition Classes
    Component
  225. type Subgraph20 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  226. type Subgraph21 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  227. type Subgraph22 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  228. type Subgraph3 = (DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  229. type Subgraph4 = (DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  230. type Subgraph5 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  231. type Subgraph6 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  232. type Subgraph7 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  233. type Subgraph8 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  234. type Subgraph9 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  235. type SubgraphUnit = Unit
    Definition Classes
    Component
  236. type Target = Unit
    Definition Classes
    Component
  237. trait UDFUtils extends RestAPIUtils with Serializable with LazyLogging

    Utility class with different UDFs to take care of miscellaneous tasks.

  238. type UnionAll = Dataset[Row]
    Definition Classes
    Component
  239. type Visualize = Unit
    Definition Classes
    Component
  240. type WindowFunction = Dataset[Row]
    Definition Classes
    Component
  241. implicit class ExtendedDataFrame extends AnyRef
    Definition Classes
    ProphecyDataFrame
  242. implicit class ExtendedStreamingTarget extends AnyRef
    Definition Classes
    ProphecyDataFrame
  243. implicit class ProphecyDataFrameReader extends AnyRef
    Definition Classes
    ProphecyDataFrame
  244. implicit class ProphecyDataFrameWriter[T] extends AnyRef
    Definition Classes
    ProphecyDataFrame

Abstract Value Members

  1. abstract def getClass(): Class[_]
    Definition Classes
    Any

Concrete Value Members

  1. final def !=(arg0: Any): Boolean
    Definition Classes
    Any
  2. final def ##(): Int
    Definition Classes
    Any
  3. final def ==(arg0: Any): Boolean
    Definition Classes
    Any
  4. val InterimState: InterimStore.type
    Definition Classes
    ProphecyDataFrame
  5. def YJJJ_to_YYYYJJJ(in_date: Column, ref_date: Column): Column

    Converts 1 digit julian year to 4 digits julian year.

    Converts 1 digit julian year to 4 digits julian year.

    in_date

    date in Julian in "YJJJ" format

    ref_date

    date in "yyyyMMdd" format

    returns

    a date in "YYYYJJJ"

    Definition Classes
    SparkFunctions
  6. def appendTrailer(pathInputData: String, pathInputTrailer: String, pathOutputConcatenated: String, configuration: Configuration): Unit

    Appends a trailer data to every single file in the data directory.

    Appends a trailer data to every single file in the data directory. A single trailer file in the pathOutputTrailer directory should correspond to a single data file in the pathOutputData directory.

    If a trailer for a given file does not exist, the file is moved as is to the output directory.

    pathInputData

    Input data files directory

    pathInputTrailer

    Input trailer files directory

    pathOutputConcatenated

    Output concatenated files directory

    configuration

    Hadoop configuration (preferably sparkSession.sparkContext.hadoopConfiguration)

    Definition Classes
    DataHelpers
  7. def arrayColumn(value: String, values: String*): Column

    Function to take variable number of values and create an array column out of it.

    Function to take variable number of values and create an array column out of it.

    value

    input value

    values

    variable number of input values.

    returns

    an array of column.

    Definition Classes
    UDFUtils
  8. val array_value: UserDefinedFunction

    UDF to find and return element in arr sequence at passed index.

    UDF to find and return element in arr sequence at passed index. If no element found then null is returned.

    Definition Classes
    UDFUtils
  9. final def asInstanceOf[T0]: T0
    Definition Classes
    Any
  10. val bigDecimalToPackedBytes: UserDefinedFunction
    Definition Classes
    SparkFunctions
  11. val call_rest_api: UserDefinedFunction

    Spark UDF that makes a single blocking rest API call to a given url.

    Spark UDF that makes a single blocking rest API call to a given url. The result of this udf is always produced, contains a proper error if it failed at any stage, and never interrupts the job execution (unless called with invalid signature).

    The default timeout can be configured through the spark.network.timeout Spark configuration option.

    Parameters:

    • method - any supported HTTP1.1 method type, e.g. POST, GET. Complete list: httpMethods.
    • url - valid url to which a request is going to be made
    • headers - an array of "key: value" headers that are past with the request
    • content - any content (by default, the supported rest api content type is application/json)

    Response - a struct with the following fields:

    • isSuccess - boolean, whether a successful response has been received
    • status - nullable integer, status code (e.g. 404, 200, etc)
    • headers - an array of name: value response headers (e.g. [Server: akka-http/10.1.10, Date: Tue, 07 Sep 2021 18:11:47 GMT])
    • content - nullable string, response back
    • error - nullable string, if the parameters passed are valid or the system failed to make a call, this field contains an error message
    Definition Classes
    RestAPIUtils
  12. val canonical_representation: UserDefinedFunction
    Definition Classes
    SparkFunctions
  13. def castDataType(sparkSession: SparkSession, df: DataFrame, column: Column, dataType: String, replaceColumn: String): DataFrame

    Function to add new typecasted column in input dataframe.

    Function to add new typecasted column in input dataframe. Newly added column is typecasted version of passed column. Typecast operation is supported for string, boolean, byte, short, int, long, float, double, decimal, date, timestamp

    sparkSession

    spark session

    df

    input dataframe

    column

    input column to be typecasted

    dataType

    datatype to cast column to.

    replaceColumn

    column name to be added in dataframe.

    returns

    new dataframe with new typecasted column.

    Definition Classes
    UDFUtils
  14. def concatenate(sources: Seq[String], destination: String, compressToGZip: Boolean = false): Unit

    Method to get data from multiple source paths and combine it into single destination path.

    Method to get data from multiple source paths and combine it into single destination path.

    sources

    multiple source paths from which to merge the data.

    destination

    destination path to combine all data to.

    compressToGZip

    flag to compress final output file into gzip format

    Definition Classes
    DataHelpers
  15. def convertInputBytesToStructType(input: Any, typeInfo: Seq[String], startByte: Int = 0): Row

    Method used for abinitio's reinterpret_as function to read necessary bytes from byteArray for input data and convert into struct format as per provided in typeInfo sequence.

    Method used for abinitio's reinterpret_as function to read necessary bytes from byteArray for input data and convert into struct format as per provided in typeInfo sequence.

    TypeInfo can have multiple entries, each could be either decimal or string type. Depending on the argument passed within decimal or string bytes are read from input byte array.

    If decimal or string argument has some integer then that many bytes are read from input byte array or if decimal or string has some string delimiter as its argument then from the current position bytes are read until string delimiter is found in input byte array.

    Definition Classes
    SparkFunctions
  16. def createDataFrameFromData(inputData: String, delimiter: String, columnName: String, columnType: String, sparkSession: SparkSession): DataFrame

    Method to read values from inputData and create dataframe with column name as columnName and column type as columnType for the values in inputData delimiter by delimiter.

    Method to read values from inputData and create dataframe with column name as columnName and column type as columnType for the values in inputData delimiter by delimiter.

    Definition Classes
    SparkFunctions
  17. def createLookup(name: String, df: DataFrame, spark: SparkSession, keyCols: List[String], rowCols: String*): UserDefinedFunction

    Function registers 4 different UDFs with spark registry.

    Function registers 4 different UDFs with spark registry. UDF for lookup_match, lookup_count, lookup_row and lookup functions are registered. This function stores the data of input dataframe in a broadcast variable, then uses this broadcast variable in different lookup functions.

    lookup : This function returns the first matching row for given input keys lookup_count : This function returns the count of all matching rows for given input keys. lookup_match : This function returns 0 if there is no matching row and 1 for some matching rows for given input keys. lookup_row : This function returns all the matching rows for given input keys.

    This function registers for upto 10 matching keys as input to these lookup functions.

    name

    UDF Name

    df

    input dataframe

    spark

    spark session

    keyCols

    columns to be used as keys in lookup functions.

    rowCols

    schema of entire row which will be stored for each matching key.

    returns

    registered UDF definitions for lookup functions. These UDF functions returns different results depending on the lookup function.

    Definition Classes
    UDFUtils
  18. def createRangeLookup(name: String, df: DataFrame, spark: SparkSession, minColumn: String, maxColumn: String, valueColumns: String*): UserDefinedFunction

    Method to create UDF which looks for passed input double in input dataframe.

    Method to create UDF which looks for passed input double in input dataframe. This function first loads the data of dataframe in broadcast variable and then defines a UDF which looks for input double value in the data stored in broadcast variable. If input double lies between passed col1 and col2 values then it adds corresponding row in the returned result. If value of input double doesn't lie between col1 and col2 then it simply returns null for current row in result.

    name

    created UDF name

    df

    input dataframe

    spark

    spark session

    minColumn

    column whose value to be considered as minimum in comparison.

    maxColumn

    column whose value to be considered as maximum in comparison.

    valueColumns

    remaining column names to be part of result.

    returns

    registers UDF which in turn returns rows corresponding to each row in dataframe on which range UDF is called.

    Definition Classes
    UDFUtils
  19. val cross_join_index_range: UserDefinedFunction
    Definition Classes
    SparkFunctions
  20. def date_add_months(inputDate: Column, months: Int): Column

    Returns the internal representation of a date resulting from adding (or subtracting) a number of months to the specified date.

    Returns the internal representation of a date resulting from adding (or subtracting) a number of months to the specified date.

    inputDate

    in yyyy-MM-dd format

    Definition Classes
    SparkFunctions
  21. def date_difference_days(laterDate: Column, earlierDate: Column): Column

    Computes number of days between two specified dates in "yyyyMMdd" format

    Computes number of days between two specified dates in "yyyyMMdd" format

    laterDate

    input date

    earlierDate

    input date

    returns

    number of days between laterDate and earlierDate or null if either one is null

    Definition Classes
    SparkFunctions
  22. val date_month_end: UserDefinedFunction
    Definition Classes
    SparkFunctions
  23. val datetime_add: UserDefinedFunction
    Definition Classes
    SparkFunctions
  24. def datetime_add_months(input: Column, months: Int): Column

    Returns the internal representation of a timestamp resulting from adding (or subtracting) a number of months to the specified timestamp.

    Returns the internal representation of a timestamp resulting from adding (or subtracting) a number of months to the specified timestamp.

    input

    timestamp in yyyy-MM-dd HH:mm:ss.SSSS format

    Definition Classes
    SparkFunctions
  25. val datetime_difference: UserDefinedFunction
    Definition Classes
    SparkFunctions
  26. def datetime_difference_hours(end: Column, start: Column): Column

    Returns the number of hours between two specified dates in standard format yyyy-MM-dd HH:mm:ss.SSSS.

    Returns the number of hours between two specified dates in standard format yyyy-MM-dd HH:mm:ss.SSSS.

    Definition Classes
    SparkFunctions
  27. def datetime_from_unixtime(seconds: Column): Column
    Definition Classes
    SparkFunctions
  28. def decimal_lpad(input: Column, len: Int, char_to_pad_with: String = "0", decimal_point_char: String = "."): Column

    Method uses a java regex to identify decimal numbers from input string.

    Method uses a java regex to identify decimal numbers from input string. This decimal number could be of 3 types 1. Simple integral number. e.g. 013334848. This part is identified by regex. 2. decimal number with explicit decimal point. e.g. 123456.90. This part is identified by combination of [0-9]+(\$$decimal_point_char)[0-9]+ and (0\$$decimal_point_char)[0-9]+ regex

    After extracting decimal number this code checks if length of decimal number is more than len parameter or not. If length is more than len parameter then it simply returns this extracted decimal number. Otherwise it first left pad decimal number with char_to_pad_with to make its length equal to len parameter and then adjusts minus sign (-) to left most part of decimal number.

    input

    input string.

    len

    length of characters.

    char_to_pad_with

    character to left pad with. default value is "0"

    decimal_point_char

    A string that specifies the character that represents the decimal point.

    returns

    a decimal string of the specified length or longer, left-padded with a specified character as needed and trimmed of leading zeros.

    Definition Classes
    SparkFunctions
  29. def decimal_lrepad(input: Column, len: Int, char_to_pad_with: String = "0", decimal_point_char: String = "."): Column

    Method uses a java regex to identify decimal numbers from input string.

    Method uses a java regex to identify decimal numbers from input string. This decimal number could be of 3 types 1. Simple integral number. e.g. 013334848. This part is identified by combination of [1-9][0-9]*[0-9] and [1-9]+ regex 2. decimal number with explicit decimal point. e.g. 123456.90. This part is identified by combination of [1-9][0-9]*(\\\$$decimal_point_char)[0-9]+ and (0\\\$$decimal_point_char)[0-9]*[0-9] regex

    After extracting decimal number this code checks if length of decimal number is more than len parameter or not. If length is more than len parameter then it simply returns this extracted decimal number. Otherwise it first left pad decimal number with char_to_pad_with to make its length equal to len parameter and then adjusts minus sign (-) to left most part of decimal number.

    input

    input string.

    len

    length of characters.

    char_to_pad_with

    character to left pad with. default value is "0"

    decimal_point_char

    A string that specifies the character that represents the decimal point.

    returns

    a decimal string of the specified length or longer, left-padded with a specified character as needed and trimmed of leading zeros.

    Definition Classes
    SparkFunctions
  30. def decimal_round(input: Column, places: Int): Column
    Definition Classes
    SparkFunctions
  31. def decimal_round_down(input: Column, right_digits: Int): Column

    Function returns a value which is rounded down to right_digits number of digits to the right of decimal point.

    Function returns a value which is rounded down to right_digits number of digits to the right of decimal point.

    Definition Classes
    SparkFunctions
  32. def decimal_round_up(input: Column, places: Int): Column

    Returns a number rounded up to a specified number of places to the right of the decimal point.

    Returns a number rounded up to a specified number of places to the right of the decimal point.

    Definition Classes
    SparkFunctions
  33. def decimal_strip(input: Column, decimal_point_char: String = "."): Column

    Function uses a java regex to identify decimal numbers from input string.

    Function uses a java regex to identify decimal numbers from input string. This decimal number could be of 3 types 1. Simple integral number. e.g. 013334848. This part is identified by combination of [1-9][0-9 ]*[0-9] and [1-9]+ regex 2. decimal number with explicit decimal point. e.g. 123456.90. This part is identified by combination of [1-9][0-9]*(\$$decimal_point_char)[0-9 ]+ and (0\$$decimal_point_char)[0-9 ]*[0-9] regex

    After extracting decimal number this code looks for minus sign before extracted number in input and appends it with decimal number if found minus sign.

    In the end it replaces all whitespaces with empty string in the final resultant decimal number.

    input

    input string

    decimal_point_char

    A string that specifies the character that represents the decimal point.

    returns

    a decimal from a string that has been trimmed of leading zeros and non-numeric characters.

    Definition Classes
    SparkFunctions
  34. def decimal_truncate(input: Column, number_of_places: Column): Column
    Definition Classes
    SparkFunctions
  35. val decodeBytes: UserDefinedFunction
    Definition Classes
    SparkFunctions
  36. val decodeString: UserDefinedFunction
    Definition Classes
    SparkFunctions
  37. val decode_datetime: UserDefinedFunction

    UDF to get record of type decode_datetime_type.

    UDF to get record of type decode_datetime_type. This record will have all its fields populated with corresponding entries in input date/timestamp.

    Returned record will have following schema.

    integer(8) year; integer(8) month; integer(8) day; integer(8) hour; integer(8) minute; integer(8) second; integer(8) microsecond;

    Note: Supported Input time is in yyyy-MM-dd HH:mm:ss.SSSSSS or yyyy-MM-dd HH:mm:ss or yyyy-MM-dd formats only. Additional handling is done to support timestamp retrieved from now() function call.

    Definition Classes
    SparkFunctions
  38. val decode_datetime_as_local: UserDefinedFunction
    Definition Classes
    SparkFunctions
  39. def directory_listing(path: String, filePrefix: String): Column
    Definition Classes
    SparkFunctions
  40. def dropColumns(sparkSession: SparkSession, df: DataFrame, columns: Column*): DataFrame

    Function to drop passed columns from input dataframe.

    Function to drop passed columns from input dataframe.

    sparkSession

    spark session

    df

    input dataframe.

    columns

    list of columns to be dropped from dataframe.

    returns

    new dataframe with dropped columns.

    Definition Classes
    UDFUtils
  41. val encodeBytes: UserDefinedFunction
    Definition Classes
    SparkFunctions
  42. val encodeString: UserDefinedFunction
    Definition Classes
    SparkFunctions
  43. val encode_date: UserDefinedFunction

    integer values specifying days relative to January 1, 1900.

    integer values specifying days relative to January 1, 1900. This function returns the internal representation of a date given the year, month, and date. encode_date returns the internal representation of the date specified by the year 1998, the month 5, and the day 18:encode_date(1998, 5, 18) = 35931

    Definition Classes
    SparkFunctions
  44. def ends_with(input: Column, suffix: String): Column

    Returns true if string columns ends with given suffix

    Returns true if string columns ends with given suffix

    Definition Classes
    SparkFunctions
  45. def equals(arg0: Any): Boolean
    Definition Classes
    Any
  46. val eval: UserDefinedFunction

    Method to return the result of evaluating a string expression in the context of a specified input column.

    Method to return the result of evaluating a string expression in the context of a specified input column. Here input column could be struct type record, simple column, array type etc. Here expr could be reference to nested column inside input column or any expression which requires values from input column for its evaulation.

    Note: Current implementation only supports scenerio where input column is of struct type and expr is simply dot separated column reference to input struct.

    Definition Classes
    SparkFunctions
  47. def executeNonSelectSQLQueries(sqlList: Seq[String], dbConnection: Connection): Unit
    Definition Classes
    DataHelpers
  48. val file_information: UserDefinedFunction

    UDF to get file information for passed input file path.

    UDF to get file information for passed input file path.

    Definition Classes
    SparkFunctions
  49. def findFirstElement(input: Column, default: Column = lit(null)): Column
    Definition Classes
    SparkFunctions
  50. def findFirstNonBlankElement(input: Column, default: Column): Column
    Definition Classes
    SparkFunctions
  51. def findLastElement(input: Column, default: Column = lit(null)): Column
    Definition Classes
    SparkFunctions
  52. def first_defined(expr1: Column, expr2: Column): Column

    Method to identify and return first non null expression.

    Method to identify and return first non null expression.

    Definition Classes
    SparkFunctions
  53. val first_defined_for_double_Udf: UserDefinedFunction
    Definition Classes
    SparkFunctions
  54. def flattenStructSchema(schema: StructType, prefix: String = null): Array[Column]
    Definition Classes
    SparkFunctions
  55. val force_error: UserDefinedFunction
    Definition Classes
    SparkFunctions
  56. def from_sv(input: Column, separator: String, schema: StructType): Column
    Definition Classes
    SparkFunctions
  57. def from_xml(content: Column, schema: StructType): Column
    Definition Classes
    SparkFunctions
  58. def ftpTo(remoteHost: String, userName: String, password: String, sourceFile: String, destFile: String, retryFailures: Boolean, retryCount: Int, retryPauseSecs: Int, mode: String, psCmd: String): (Boolean, Boolean, String, String)
    Definition Classes
    DataHelpers
  59. def generateDataFrameWithSequenceColumn(start: Int, end: Int, columnName: String, sparkSession: SparkSession): DataFrame

    Method to create dataframe with single column containing increasing sequence id from start to end.

    Method to create dataframe with single column containing increasing sequence id from start to end.

    Definition Classes
    SparkFunctions
  60. def generate_sequence(start: Int, end: Int, step: Int = 1): Column

    Function to create sequence of array between two passed numbers

    Function to create sequence of array between two passed numbers

    start

    starting point of generated sequence

    end

    terminating point of generated sequence.

    returns

    column containing sequence of integers.

    Definition Classes
    SparkFunctions
  61. val generate_sequence: UserDefinedFunction

    UDF to generate column with sequence of integers between two passed start and end columns.

    UDF to generate column with sequence of integers between two passed start and end columns.

    Definition Classes
    SparkFunctions
  62. val getByteFromByteArray: UserDefinedFunction

    UDF to get last Byte from ByteArray of input data.

    UDF to get last Byte from ByteArray of input data.

    Definition Classes
    SparkFunctions
  63. def getColumnInSecondArrayByFirstNonBlankPositionInFirstArray(nonBlankEntryExpr: Column, firstArray: Column, secondArray: Column): Column
    Definition Classes
    SparkFunctions
  64. def getContentAsStream(content: String): StringAsStream
    Definition Classes
    SparkFunctions
  65. def getEmptyLogDataFrame(sparkSession: SparkSession): DataFrame

    Method to get empty dataframe with below abinitio log schema.

    Method to get empty dataframe with below abinitio log schema.

    record string("|") node, timestamp, component, subcomponent, event_type; string("|\n") event_text; end

    Definition Classes
    DataHelpers
  66. def getFebruaryDay(year: Column): Column

    Computes number of days in February month in a given year

    Computes number of days in February month in a given year

    year

    year whose number of days in February needs to be calculated

    returns

    number of days

    Definition Classes
    SparkFunctions
  67. def getFieldFromStructByPosition(column: Column, position: Int): Column

    Method to get field at specific position from struct column

    Method to get field at specific position from struct column

    Definition Classes
    SparkFunctions
  68. val getIntFromByteArray: UserDefinedFunction

    UDF to get integer comprising of last 4 Bytes from ByteArray of input data.

    UDF to get integer comprising of last 4 Bytes from ByteArray of input data.

    Definition Classes
    SparkFunctions
  69. val getLongArrayFromByteArray: UserDefinedFunction

    UDF to get long comprising of last 8 Bytes from ByteArray of input data.

    UDF to get long comprising of last 8 Bytes from ByteArray of input data.

    Definition Classes
    SparkFunctions
  70. val getLongFromByteArray: UserDefinedFunction

    UDF to get long comprising of last 8 Bytes from ByteArray of input data.

    UDF to get long comprising of last 8 Bytes from ByteArray of input data.

    Definition Classes
    SparkFunctions
  71. def getMTimeDataframe(filepath: String, format: String, spark: SparkSession): DataFrame
    Definition Classes
    SparkFunctions
  72. val getShortFromByteArray: UserDefinedFunction

    UDF to get short comprising of last 2 Bytes from ByteArray of input data.

    UDF to get short comprising of last 2 Bytes from ByteArray of input data.

    Definition Classes
    SparkFunctions
  73. def hashCode(): Int
    Definition Classes
    Any
  74. val hash_MD5: UserDefinedFunction
    Definition Classes
    SparkFunctions
  75. val instr_udf: UserDefinedFunction
    Definition Classes
    SparkFunctions
  76. final def isInstanceOf[T0]: Boolean
    Definition Classes
    Any
  77. def isNullOrEmpty(input: Column): Column

    Method to check if current column is null or has empty value.

    Method to check if current column is null or has empty value.

    Definition Classes
    SparkFunctions
  78. def is_ascii(input: Column): Column

    Checks if a string is ascii

    Checks if a string is ascii

    input

    column to be checked

    returns

    true if the input string is ascii otherwise false

    Definition Classes
    SparkFunctions
  79. def is_blank(input: Column): Column

    Method to identify if input string is a blank string or not.

    Method to identify if input string is a blank string or not.

    input

    input string.

    returns

    return 1 if given string contains all blank character or is a zero length string, otherwise it returns 0

    Definition Classes
    SparkFunctions
  80. val is_bzero: UserDefinedFunction

    Tests whether an object is composed of all binary zero bytes.

    Tests whether an object is composed of all binary zero bytes. This function returns: 1. 1 if obj contains only binary zero bytes or is a zero-length string 2. 0 if obj contains any non-zero bytes 3. NULL if obj is NULL

    Definition Classes
    SparkFunctions
  81. def is_numeric_ascii(input: Column): Column

    Checks if an input string contains only ascii code and numbers

    Checks if an input string contains only ascii code and numbers

    input

    string to be checked

    returns

    true if input string contains only ascii code and numbers or null if input is null

    Definition Classes
    SparkFunctions
  82. def is_valid(input: Column, isNullable: Boolean, formatInfo: Option[Any], len: Option[Seq[Int]]): Column

    Method to identify if passed input column is a valid expression after typecasting to passed dataType.

    Method to identify if passed input column is a valid expression after typecasting to passed dataType. Also while typecasting if len is present then this function also makes sure the max length of input column after typecasting operation is not greater than len.

    input

    input column expression to be identified if is valid.

    formatInfo

    datatype to which input column expression must be typecasted. If datatype is a string then it is treated as timestamp format. If it is a list of string then it is treated as having current timestamp format and and new timestamp format to which input column needs to be typecasted.

    len

    max length of input column after typecasting it to dataType.

    returns

    0 if input column is not valid after typecasting or 1 if it is valid.

    Definition Classes
    SparkFunctions
  83. def is_valid(input: Column, isNullable: Boolean, formatInfo: Option[Any]): Column
    Definition Classes
    SparkFunctions
  84. def is_valid(input: Column, formatInfo: Option[Any], len: Option[Seq[Int]]): Column
    Definition Classes
    SparkFunctions
  85. def is_valid(input: Column, formatInfo: Option[Any]): Column
    Definition Classes
    SparkFunctions
  86. def is_valid(input: Column, isNullable: Boolean): Column
    Definition Classes
    SparkFunctions
  87. def is_valid(input: Column): Column
    Definition Classes
    SparkFunctions
  88. def is_valid_date(dateFormat: String, inDate: Column): Column

    Validates date against a input format

    Validates date against a input format

    dateFormat

    A pattern such as yyyy-MM-dd or yyyy-MM-dd HH:mm:ss.SSSS or dd.MM.yyyy

    inDate

    Input date to be validated

    returns

    true if the input date is valid otherwise false

    Definition Classes
    SparkFunctions
  89. def loadBinaryFileAsBinaryDataFrame(filePath: String, lineDelimiter: String = "\n", minPartition: Int = 1, rowName: String = "line", spark: SparkSession): DataFrame
    Definition Classes
    DataHelpers
  90. def loadBinaryFileAsStringDataFrame(filePath: String, lineDelimiter: String = "\n", charSetEncoding: String = "Cp1047", minPartition: Int = 1, rowName: String = "line", spark: SparkSession): DataFrame
    Definition Classes
    DataHelpers
  91. def loadFixedWindowBinaryFileAsDataFrame(filePath: String, lineLength: Int, minPartition: Int = 1, rowName: String = "line", spark: SparkSession): DataFrame
    Definition Classes
    DataHelpers
  92. lazy val logger: Logger
    Attributes
    protected
    Definition Classes
    LazyLogging
    Annotations
    @transient()
  93. def lookup(lookupName: String, cols: Column*): Column

    By default returns only the first matching record

    By default returns only the first matching record

    Definition Classes
    UDFUtils
  94. def lookup_count(lookupName: String, cols: Column*): Column
    Definition Classes
    UDFUtils
  95. def lookup_last(lookupName: String, cols: Column*): Column

    Returns the last matching record

    Returns the last matching record

    Definition Classes
    UDFUtils
  96. def lookup_match(lookupName: String, cols: Column*): Column

    returns

    Boolean Column

    Definition Classes
    UDFUtils
  97. def lookup_nth(lookupName: String, cols: Column*): Column
    Definition Classes
    UDFUtils
  98. def lookup_range(lookupName: String, input: Column): Column
    Definition Classes
    UDFUtils
  99. def lookup_row(lookupName: String, cols: Column*): Column
    Definition Classes
    UDFUtils
  100. def lookup_row_reverse(lookupName: String, cols: Column*): Column
    Definition Classes
    UDFUtils
  101. val make_byte_flags: UserDefinedFunction

    UDF to return a flag for each character if it is present or not in input String.

    UDF to return a flag for each character if it is present or not in input String.

    Definition Classes
    SparkFunctions
  102. def make_constant_vector(size: Int, seedVal: Int): Array[Int]

    Method to create array of size "size" containing seedVal as each entry

    Method to create array of size "size" containing seedVal as each entry

    Definition Classes
    SparkFunctions
  103. def make_constant_vector(size: Int, seedVal: Column): Column

    Method to create array of size "size" containing seedVal as each entry

    Method to create array of size "size" containing seedVal as each entry

    Definition Classes
    SparkFunctions
  104. def measure[T](fn: ⇒ T)(caller: String = findCaller()): T
    Definition Classes
    UDFUtils
  105. val multifile_information: UserDefinedFunction

    UDF to get multifile information for passed input file path.

    UDF to get multifile information for passed input file path.

    Definition Classes
    SparkFunctions
  106. val murmur: UserDefinedFunction

    UDF for murmur hash generation for any column type

    UDF for murmur hash generation for any column type

    Definition Classes
    SparkFunctions
  107. def now(): Column

    Method to get current timestamp.

    Method to get current timestamp.

    returns

    current timestamp in YYYYMMddHHmmssSSSSSS format.

    Definition Classes
    SparkFunctions
  108. def numberOfPartitions(in: DataFrame): Column
    Definition Classes
    SparkFunctions
  109. val number_grouping: UserDefinedFunction

    udf to group input decimal into multiple groups separated by separator

    udf to group input decimal into multiple groups separated by separator

    Definition Classes
    SparkFunctions
  110. val packedBytesStringToDecimal: UserDefinedFunction
    Definition Classes
    SparkFunctions
  111. val packedBytesToDecimal: UserDefinedFunction
    Definition Classes
    SparkFunctions
  112. val re_get_match: UserDefinedFunction

    Returns the first string in a target string that matches a regular expression.

    Returns the first string in a target string that matches a regular expression.

    Definition Classes
    SparkFunctions
  113. val re_get_match_with_index: UserDefinedFunction
    Definition Classes
    SparkFunctions
  114. val re_index: UserDefinedFunction

    UDF wrapper over re_index function.

    UDF wrapper over re_index function.

    Definition Classes
    SparkFunctions
  115. val re_index_with_offset: UserDefinedFunction

    Returns the first string in a target string that matches a regular expression.

    Returns the first string in a target string that matches a regular expression.

    Definition Classes
    SparkFunctions
  116. def re_replace(target: Column, pattern: String, replacement: String, offset: Int = 0): Column

    Replaces all substrings in a target string that match a specified regular expression.

    Replaces all substrings in a target string that match a specified regular expression.

    target

    A string that the function searches for a substring that matches pattern_expr.

    pattern

    regular expression

    replacement

    replacement string

    offset

    Number of characters, from the beginning of str, to skip before searching.

    returns

    a replaced string in which all substrings, which matches a specified regular expression, are replaced.

    Definition Classes
    SparkFunctions
  117. def re_replace_first(target: Column, pattern: String, replacement: String, offset: Column = lit(0)): Column

    Replaces only the first regex matching occurrence in the target string.

    Replaces only the first regex matching occurrence in the target string.

    target

    A string that the function searches for a substring that matches pattern_expr.

    pattern

    regular expression

    replacement

    replacement string

    returns

    a replaced string in which first substring, which matches a specified regular expression, is replaced.

    Definition Classes
    SparkFunctions
  118. val re_split_no_empty: UserDefinedFunction

    UDF to split input string via pattern string and remove all empty subtrings.

    UDF to split input string via pattern string and remove all empty subtrings.

    Definition Classes
    SparkFunctions
  119. val readBytesIntoInteger: UserDefinedFunction
    Definition Classes
    SparkFunctions
  120. val readBytesIntoLong: UserDefinedFunction
    Definition Classes
    SparkFunctions
  121. val readBytesStringIntoInteger: UserDefinedFunction
    Definition Classes
    SparkFunctions
  122. val readBytesStringIntoLong: UserDefinedFunction
    Definition Classes
    SparkFunctions
  123. def readHiveTable(spark: SparkSession, database: String, table: String, partition: String = ""): DataFrame

    Method to read data from hive table.

    Method to read data from hive table.

    spark

    spark session

    database

    hive database

    table

    hive table.

    partition

    hive table partition to read data specifically from if provided.

    returns

    dataframe with data read from Hive Table.

    Definition Classes
    DataHelpers
  124. def readHiveTableInChunks(spark: SparkSession, database: String, table: String, partitionKey: String, partitionValue: String): DataFrame

    Reads a full hive table partition, by reading every subpartition separately and performing a union on all the final DataFrames

    Reads a full hive table partition, by reading every subpartition separately and performing a union on all the final DataFrames

    This function is meant to temporarily solve the problem with Hive metastore crashing when querying too many partitions at the same time.

    spark

    spark session

    database

    hive database name

    table

    hive table name

    partitionKey

    top-level partition's key

    partitionValue

    top-level partition's value

    returns

    A complete DataFrame with the selected hive table partition

    Definition Classes
    DataHelpers
  125. val read_file: UserDefinedFunction
    Definition Classes
    SparkFunctions
  126. val record_info: UserDefinedFunction
    Definition Classes
    SparkFunctions
  127. val record_info_with_includes: UserDefinedFunction
    Definition Classes
    SparkFunctions
  128. def registerAllUDFs(spark: SparkSession): Unit
    Definition Classes
    SparkFunctions
  129. def registerProphecyUdfs(spark: SparkSession): Unit
    Definition Classes
    UDFUtils
  130. def register_output_schema(portName: String, schema: StructType): Unit
    Definition Classes
    Component
  131. def remove_non_digit(input: Column): Column

    Method removes any non-digit characters from the specified string column.

    Method removes any non-digit characters from the specified string column.

    input

    input String Column

    returns

    Cleaned string column or null

    Definition Classes
    SparkFunctions
  132. def replaceBlankColumnWithNull(input: Column): Column

    Method to replace String Columns with Empty value to Null.

    Method to replace String Columns with Empty value to Null.

    Definition Classes
    SparkFunctions
  133. def replaceString(sparkSession: SparkSession, df: DataFrame, outputCol: String, inputCol: String, replaceWith: String, value: String, values: String*): DataFrame

    Function to add new column in passed dataframe.

    Function to add new column in passed dataframe. Newly added column value is decided by the presence of value corresponding to inputCol in array comprised of value and values. If inputCol is found then value of replaceWith is added in new column otherwise inputCol value is added.

    sparkSession

    spark session.

    df

    input dataframe.

    outputCol

    name of new column to be added.

    inputCol

    column name whose value is searched.

    replaceWith

    value with which to replace searched value if found.

    value

    element to be combined in array column

    values

    all values to be combined in array column for searching purpose.

    returns

    dataframe with new column with column name outputCol

    Definition Classes
    UDFUtils
  134. def replaceStringNull(sparkSession: SparkSession, df: DataFrame, outputCol: String, inputCol: String, replaceWith: String, value: String, values: String*): DataFrame

    Function to add new column in passed dataframe.

    Function to add new column in passed dataframe. Newly added column value is decided by the presence of value corresponding to inputCol in array comprised of value and values and null. If inputCol is found then value of replaceWith is added in new column otherwise inputCol value is added.

    sparkSession

    spark session.

    df

    input dataframe.

    outputCol

    name of new column to be added.

    inputCol

    column name whose value is searched.

    replaceWith

    value with which to replace searched value if found.

    value

    element to be combined in array column

    values

    all values to be combined in array column for searching purpose.

    returns

    dataframe with new column with column name outputCol

    Definition Classes
    UDFUtils
  135. def replaceStringWithNull(sparkSession: SparkSession, df: DataFrame, outputCol: String, inputCol: String, value: String, values: String*): DataFrame

    Function to add new column in passed dataframe.

    Function to add new column in passed dataframe. Newly added column value is decided by the presence of value corresponding to inputCol in array comprised of value and values and null. If inputCol is found then value of null is added in new column otherwise inputCol value is added.

    sparkSession

    spark session.

    df

    input dataframe.

    outputCol

    name of new Column to be added.

    inputCol

    column name whose value is searched.

    value

    element to be combined in array column.

    values

    all values to be combined in array column for searching purpose.

    returns

    dataframe with new column with column name outputCol

    Definition Classes
    UDFUtils
  136. def replace_null_with_blank(input: Column): Column
    Definition Classes
    SparkFunctions
  137. val replace_string: UserDefinedFunction

    UDF to find str in input sequence toBeReplaced and return replace if found.

    UDF to find str in input sequence toBeReplaced and return replace if found. Otherwise str is returned.

    Definition Classes
    UDFUtils
  138. val replace_string_with_null: UserDefinedFunction

    UDF to find str in input sequence toBeReplaced and return null if found.

    UDF to find str in input sequence toBeReplaced and return null if found. Otherwise str is returned.

    Definition Classes
    UDFUtils
  139. def scanf_double(format: Column, value: Column): Column
    Definition Classes
    SparkFunctions
  140. def scanf_long(format: Column, value: Column): Column
    Definition Classes
    SparkFunctions
  141. def schemaRowCompareResult(row1: StructType, row2: StructType): Column
    Definition Classes
    SparkFunctions
  142. def sign_explicit(c: Column): Column

    Adds an explicit sign to the number.

    Adds an explicit sign to the number. E.g. 2 -> +2; -004 -> -004; 0 -> +0

    Definition Classes
    SparkFunctions
  143. val sign_explicit_Udf: UserDefinedFunction
    Definition Classes
    SparkFunctions
  144. def sign_reserved(c: Column): Column
    Definition Classes
    SparkFunctions
  145. val sign_reserved_Udf: UserDefinedFunction
    Definition Classes
    SparkFunctions
  146. def splitIntoMultipleColumns(sparkSession: SparkSession, df: DataFrame, colName: String, pattern: String, prefix: String = null): DataFrame

    Function to split column with colName in input dataframe using split pattern into multiple columns.

    Function to split column with colName in input dataframe using split pattern into multiple columns. If prefix name is provided each new generated column is prefixed with prefix followed by column number, otherwise original column name is used.

    sparkSession

    spark session.

    df

    input dataframe.

    colName

    column in dataframe which needs to be split into multiple columns.

    pattern

    regex with which column in input dataframe will be split into multiple columns.

    prefix

    column prefix to be used with all newly generated columns.

    returns

    new dataframe with new columns where new column values are generated after splitting original column colName.

    Definition Classes
    UDFUtils
  147. val splitIntoMultipleColumnsUdf: UserDefinedFunction

    UDF to break input string into multiple string via delimiter.

    UDF to break input string into multiple string via delimiter. Number of strings after split are adjusted as per passed width parameter. If number of strings are less then empty strings are added otherwise in case of more number of strings, first width number of entries are picked and remaining are discarded.

    Definition Classes
    SparkFunctions
  148. def starts_with(input: Column, prefix: String): Column

    Returns true if string columns starts with given prefix

    Returns true if string columns starts with given prefix

    Definition Classes
    SparkFunctions
  149. def string_char(inputStr: Column, index: Int): Column

    Method to return character code of character at index position in inputStr string.

    Method to return character code of character at index position in inputStr string.

    inputStr

    input string

    index

    location of character to get code.

    returns

    integer column.

    Definition Classes
    SparkFunctions
  150. val string_cleanse: UserDefinedFunction

    This implementation is incorrect.

    This implementation is incorrect.

    Definition Classes
    SparkFunctions
  151. def string_compare(input1: Column, input2: Column): Column
    Definition Classes
    SparkFunctions
  152. val string_concat_in_loop: UserDefinedFunction
    Definition Classes
    SparkFunctions
  153. val string_convert_explicit: UserDefinedFunction

    Converts a string from one character set to another, replacing inconvertible characters with a specified string.

    Converts a string from one character set to another, replacing inconvertible characters with a specified string.

    Definition Classes
    SparkFunctions
  154. val string_filter: UserDefinedFunction

    Method which returns string of characters present in both of the strings in the same order as appearing in first string

    Method which returns string of characters present in both of the strings in the same order as appearing in first string

    Definition Classes
    SparkFunctions
  155. val string_filter_out: UserDefinedFunction

    Compares two input strings, then returns characters that appear in one string but not in the other.

    Compares two input strings, then returns characters that appear in one string but not in the other.

    Definition Classes
    SparkFunctions
  156. val string_index: UserDefinedFunction

    UDF to find index of seekStr in inputStr.

    UDF to find index of seekStr in inputStr. Returned index will be 1 based index.

    Definition Classes
    SparkFunctions
  157. val string_index_with_offset: UserDefinedFunction

    UDF to find index of seekStr in inputStr from offset index onwards.

    UDF to find index of seekStr in inputStr from offset index onwards. Returned string position is 1 based position.

    Definition Classes
    SparkFunctions
  158. def string_is_alphabetic(input: Column): Column

    Method which returns true if input string contains all alphabetic characters, or false otherwise.

    Method which returns true if input string contains all alphabetic characters, or false otherwise.

    Definition Classes
    SparkFunctions
  159. def string_is_numeric(input: Column): Column

    Method which returns true if input string contains all numeric characters, or false otherwise.

    Method which returns true if input string contains all numeric characters, or false otherwise.

    Definition Classes
    SparkFunctions
  160. def string_join(column: Column, delimiter: String): Column

    Concatenates the elements of column using the delimiter.

    Concatenates the elements of column using the delimiter.

    Definition Classes
    SparkFunctions
  161. def string_length(input: Column): Column
    Definition Classes
    SparkFunctions
  162. val string_like: UserDefinedFunction

    Method to test whether a string matches a specified pattern.

    Method to test whether a string matches a specified pattern. This function returns 1 if the input string matches a specified pattern, and 0 if the string does not match the pattern.

    In abinitio version % character in pattern means to match zero or more characters and _ character means matches a single character.

    Definition Classes
    SparkFunctions
  163. def string_lpad(input: Column, len: Int, pad_char: String = " "): Column

    Left-pad the input string column with pad_char to a length of len.

    Left-pad the input string column with pad_char to a length of len. If length of input column is more than len then returns input column unmodified.

    Definition Classes
    SparkFunctions
  164. def string_lrepad(input: Column, len: Int, char_to_pad_with: String = " "): Column

    function trims the string and then pad the string with given character upto given length.

    function trims the string and then pad the string with given character upto given length. if the length of trimmed string is equal to or greater than given length than it return input string

    input

    input string

    len

    length in number of characters.

    char_to_pad_with

    A character used to pad input string to length len.

    returns

    string of a specified length, trimmed of leading and trailing blanks and left-padded with a given character.

    Definition Classes
    SparkFunctions
  165. def string_pad(input: Column, len: Int, char_to_pad_with: String = " "): Column

    function pads input on the right with the character char_to_pad_with to make the string length len.

    function pads input on the right with the character char_to_pad_with to make the string length len. If str is already len or more characters long, the function returns input unmodified.

    Definition Classes
    SparkFunctions
  166. val string_pad: UserDefinedFunction
    Definition Classes
    SparkFunctions
  167. val string_pad_with_char: UserDefinedFunction
    Definition Classes
    SparkFunctions
  168. def string_prefix(input: Column, length: Column): Column
    Definition Classes
    SparkFunctions
  169. def string_repad(input: Column, len: Int, char_to_pad_with: String = " "): Column

    function trims the string and then pad the string on right side with given character upto given length.

    function trims the string and then pad the string on right side with given character upto given length. if the length of trimmed string is equal to or greater than given length than it return input string

    input

    input string

    len

    length in number of characters.

    char_to_pad_with

    A character used to pad input string to length len.

    returns

    string of a specified length, trimmed of leading and trailing blanks and left-padded with a given character.

    Definition Classes
    SparkFunctions
  170. def string_replace(input: Column, seekStr: Column, newStr: Column, offset: Column = lit(0)): Column

    Function to replace occurrence of seekStr with newStr string in input string after offset characters from first character.

    Function to replace occurrence of seekStr with newStr string in input string after offset characters from first character.

    input

    input string on which to perform replace operation.

    seekStr

    string to be replaced in input string.

    newStr

    string to be used instead of seekStr in input string.

    offset

    number of characters to skip from begining in input string before performing string_replace operation.

    returns

    modified string where seekStr is replaced with newStr in input string.

    Definition Classes
    SparkFunctions
  171. val string_replace_first: UserDefinedFunction
    Definition Classes
    SparkFunctions
  172. val string_replace_first_in_loop: UserDefinedFunction
    Definition Classes
    SparkFunctions
  173. val string_replace_in_loop: UserDefinedFunction
    Definition Classes
    SparkFunctions
  174. val string_representation: UserDefinedFunction
    Definition Classes
    SparkFunctions
  175. val string_rindex: UserDefinedFunction

    Returns the index of the first character of the last occurrence of a seek string within another input string.

    Returns the index of the first character of the last occurrence of a seek string within another input string. Returned index is 1 based.

    Definition Classes
    SparkFunctions
  176. val string_rindex_with_offset: UserDefinedFunction

    UDF to find index of seekStr in inputStr from end of inputStr skipping offset number of characters from end.

    UDF to find index of seekStr in inputStr from end of inputStr skipping offset number of characters from end. Offset index is number of characters, from the end of str, to skip before searching. Returned string position is 1 based position.

    Definition Classes
    SparkFunctions
  177. val string_split: UserDefinedFunction

    UDF to split input string via delimiter string.

    UDF to split input string via delimiter string.

    Definition Classes
    SparkFunctions
  178. val string_split_no_empty: UserDefinedFunction

    UDF to split input string via delimiter string and remove all empty subtrings.

    UDF to split input string via delimiter string and remove all empty subtrings.

    Definition Classes
    SparkFunctions
  179. def string_substring(input: Column, start_position: Column, length: Column): Column

    Method to find substring of input string.

    Method to find substring of input string.

    input

    string on which to find substring.

    start_position

    1 based starting position to find substring from.

    length

    total length of substring to be found.

    returns

    substring of input string

    Definition Classes
    SparkFunctions
  180. def string_suffix(input: Column, len: Int): Column
    Definition Classes
    SparkFunctions
  181. val take_last_nth: UserDefinedFunction

    UDF to return nth element from last in passed array of elements.

    UDF to return nth element from last in passed array of elements. In case input sequence has less number of elements than n then first element is returned.

    Definition Classes
    UDFUtils
  182. val take_nth: UserDefinedFunction

    UDF to take Nth element from beginning.

    UDF to take Nth element from beginning. In case input sequence has less element than N then exception is thrown.

    Definition Classes
    UDFUtils
  183. val test_characters_all: UserDefinedFunction

    UDF to identify the number of characters in inputStr which are present in charFlag

    UDF to identify the number of characters in inputStr which are present in charFlag

    Definition Classes
    SparkFunctions
  184. def timezone_to_utc(timezone: String, time: Column): Column

    Method to convert

    Method to convert

    Definition Classes
    SparkFunctions
  185. def toString(): String
    Definition Classes
    Any
  186. def today(): Column

    Method to return integer value representing number of days to today from “1-1-1990”.

    Method to return integer value representing number of days to today from “1-1-1990”.

    returns

    integer value

    Definition Classes
    SparkFunctions
  187. val translate_bytes: UserDefinedFunction

    UDF to return a string in the native character set made up of bytes from the given map.

    UDF to return a string in the native character set made up of bytes from the given map. Each byte of the result is the value of map indexed by the character code of the corresponding byte of the input string str. The function returns NULL if any argument is NULL.

    Definition Classes
    SparkFunctions
  188. val truncateMicroSeconds: UserDefinedFunction

    UDF to truncate microseconds part of timestamp.

    UDF to truncate microseconds part of timestamp. This is needed as abinitio and spark has some incompatibility in microseconds part of timestamp format.

    Definition Classes
    SparkFunctions
  189. val type_info: UserDefinedFunction
    Definition Classes
    SparkFunctions
  190. val type_info_with_includes: UserDefinedFunction
    Definition Classes
    SparkFunctions
  191. def unionAll(df: DataFrame*): DataFrame

    Method to take union of all passed dataframes.

    Method to take union of all passed dataframes.

    df

    list of dataframes for which to take union of.

    returns

    union of all passed input dataframes.

    Definition Classes
    DataHelpers
  192. val unique_identifier: UserDefinedFunction
    Definition Classes
    SparkFunctions
  193. val url_encode_escapes: UserDefinedFunction
    Definition Classes
    SparkFunctions
  194. val windowSpec: WindowSpec
    Definition Classes
    SparkFunctions
  195. def withSparkProperty[T](key: String, value: String, spark: SparkSession)(body: ⇒ T): T
  196. def withSubgraphName[T](value: String, spark: SparkSession)(body: ⇒ T): T
  197. def withTargetId[T](value: String, spark: SparkSession)(body: ⇒ T): T
  198. def writeDataFrame(df: DataFrame, path: String, spark: SparkSession, props: Map[String, String], format: String, partitionColumns: List[String] = Nil, bucketColumns: List[String] = Nil, numBuckets: Option[Int] = None, sortColumns: List[String] = Nil, tableName: Option[String] = None, databaseName: Option[String] = None): Unit

    Method to write data passed in dataframe in specific file format.

    Method to write data passed in dataframe in specific file format.

    df

    dataframe containing data.

    path

    path to write data to.

    spark

    spark session.

    props

    underlying data source specific properties.

    format

    file format in which to persist data. Supported file formats are csv, text, json, parquet, orc

    partitionColumns

    columns to be used for partitioning.

    bucketColumns

    used to bucket the output by the given columns. If specified, the output is laid out on the file-system similar to Hive's bucketing scheme.

    numBuckets

    number of buckets to be used.

    sortColumns

    columns on which to order data while persisting.

    tableName

    table name for persisting data.

    databaseName

    database name for persisting data.

    Definition Classes
    DataHelpers
  199. val writeIntegerToBytes: UserDefinedFunction
    Definition Classes
    SparkFunctions
  200. val writeLongToBytes: UserDefinedFunction
    Definition Classes
    SparkFunctions
  201. lazy val write_to_log: UserDefinedFunction

    UDF to write logging parameters to log port.

    UDF to write logging parameters to log port.

    Definition Classes
    DataHelpers
  202. val xmlToJSON: UserDefinedFunction
    Definition Classes
    SparkFunctions
  203. def yyyyMMdd_to_YYYYJJJ(in_date: Column): Column

    Converts yyyyyMMdd to YYYYJJJ

    Converts yyyyyMMdd to YYYYJJJ

    in_date

    date in yyyyMMdd format

    returns

    a date converted to YYYYJJJ

    Definition Classes
    SparkFunctions
  204. def zip_eventInfo_arrays(column1: Column, column2: Column): Column

    Method to zip two arrays with first one having event_type and second one having event_text

    Method to zip two arrays with first one having event_type and second one having event_text

    Definition Classes
    SparkFunctions
  205. object AbinitioDMLs
  206. object CDC
  207. object Component
  208. object DataFrameValidator
  209. object DataHelpers
  210. object FixedFileFormatImplicits
  211. object FixedFormatHelper
  212. object FixedFormatSchemaImplicits
  213. object RestAPIUtils
  214. object SchemaUtils
  215. object SparkFunctions
  216. object LongSequence
    Definition Classes
    SparkFunctions
  217. object LongWrappedArray
    Definition Classes
    SparkFunctions
  218. object SparkTestingUtils

Inherited from FixedFileFormatImplicits

Inherited from SparkFunctions

Inherited from DataHelpers

Inherited from Component

Inherited from UDFUtils

Inherited from LazyLogging

Inherited from Serializable

Inherited from Serializable

Inherited from RestAPIUtils

Inherited from ProphecyDataFrame

Inherited from AnyRef

Inherited from Any

Ungrouped