Packages

package libs

Ordering
  1. Alphabetic
  2. By Inheritance
Inherited
  1. libs
  2. ProphecySecrets
  3. FutureExtensions
  4. ExceptionExtensions
  5. StringExtensions
  6. TraversableExtensions
  7. Extension
  8. FixedFileFormatImplicits
  9. SparkFunctions
  10. DataHelpers
  11. LazyLogging
  12. Component
  13. UDFUtils
  14. Serializable
  15. Serializable
  16. RestAPIUtils
  17. ProphecyDataFrame
  18. AnyRef
  19. Any
  1. Hide All
  2. Show All
Visibility
  1. Public
  2. All

Type Members

  1. type Aggregate = Dataset[Row]
    Definition Classes
    Component
  2. class CDC extends AnyRef
  3. class CLIConf extends ScallopConf
  4. trait Component extends AnyRef
  5. case class UsesDataset(id: String, version: Int = -1) extends Annotation with StaticAnnotation with Product with Serializable
    Definition Classes
    Component
  6. case class UsesRuleset(id: String) extends Annotation with StaticAnnotation with Product with Serializable
    Definition Classes
    Component
  7. case class Visual(id: String = "ID", label: String = "Label", x: Long = 0, y: Long = 0, phase: Int = 0, mode: String = "batch", interimMode: String = "full", detailedStats: Boolean = false) extends Annotation with StaticAnnotation with Product with Serializable
    Definition Classes
    Component
  8. trait ConfigBase extends AnyRef
  9. abstract class ConfigurationFactory[C <: ConfigBase] extends AnyRef
  10. type CreateData = Dataset[Row]
    Definition Classes
    Component
  11. type DataFrame1 = Dataset[Row]
    Definition Classes
    Component
  12. type DataFrame10 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  13. type DataFrame11 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  14. type DataFrame12 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  15. type DataFrame13 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  16. type DataFrame14 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  17. type DataFrame15 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  18. type DataFrame16 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  19. type DataFrame17 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  20. type DataFrame18 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  21. type DataFrame19 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  22. type DataFrame2 = (DataFrame, DataFrame)
    Definition Classes
    Component
  23. type DataFrame20 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  24. type DataFrame21 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  25. type DataFrame22 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  26. type DataFrame3 = (DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  27. type DataFrame4 = (DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  28. type DataFrame5 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  29. type DataFrame6 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  30. type DataFrame7 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  31. type DataFrame8 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  32. type DataFrame9 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  33. trait DataHelpers extends LazyLogging

    Helper Utilities for reading/writing data from/to different data sources.

  34. type DataQualityTest = Dataset[Row]
    Definition Classes
    Component
  35. type DatabaseInput = Dataset[Row]
    Definition Classes
    Component
  36. type Deduplicate = Dataset[Row]
    Definition Classes
    Component
  37. case class Description(comment: String) extends Annotation with StaticAnnotation with Product with Serializable
  38. case class ExecutionMetricsException(message: String, cause: Option[Throwable] = None) extends ProphecyLibsException with Product with Serializable
  39. implicit class ExtendedDataFrameGlobal extends ExtendedDataFrame
  40. implicit class ExtendedStreamingTargetGlobal extends ExtendedStreamingTarget
  41. trait FFAST extends Positional
  42. case class FFCompoundSchemaRow(compound: FFCompoundType, rows: Seq[FFSchemaRow]) extends FFSchemaRow with Product with Serializable
  43. sealed trait FFCompoundType extends FFAST
  44. case class FFConditionalSchemaRow(condition: String, schemaRow: FFSchemaRow) extends FFSchemaRow with Product with Serializable
  45. sealed trait FFDataFormat extends FFAST
  46. case class FFDateFormat(name: FFTypeName, format: Option[String], miscProperties: Map[String, Any] = Map()) extends FFDataFormat with Product with Serializable
  47. case class FFDateTimeFormat(name: FFTypeName, format: Option[String], miscProperties: Map[String, Any] = Map()) extends FFDataFormat with Product with Serializable
  48. sealed trait FFDefaultVal extends FFAST
  49. case class FFDoubleDefaultVal(value: Double) extends FFDefaultVal with Product with Serializable
  50. case class FFExpressionDefaultVal(value: CustomExpression) extends FFDefaultVal with Product with Serializable
  51. case class FFIncludeFileRow(filePath: String) extends FFSchemaRow with Product with Serializable
  52. case class FFIntDefaultVal(value: Int) extends FFDefaultVal with Product with Serializable
  53. case class FFNoDefaultVal() extends FFDefaultVal with Product with Serializable
  54. case class FFNullDefaultVal(value: Option[Any] = None) extends FFDefaultVal with Product with Serializable
  55. case class FFNumberArrayFormat(name: FFTypeName, precision: Option[Int], scale: Option[Int], arraySizeInfo: Option[String], miscProperties: Map[String, Any] = ...) extends FFDataFormat with Product with Serializable
  56. case class FFNumberFormat(name: FFTypeName, precision: Option[Int], scale: Option[Int], miscProperties: Map[String, Any] = ...) extends FFDataFormat with Product with Serializable
  57. case class FFRecordType(startType: String) extends FFAST with Product with Serializable
  58. case class FFSchemaRecord(recordType: String, rows: Seq[FFSchemaRow]) extends FFAST with Product with Serializable
  59. sealed trait FFSchemaRow extends FFAST
  60. case class FFSimpleSchemaList(rows: Seq[FFSimpleSchemaRow]) extends FFSchemaRow with Product with Serializable
  61. case class FFSimpleSchemaRow(name: String, format: FFDataFormat, value: FFDefaultVal) extends FFSchemaRow with Product with Serializable
  62. case class FFStringArrayFormat(name: FFTypeName, precision: Option[Int], arraySizeInfo: Option[String]) extends FFDataFormat with Product with Serializable
  63. case class FFStringDefaultVal(value: String) extends FFDefaultVal with Product with Serializable
  64. case class FFStringFormat(name: FFTypeName, precision: Option[Int], props: Option[Map[String, String]] = None) extends FFDataFormat with Product with Serializable
  65. case class FFStructArrayType(name1: String, arraySizeInfo: Option[String], typeName: Option[String] = None) extends FFCompoundType with Product with Serializable
  66. case class FFStructFormat(name: FFTypeName, precision: Option[Int]) extends FFDataFormat with Product with Serializable
  67. case class FFStructType(name1: String, typeName: Option[String] = None) extends FFCompoundType with Product with Serializable
  68. case class FFTypeName(name: String, delimiter: Option[String]) extends FFAST with Product with Serializable
  69. case class FFTypeNameWithProperties(name: String, delimiter: Option[String], miscProperties: Map[String, Any] = Map("packed" → false)) extends FFAST with Product with Serializable
  70. case class FFUnionType(name: Option[String] = None, typeName: Option[String] = None) extends FFCompoundType with Product with Serializable
  71. case class FFUnknownFormat(name: FFTypeName, arraySizeInfo: Option[String]) extends FFDataFormat with Product with Serializable
  72. case class FFVoidFormat(name: FFTypeName, size: Option[Int]) extends FFDataFormat with Product with Serializable
  73. type FileInput = Dataset[Row]
    Definition Classes
    Component
  74. type FileIntermediate = Dataset[Row]
    Definition Classes
    Component
  75. type FileOutput = Unit
    Definition Classes
    Component
  76. type Filter = Dataset[Row]
    Definition Classes
    Component
  77. class FixedFileFormat extends FileFormat with DataSourceRegister with Serializable
  78. implicit class FixedFileFormatDataFrameGlobal extends FixedFileFormatDataFrame
  79. trait FixedFileFormatImplicits extends AnyRef
  80. implicit class FixedFileFormatDataFrame extends AnyRef
    Definition Classes
    FixedFileFormatImplicits
  81. implicit class FixedFileFormatSpark extends AnyRef
    Definition Classes
    FixedFileFormatImplicits
  82. type FixedFileOutput = Unit
    Definition Classes
    Component
  83. class FixedFormatOutputWriter extends OutputWriter
  84. type FlattenSchema = Dataset[Row]
    Definition Classes
    Component
  85. type Generate = Dataset[Row]
    Definition Classes
    Component
  86. type HashPartition = Dataset[Row]
    Definition Classes
    Component
  87. type Join = Dataset[Row]
    Definition Classes
    Component
  88. type Limit = Dataset[Row]
    Definition Classes
    Component
  89. type Lookup = UserDefinedFunction
    Definition Classes
    Component
  90. case class LookupDataset(datasetId: String, columnName: String) extends Annotation with StaticAnnotation with Product with Serializable
  91. type LookupFileInput = UserDefinedFunction
    Definition Classes
    Component
  92. type LookupUnit = Unit
    Definition Classes
    Component
  93. trait LookupUtils extends AnyRef
  94. class MDumpReader extends AnyRef
  95. type MultiFileRead = Dataset[Row]
    Definition Classes
    Component
  96. type MultiFileWrite = Unit
    Definition Classes
    Component
  97. type MultiFileWriteUnit = Unit
    Definition Classes
    Component
  98. type MultiJoin = Dataset[Row]
    Definition Classes
    Component
  99. type Normalize = Dataset[Row]
    Definition Classes
    Component
  100. type OrderBy = Dataset[Row]
    Definition Classes
    Component
  101. type OrderByPartition = Dataset[Row]
    Definition Classes
    Component
  102. type Prepare = Dataset[Row]
    Definition Classes
    Component
  103. class ProphecyLibsException extends Exception
  104. case class ProphecySparkAnalysisException(gemName: Option[String], e: AnalysisException) extends ProphecyLibsException with Product with Serializable
  105. case class PySparkException(className: String, message: String, trace: String, cause: Option[Throwable]) extends ProphecyLibsException with Product with Serializable
  106. type ReadSV = Dataset[Row]
    Definition Classes
    Component
  107. type Reformat = Dataset[Row]
    Definition Classes
    Component
  108. type Repartition = Dataset[Row]
    Definition Classes
    Component
  109. trait RestAPIUtils extends AnyRef

    Spark utilities for handling rest api connections.

  110. type RoundRobinPartition = Dataset[Row]
    Definition Classes
    Component
  111. type RowDistributor = Dataset[Row]
    Definition Classes
    Component
  112. type RowDistributor1 = Dataset[Row]
    Definition Classes
    Component
  113. type RowDistributor10 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  114. type RowDistributor11 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  115. type RowDistributor12 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  116. type RowDistributor13 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  117. type RowDistributor14 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  118. type RowDistributor15 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  119. type RowDistributor16 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  120. type RowDistributor17 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  121. type RowDistributor18 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  122. type RowDistributor19 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  123. type RowDistributor2 = (DataFrame, DataFrame)
    Definition Classes
    Component
  124. type RowDistributor20 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  125. type RowDistributor21 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  126. type RowDistributor22 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  127. type RowDistributor3 = (DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  128. type RowDistributor4 = (DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  129. type RowDistributor5 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  130. type RowDistributor6 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  131. type RowDistributor7 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  132. type RowDistributor8 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  133. type RowDistributor9 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  134. type Ruleset = Dataset[Row]
    Definition Classes
    Component
  135. type SQLStatement = Dataset[Row]
    Definition Classes
    Component
  136. type SQLStatement1 = Dataset[Row]
    Definition Classes
    Component
  137. type SQLStatement10 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  138. type SQLStatement11 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  139. type SQLStatement12 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  140. type SQLStatement13 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  141. type SQLStatement14 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  142. type SQLStatement15 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  143. type SQLStatement16 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  144. type SQLStatement17 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  145. type SQLStatement18 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  146. type SQLStatement19 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  147. type SQLStatement2 = (DataFrame, DataFrame)
    Definition Classes
    Component
  148. type SQLStatement20 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  149. type SQLStatement21 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  150. type SQLStatement22 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  151. type SQLStatement3 = (DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  152. type SQLStatement4 = (DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  153. type SQLStatement5 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  154. type SQLStatement6 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  155. type SQLStatement7 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  156. type SQLStatement8 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  157. type SQLStatement9 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  158. type SQLStatementUnit = Unit
    Definition Classes
    Component
  159. type Scan = Dataset[Row]
    Definition Classes
    Component
  160. type SchemaTransformer = Dataset[Row]
    Definition Classes
    Component
  161. type Script = Dataset[Row]
    Definition Classes
    Component
  162. type Script1 = Dataset[Row]
    Definition Classes
    Component
  163. type Script10 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  164. type Script11 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  165. type Script12 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  166. type Script13 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  167. type Script14 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  168. type Script15 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  169. type Script16 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  170. type Script17 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  171. type Script18 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  172. type Script19 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  173. type Script2 = (DataFrame, DataFrame)
    Definition Classes
    Component
  174. type Script20 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  175. type Script21 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  176. type Script22 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  177. type Script3 = (DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  178. type Script4 = (DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  179. type Script5 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  180. type Script6 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  181. type Script7 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  182. type Script8 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  183. type Script9 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  184. type ScriptUnit = Unit
    Definition Classes
    Component
  185. type Select = Dataset[Row]
    Definition Classes
    Component
  186. type Sequence = Dataset[Row]
    Definition Classes
    Component
  187. type SetOperation = Dataset[Row]
    Definition Classes
    Component
  188. type Source = Dataset[Row]
    Definition Classes
    Component
  189. trait SparkFunctions extends AnyRef

    Library of all spark functions which implements different abinitio functions used in abinitio workflows.

  190. class RecordIterator extends Serializable
    Definition Classes
    SparkFunctions
  191. class StringAsStream extends Serializable
    Definition Classes
    SparkFunctions
  192. type StreamingTarget = StreamingQuery
    Definition Classes
    Component
  193. type SubGraph = Dataset[Row]
    Definition Classes
    Component
  194. type SubGraph1 = Dataset[Row]
    Definition Classes
    Component
  195. type SubGraph10 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  196. type SubGraph11 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  197. type SubGraph12 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  198. type SubGraph13 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  199. type SubGraph14 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  200. type SubGraph15 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  201. type SubGraph16 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  202. type SubGraph17 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  203. type SubGraph18 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  204. type SubGraph19 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  205. type SubGraph2 = (DataFrame, DataFrame)
    Definition Classes
    Component
  206. type SubGraph20 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  207. type SubGraph21 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  208. type SubGraph22 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  209. type SubGraph3 = (DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  210. type SubGraph4 = (DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  211. type SubGraph5 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  212. type SubGraph6 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  213. type SubGraph7 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  214. type SubGraph8 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  215. type SubGraph9 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  216. type SubGraphUnit = Unit
    Definition Classes
    Component
  217. type Subgraph = Dataset[Row]
    Definition Classes
    Component
  218. type Subgraph1 = Dataset[Row]
    Definition Classes
    Component
  219. type Subgraph10 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  220. type Subgraph11 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  221. type Subgraph12 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  222. type Subgraph13 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  223. type Subgraph14 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  224. type Subgraph15 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  225. type Subgraph16 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  226. type Subgraph17 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  227. type Subgraph18 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  228. type Subgraph19 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  229. type Subgraph2 = (DataFrame, DataFrame)
    Definition Classes
    Component
  230. type Subgraph20 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  231. type Subgraph21 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  232. type Subgraph22 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  233. type Subgraph3 = (DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  234. type Subgraph4 = (DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  235. type Subgraph5 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  236. type Subgraph6 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  237. type Subgraph7 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  238. type Subgraph8 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  239. type Subgraph9 = (DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame, DataFrame)
    Definition Classes
    Component
  240. type SubgraphUnit = Unit
    Definition Classes
    Component
  241. type Target = Unit
    Definition Classes
    Component
  242. trait UDFUtils extends RestAPIUtils with Serializable

    Utility class with different UDFs to take care of miscellaneous tasks.

  243. case class LookupCondition(lookupColumn: String, comparisonOp: String, inputVariableName: String) extends Product with Serializable
    Definition Classes
    UDFUtils
  244. type UnionAll = Dataset[Row]
    Definition Classes
    Component
  245. implicit class UriString extends AnyRef
  246. type Visualize = Unit
    Definition Classes
    Component
  247. type WindowFunction = Dataset[Row]
    Definition Classes
    Component
  248. case class SecretValue(providerType: Option[String] = Some("Databricks"), secretScope: Option[String] = None, secretKey: Option[String] = None) extends Product with Serializable
    Definition Classes
    ProphecySecrets
  249. implicit class ExceptionExtension extends AnyRef
    Definition Classes
    ExceptionExtensions
  250. implicit class FutureExtension[T] extends AnyRef
    Definition Classes
    FutureExtensions
  251. implicit class OptionStringExtension extends AnyRef
    Definition Classes
    StringExtensions
  252. implicit class StringExtension extends OptionStringExtension
    Definition Classes
    StringExtensions
  253. implicit class SeqWrapper[+A, M[X] <: TraversableOnce[X]] extends LazyLogging
    Definition Classes
    TraversableExtensions
  254. implicit class ExtendedDataFrame extends AnyRef
    Definition Classes
    ProphecyDataFrame
  255. implicit class ExtendedStreamingTarget extends AnyRef
    Definition Classes
    ProphecyDataFrame
  256. implicit class ProphecyDataFrameReader extends AnyRef
    Definition Classes
    ProphecyDataFrame
  257. implicit class ProphecyDataFrameWriter[T] extends AnyRef
    Definition Classes
    ProphecyDataFrame

Abstract Value Members

  1. abstract def getClass(): Class[_]
    Definition Classes
    Any

Concrete Value Members

  1. final def !=(arg0: Any): Boolean
    Definition Classes
    Any
  2. final def ##(): Int
    Definition Classes
    Any
  3. final def ==(arg0: Any): Boolean
    Definition Classes
    Any
  4. final val ASCII_NEWLINE_INT_VALUE: Int(10)
    Definition Classes
    FixedFileFormatImplicits
  5. final val EBCDIC_NEWLINE_INT_VALUE: Int(21)
    Definition Classes
    FixedFileFormatImplicits
  6. val InterimState: InterimStore.type
    Definition Classes
    ProphecyDataFrame
  7. final val SKIP_FOOTER_LINES: String("skip_footer_lines")
    Definition Classes
    FixedFileFormatImplicits
  8. final val SKIP_HEADER_LINES: String("skip_header_lines")
    Definition Classes
    FixedFileFormatImplicits
  9. def YJJJ_to_YYYYJJJ(in_date: Column, ref_date: Column): Column

    Converts 1 digit julian year to 4 digits julian year.

    Converts 1 digit julian year to 4 digits julian year.

    in_date

    date in Julian in "YJJJ" format

    ref_date

    date in "yyyyMMdd" format

    returns

    a date in "YYYYJJJ"

    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  10. lazy val adjustCenturyDateInCyyFormat: UserDefinedFunction

    begining of input should have Cyy

    begining of input should have Cyy

    Definition Classes
    SparkFunctions
  11. def adjustStringRegexPattern(input: String): String
    Definition Classes
    SparkFunctions
  12. def appendTrailer(pathInputData: String, pathInputTrailer: String, pathOutputConcatenated: String, configuration: Configuration): Unit

    Appends a trailer data to every single file in the data directory.

    Appends a trailer data to every single file in the data directory. A single trailer file in the pathOutputTrailer directory should correspond to a single data file in the pathOutputData directory.

    If a trailer for a given file does not exist, the file is moved as is to the output directory.

    pathInputData

    Input data files directory

    pathInputTrailer

    Input trailer files directory

    pathOutputConcatenated

    Output concatenated files directory

    configuration

    Hadoop configuration (preferably sparkSession.sparkContext.hadoopConfiguration)

    Definition Classes
    DataHelpers
  13. def arrayColumn(value: String, values: String*): Column

    Function to take variable number of values and create an array column out of it.

    Function to take variable number of values and create an array column out of it.

    value

    input value

    values

    variable number of input values.

    returns

    an array of column.

    Definition Classes
    UDFUtils
  14. lazy val array_value: UserDefinedFunction

    UDF to find and return element in arr sequence at passed index.

    UDF to find and return element in arr sequence at passed index. If no element found then null is returned.

    Definition Classes
    UDFUtils
  15. final def asInstanceOf[T0]: T0
    Definition Classes
    Any
  16. lazy val bigDecimalToPackedBytes: UserDefinedFunction
    Definition Classes
    SparkFunctions
  17. lazy val call_rest_api: UserDefinedFunction

    Spark UDF that makes a single blocking rest API call to a given url.

    Spark UDF that makes a single blocking rest API call to a given url. The result of this udf is always produced, contains a proper error if it failed at any stage, and never interrupts the job execution (unless called with invalid signature).

    The default timeout can be configured through the spark.network.timeout Spark configuration option.

    Parameters:

    • method - any supported HTTP1.1 method type, e.g. POST, GET. Complete list: [httpMethods].
    • url - valid url to which a request is going to be made
    • headers - an array of "key: value" headers that are past with the request
    • content - any content (by default, the supported rest api content type is application/json)

    Response - a struct with the following fields:

    • isSuccess - boolean, whether a successful response has been received
    • status - nullable integer, status code (e.g. 404, 200, etc)
    • headers - an array of name: value response headers (e.g. [Server: akka-http/10.1.10, Date: Tue, 07 Sep 2021 18:11:47 GMT])
    • content - nullable string, response back
    • error - nullable string, if the parameters passed are valid or the system failed to make a call, this field contains an error message
    Definition Classes
    RestAPIUtils
  18. def call_udf(udfName: String, cols: Column*): Column

    Taken from upstream Spark

    Taken from upstream Spark

    Definition Classes
    UDFUtils
    Annotations
    @varargs()
  19. lazy val canonical_representation: UserDefinedFunction
    Definition Classes
    SparkFunctions
  20. def castDataType(sparkSession: SparkSession, df: DataFrame, column: Column, dataType: String, replaceColumn: String): DataFrame

    Function to add new typecasted column in input dataframe.

    Function to add new typecasted column in input dataframe. Newly added column is typecasted version of passed column. Typecast operation is supported for string, boolean, byte, short, int, long, float, double, decimal, date, timestamp

    sparkSession

    spark session

    df

    input dataframe

    column

    input column to be typecasted

    dataType

    datatype to cast column to.

    replaceColumn

    column name to be added in dataframe.

    returns

    new dataframe with new typecasted column.

    Definition Classes
    UDFUtils
  21. lazy val char_string: UserDefinedFunction
    Definition Classes
    SparkFunctions
  22. def computeChecksum(df: DataFrame): DataFrame
    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  23. def concatenate(sources: Seq[String], destination: String, compressToGZip: Boolean = false): Unit

    Method to get data from multiple source paths and combine it into single destination path.

    Method to get data from multiple source paths and combine it into single destination path.

    sources

    multiple source paths from which to merge the data.

    destination

    destination path to combine all data to.

    compressToGZip

    flag to compress final output file into gzip format

    Definition Classes
    DataHelpers
  24. def concatenateFiles(spark: SparkSession, format: String, mode: String, inputDir: String, outputFileName: String, deleteTempPath: Boolean = true, fileFormatHasHeaders: Boolean = false): Unit

    Method to get data from multiple part files in source directory and combine it into single file.

    Method to get data from multiple part files in source directory and combine it into single file.

    spark

    spark session

    format

    file extension. e.g.: ".csv", ".txt"

    mode

    write mode in spark. Can be overwrite, append, error and ignore

    inputDir

    input directory containing part files

    outputFileName

    output single file path

    deleteTempPath

    flag to delete temp source directory

    fileFormatHasHeaders

    flag to exclude headers from file

    Definition Classes
    DataHelpers
    Annotations
    @Py4JWhitelist()
  25. def convertInputBytesToStructType(input: Any, typeInfo: Seq[String], startByte: Int = 0): Row

    Method used for abinitio's reinterpret_as function to read necessary bytes from byteArray for input data and convert into struct format as per provided in typeInfo sequence.

    Method used for abinitio's reinterpret_as function to read necessary bytes from byteArray for input data and convert into struct format as per provided in typeInfo sequence.

    TypeInfo can have multiple entries, each could be either decimal or string type. Depending on the argument passed within decimal or string bytes are read from input byte array.

    If decimal or string argument has some integer then that many bytes are read from input byte array or if decimal or string has some string delimiter as its argument then from the current position bytes are read until string delimiter is found in input byte array.

    Definition Classes
    SparkFunctions
  26. def createDataFrameFromData(inputData: String, delimiter: String, columnName: String, columnType: String, sparkSession: SparkSession): DataFrame

    Method to read values from inputData and create dataframe with column name as columnName and column type as columnType for the values in inputData delimiter by delimiter.

    Method to read values from inputData and create dataframe with column name as columnName and column type as columnType for the values in inputData delimiter by delimiter.

    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  27. def createExtendedLookup(name: String, df: DataFrame, spark: SparkSession, conditions: List[LookupCondition], inputParams: List[String], valueColumns: String*): UserDefinedFunction

    Extended Lookup creates a special lookup to support the informatica lookup node functionality

    Extended Lookup creates a special lookup to support the informatica lookup node functionality

    conditions

    : condition used to filter the rows

    inputParams

    : input parameters

    Definition Classes
    UDFUtils
  28. def createLookup(name: String, df: DataFrame, spark: SparkSession, keyCols: List[String], rowCols: String*): UserDefinedFunction

    Function registers 4 different UDFs with spark registry.

    Function registers 4 different UDFs with spark registry. UDF for lookup_match, lookup_count, lookup_row and lookup functions are registered. This function stores the data of input dataframe in a broadcast variable, then uses this broadcast variable in different lookup functions.

    lookup : This function returns the first matching row for given input keys lookup_count : This function returns the count of all matching rows for given input keys. lookup_match : This function returns 0 if there is no matching row and 1 for some matching rows for given input keys. lookup_row : This function returns all the matching rows for given input keys.

    This function registers for upto 10 matching keys as input to these lookup functions.

    name

    UDF Name

    df

    input dataframe

    spark

    spark session

    keyCols

    columns to be used as keys in lookup functions.

    rowCols

    schema of entire row which will be stored for each matching key.

    returns

    registered UDF definitions for lookup functions. These UDF functions returns different results depending on the lookup function.

    Definition Classes
    UDFUtils
  29. def createRangeLookup(name: String, df: DataFrame, spark: SparkSession, minColumn: String, maxColumn: String, valueColumns: String*): UserDefinedFunction

    Method to create UDF which looks for passed input double in input dataframe.

    Method to create UDF which looks for passed input double in input dataframe. This function first loads the data of dataframe in broadcast variable and then defines a UDF which looks for input double value in the data stored in broadcast variable. If input double lies between passed col1 and col2 values then it adds corresponding row in the returned result. If value of input double doesn't lie between col1 and col2 then it simply returns null for current row in result.

    name

    created UDF name

    df

    input dataframe

    spark

    spark session

    minColumn

    column whose value to be considered as minimum in comparison.

    maxColumn

    column whose value to be considered as maximum in comparison.

    valueColumns

    remaining column names to be part of result.

    returns

    registers UDF which in turn returns rows corresponding to each row in dataframe on which range UDF is called.

    Definition Classes
    UDFUtils
  30. implicit def createSparkSessionExtension(spark: SparkSession): ProphecySparkSession
    Definition Classes
    Extension
  31. def createTmpDir(spark: SparkSession): String
    Definition Classes
    DataHelpers
  32. lazy val cross_join_index_range: UserDefinedFunction
    Definition Classes
    SparkFunctions
  33. def date_add_months(inputDate: Column, months: Int): Column

    Returns the internal representation of a date resulting from adding (or subtracting) a number of months to the specified date.

    Returns the internal representation of a date resulting from adding (or subtracting) a number of months to the specified date.

    inputDate

    in yyyy-MM-dd format

    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  34. def date_difference_days(laterDate: Column, earlierDate: Column): Column

    Computes number of days between two specified dates in "yyyyMMdd" format

    Computes number of days between two specified dates in "yyyyMMdd" format

    laterDate

    input date

    earlierDate

    input date

    returns

    number of days between laterDate and earlierDate or null if either one is null

    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  35. lazy val date_month_end: UserDefinedFunction
    Definition Classes
    SparkFunctions
  36. lazy val datetime_add: UserDefinedFunction
    Definition Classes
    SparkFunctions
  37. def datetime_add_months(input: Column, months: Int): Column

    Returns the internal representation of a timestamp resulting from adding (or subtracting) a number of months to the specified timestamp.

    Returns the internal representation of a timestamp resulting from adding (or subtracting) a number of months to the specified timestamp.

    input

    timestamp in yyyy-MM-dd HH:mm:ss.SSSS format

    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  38. lazy val datetime_difference: UserDefinedFunction
    Definition Classes
    SparkFunctions
  39. def datetime_difference_hours(end: Column, start: Column): Column

    Returns the number of hours between two specified dates in standard format yyyy-MM-dd HH:mm:ss.SSSS.

    Returns the number of hours between two specified dates in standard format yyyy-MM-dd HH:mm:ss.SSSS.

    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  40. def datetime_difference_minutes(end: Column, start: Column): Column

    Returns the number of minutes between two specified dates in standard format yyyy-MM-dd HH:mm:ss.SSSS.

    Returns the number of minutes between two specified dates in standard format yyyy-MM-dd HH:mm:ss.SSSS.

    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  41. def datetime_difference_seconds(end: Column, start: Column): Column

    Returns the number of minutes between two specified dates in standard format yyyy-MM-dd HH:mm:ss.SSSS.

    Returns the number of minutes between two specified dates in standard format yyyy-MM-dd HH:mm:ss.SSSS.

    Definition Classes
    SparkFunctions
  42. def datetime_from_unixtime(seconds: Column): Column
    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  43. def decimal_lpad(input: Column, len: Int, char_to_pad_with: String = "0", decimal_point_char: String = "."): Column

    Method uses a java regex to identify decimal numbers from input string.

    Method uses a java regex to identify decimal numbers from input string. This decimal number could be of 3 types 1. Simple integral number. e.g. 013334848. This part is identified by regex. 2. decimal number with explicit decimal point. e.g. 123456.90. This part is identified by combination of [0-9]+(\$$decimal_point_char)[0-9]+ and (0\$$decimal_point_char)[0-9]+ regex

    After extracting decimal number this code checks if length of decimal number is more than len parameter or not. If length is more than len parameter then it simply returns this extracted decimal number. Otherwise it first left pad decimal number with char_to_pad_with to make its length equal to len parameter and then adjusts minus sign (-) to left most part of decimal number.

    input

    input string.

    len

    length of characters.

    char_to_pad_with

    character to left pad with. default value is "0"

    decimal_point_char

    A string that specifies the character that represents the decimal point.

    returns

    a decimal string of the specified length or longer, left-padded with a specified character as needed and trimmed of leading zeros.

    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  44. def decimal_lrepad(input: Column, len: Int, char_to_pad_with: String = "0", decimal_point_char: String = "."): Column

    Method uses a java regex to identify decimal numbers from input string.

    Method uses a java regex to identify decimal numbers from input string. This decimal number could be of 3 types 1. Simple integral number. e.g. 013334848. This part is identified by combination of [1-9][0-9]*[0-9] and [1-9]+ regex 2. decimal number with explicit decimal point. e.g. 123456.90. This part is identified by combination of [1-9][0-9]*(\\\$$decimal_point_char)[0-9]+ and (0\\\$$decimal_point_char)[0-9]*[0-9] regex

    After extracting decimal number this code checks if length of decimal number is more than len parameter or not. If length is more than len parameter then it simply returns this extracted decimal number. Otherwise it first left pad decimal number with char_to_pad_with to make its length equal to len parameter and then adjusts minus sign (-) to left most part of decimal number.

    input

    input string.

    len

    length of characters.

    char_to_pad_with

    character to left pad with. default value is "0"

    decimal_point_char

    A string that specifies the character that represents the decimal point.

    returns

    a decimal string of the specified length or longer, left-padded with a specified character as needed and trimmed of leading zeros.

    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  45. def decimal_round(input: Column, places: Int): Column
    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  46. def decimal_round_down(input: Column, right_digits: Int): Column

    Function returns a value which is rounded down to right_digits number of digits to the right of decimal point.

    Function returns a value which is rounded down to right_digits number of digits to the right of decimal point.

    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  47. def decimal_round_up(input: Column, places: Int): Column

    Returns a number rounded up to a specified number of places to the right of the decimal point.

    Returns a number rounded up to a specified number of places to the right of the decimal point.

    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  48. def decimal_strip(input: Column, decimal_point_char: String = "."): Column

    Function uses a java regex to identify decimal numbers from input string.

    Function uses a java regex to identify decimal numbers from input string. This decimal number could be of 3 types 1. Simple integral number. e.g. 013334848. This part is identified by combination of [1-9][0-9 ]*[0-9] and [1-9]+ regex 2. decimal number with explicit decimal point. e.g. 123456.90. This part is identified by combination of [1-9][0-9]*(\$$decimal_point_char)[0-9 ]+ and (0\$$decimal_point_char)[0-9 ]*[0-9] regex

    After extracting decimal number this code looks for minus sign before extracted number in input and appends it with decimal number if found minus sign.

    In the end it replaces all whitespaces with empty string in the final resultant decimal number.

    input

    input string

    decimal_point_char

    A string that specifies the character that represents the decimal point.

    returns

    a decimal from a string that has been trimmed of leading zeros and non-numeric characters.

    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  49. def decimal_truncate(input: Column, number_of_places: Column): Column
    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  50. lazy val decodeBytes: UserDefinedFunction
    Definition Classes
    SparkFunctions
  51. lazy val decodeString: UserDefinedFunction
    Definition Classes
    SparkFunctions
  52. lazy val decode_datetime: UserDefinedFunction

    UDF to get record of type decode_datetime_type.

    UDF to get record of type decode_datetime_type. This record will have all its fields populated with corresponding entries in input date/timestamp.

    Returned record will have following schema.

    integer(8) year; integer(8) month; integer(8) day; integer(8) hour; integer(8) minute; integer(8) second; integer(8) microsecond;

    Note: Supported Input time is in yyyy-MM-dd HH:mm:ss.SSSSSS or yyyy-MM-dd HH:mm:ss or yyyy-MM-dd formats only. Additional handling is done to support timestamp retrieved from now() function call.

    Definition Classes
    SparkFunctions
  53. lazy val decode_datetime_as_local: UserDefinedFunction
    Definition Classes
    SparkFunctions
  54. def deleteDir(inputDir: String, spark: SparkSession): Boolean
    Definition Classes
    DataHelpers
  55. lazy val directory_listing: UserDefinedFunction
    Definition Classes
    SparkFunctions
  56. lazy val directory_listing_dir_only: UserDefinedFunction
    Definition Classes
    SparkFunctions
  57. def directory_listing_scala(path: String, filePrefix: String): Column
    Definition Classes
    SparkFunctions
  58. def dropColumns(sparkSession: SparkSession, df: DataFrame, columns: Column*): DataFrame

    Function to drop passed columns from input dataframe.

    Function to drop passed columns from input dataframe.

    sparkSession

    spark session

    df

    input dataframe.

    columns

    list of columns to be dropped from dataframe.

    returns

    new dataframe with dropped columns.

    Definition Classes
    UDFUtils
  59. lazy val encodeBytes: UserDefinedFunction
    Definition Classes
    SparkFunctions
  60. lazy val encodeString: UserDefinedFunction
    Definition Classes
    SparkFunctions
  61. lazy val encode_date: UserDefinedFunction

    integer values specifying days relative to January 1, 1900.

    integer values specifying days relative to January 1, 1900. This function returns the internal representation of a date given the year, month, and date. encode_date returns the internal representation of the date specified by the year 1998, the month 5, and the day 18:encode_date(1998, 5, 18) = 35931

    Definition Classes
    SparkFunctions
  62. lazy val encrypt_idwdata: UserDefinedFunction
    Definition Classes
    SparkFunctions
  63. def ends_with(input: Column, suffix: String): Column

    Returns true if string columns ends with given suffix

    Returns true if string columns ends with given suffix

    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  64. def equals(arg0: Any): Boolean
    Definition Classes
    Any
  65. lazy val eval: UserDefinedFunction

    Method to return the result of evaluating a string expression in the context of a specified input column.

    Method to return the result of evaluating a string expression in the context of a specified input column. Here input column could be struct type record, simple column, array type etc. Here expr could be reference to nested column inside input column or any expression which requires values from input column for its evaulation.

    Note: Current implementation only supports scenerio where input column is of struct type and expr is simply dot separated column reference to input struct.

    Definition Classes
    SparkFunctions
  66. def executeNonSelectSQLQueries(sqlList: Seq[String], dbConnection: Connection): Unit
    Definition Classes
    DataHelpers
  67. def extended_lookup(lookupName: String, cols: Column*): Column
    Definition Classes
    UDFUtils
  68. def extended_lookup_any(lookupName: String, cols: Column*): Column
    Definition Classes
    UDFUtils
  69. def extended_lookup_first(lookupName: String, cols: Column*): Column
    Definition Classes
    UDFUtils
  70. def extended_lookup_last(lookupName: String, cols: Column*): Column
    Definition Classes
    UDFUtils
  71. lazy val ff3_encrypt_idwdata: UserDefinedFunction
    Definition Classes
    SparkFunctions
  72. lazy val file_information: UserDefinedFunction

    UDF to get file information for passed input file path.

    UDF to get file information for passed input file path.

    Definition Classes
    SparkFunctions
  73. def findFirstElement(input: Column, default: Column = lit(null)): Column
    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  74. def findFirstNonBlankElement(input: Column, default: Column): Column
    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  75. def findLastElement(input: Column, default: Column = lit(null)): Column
    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  76. def first_defined(expr1: Column, expr2: Column): Column

    Method to identify and return first non null expression.

    Method to identify and return first non null expression.

    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  77. lazy val first_defined_for_double_Udf: UserDefinedFunction
    Definition Classes
    SparkFunctions
  78. def flattenStructSchema(schema: StructType, prefix: String = null): Array[Column]
    Definition Classes
    SparkFunctions
  79. lazy val force_error: UserDefinedFunction
    Definition Classes
    SparkFunctions
  80. def format_decimal(input: Column, scale: Int): Column
    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  81. def from_sv(input: Column, separator: String, schema: StructType): Column
    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  82. def from_xml(content: Column, schema: StructType): Column
    Definition Classes
    SparkFunctions
  83. def from_xml(content: Column, schemaJSON: String): Column
    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  84. def generateDataFrameWithSequenceColumn(start: Int, end: Int, columnName: String, sparkSession: SparkSession): DataFrame

    Method to create dataframe with single column containing increasing sequence id from start to end.

    Method to create dataframe with single column containing increasing sequence id from start to end.

    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  85. def generate_sequence(start: Int, end: Int, step: Int = 1): Column

    Function to create sequence of array between two passed numbers

    Function to create sequence of array between two passed numbers

    start

    starting point of generated sequence

    end

    terminating point of generated sequence.

    returns

    column containing sequence of integers.

    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  86. lazy val generate_sequence: UserDefinedFunction

    UDF to generate column with sequence of integers between two passed start and end columns.

    UDF to generate column with sequence of integers between two passed start and end columns.

    Definition Classes
    SparkFunctions
  87. lazy val getByteFromByteArray: UserDefinedFunction

    UDF to get last Byte from ByteArray of input data.

    UDF to get last Byte from ByteArray of input data.

    Definition Classes
    SparkFunctions
  88. def getColumnInSecondArrayByFirstNonBlankPositionInFirstArray(nonBlankEntryExpr: Column, firstArray: Column, secondArray: Column): Column
    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  89. def getContentAsStream(content: String): StringAsStream
    Definition Classes
    SparkFunctions
  90. lazy val getDefaultedElseTrimmed: UserDefinedFunction
    Definition Classes
    SparkFunctions
  91. def getEmptyDataFrame(sparkSession: SparkSession): DataFrame
    Definition Classes
    DataHelpers
  92. def getEmptyDataFrameWithSchema(sparkSession: SparkSession, schema: StructType): DataFrame
    Definition Classes
    DataHelpers
  93. def getEmptyLogDataFrame(sparkSession: SparkSession): DataFrame

    Method to get empty dataframe with below abinitio log schema.

    Method to get empty dataframe with below abinitio log schema.

    record string("|") node, timestamp, component, subcomponent, event_type; string("|\n") event_text; end

    Definition Classes
    DataHelpers
    Annotations
    @Py4JWhitelist()
  94. def getFebruaryDay(year: Column): Column

    Computes number of days in February month in a given year

    Computes number of days in February month in a given year

    year

    year whose number of days in February needs to be calculated

    returns

    number of days

    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  95. def getFieldFromStructByPosition(column: Column, position: Int): Column

    Method to get field at specific position from struct column

    Method to get field at specific position from struct column

    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  96. lazy val getIntArrayFromByteArray: UserDefinedFunction

    UDF to get long comprising of last 8 Bytes from ByteArray of input data.

    UDF to get long comprising of last 8 Bytes from ByteArray of input data.

    Definition Classes
    SparkFunctions
  97. lazy val getIntFromByteArray: UserDefinedFunction

    UDF to get integer comprising of last 4 Bytes from ByteArray of input data.

    UDF to get integer comprising of last 4 Bytes from ByteArray of input data.

    Definition Classes
    SparkFunctions
  98. lazy val getLongArrayFromByteArray: UserDefinedFunction

    UDF to get long comprising of last 8 Bytes from ByteArray of input data.

    UDF to get long comprising of last 8 Bytes from ByteArray of input data.

    Definition Classes
    SparkFunctions
  99. lazy val getLongFromByteArray: UserDefinedFunction

    UDF to get long comprising of last 8 Bytes from ByteArray of input data.

    UDF to get long comprising of last 8 Bytes from ByteArray of input data.

    Definition Classes
    SparkFunctions
  100. def getMTimeDataframe(filepath: String, format: String, spark: SparkSession): DataFrame
    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  101. lazy val getShortFromByteArray: UserDefinedFunction

    UDF to get short comprising of last 2 Bytes from ByteArray of input data.

    UDF to get short comprising of last 2 Bytes from ByteArray of input data.

    Definition Classes
    SparkFunctions
  102. def hashCode(): Int
    Definition Classes
    Any
  103. lazy val hash_MD5: UserDefinedFunction
    Definition Classes
    SparkFunctions
  104. lazy val hash_SHA1: UserDefinedFunction
    Definition Classes
    SparkFunctions
  105. def hash_SHA512(input: Column): Column
    Definition Classes
    SparkFunctions
  106. def hash_value(input: Column, keys: Seq[String], hashAlgorithm: String): Column
    Definition Classes
    SparkFunctions
  107. lazy val instr_extended_udf: UserDefinedFunction
    Definition Classes
    SparkFunctions
  108. lazy val instr_udf: UserDefinedFunction
    Definition Classes
    SparkFunctions
  109. final def isInstanceOf[T0]: Boolean
    Definition Classes
    Any
  110. lazy val isNotEqualToValue: UserDefinedFunction
    Definition Classes
    SparkFunctions
  111. lazy val isNotNullAndNotBlank: UserDefinedFunction
    Definition Classes
    SparkFunctions
  112. lazy val isNullOrBlank: UserDefinedFunction
    Definition Classes
    SparkFunctions
  113. def isNullOrEmpty(input: Column): Column

    Method to check if current column is null or has empty value.

    Method to check if current column is null or has empty value.

    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  114. def is_ascii(input: Column): Column

    Checks if a string is ascii

    Checks if a string is ascii

    input

    column to be checked

    returns

    true if the input string is ascii otherwise false

    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  115. def is_blank(input: Column): Column

    Method to identify if input string is a blank string or not.

    Method to identify if input string is a blank string or not.

    input

    input string.

    returns

    return 1 if given string contains all blank character or is a zero length string, otherwise it returns 0

    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  116. lazy val is_blank_udf: UserDefinedFunction
    Definition Classes
    SparkFunctions
  117. lazy val is_bzero: UserDefinedFunction

    Tests whether an object is composed of all binary zero bytes.

    Tests whether an object is composed of all binary zero bytes. This function returns: 1. 1 if obj contains only binary zero bytes or is a zero-length string 2. 0 if obj contains any non-zero bytes 3. NULL if obj is NULL

    Definition Classes
    SparkFunctions
  118. def is_numeric_ascii(input: Column): Column

    Checks if an input string contains only ascii code and numbers

    Checks if an input string contains only ascii code and numbers

    input

    string to be checked

    returns

    true if input string contains only ascii code and numbers or null if input is null

    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  119. def is_valid(input: Column, isNullable: Boolean, formatInfo: Option[Any], len: Option[Seq[Int]]): Column

    Method to identify if passed input column is a valid expression after typecasting to passed dataType.

    Method to identify if passed input column is a valid expression after typecasting to passed dataType. Also while typecasting if len is present then this function also makes sure the max length of input column after typecasting operation is not greater than len.

    input

    input column expression to be identified if is valid.

    formatInfo

    datatype to which input column expression must be typecasted. If datatype is a string then it is treated as timestamp format. If it is a list of string then it is treated as having current timestamp format and and new timestamp format to which input column needs to be typecasted.

    len

    max length of input column after typecasting it to dataType.

    returns

    0 if input column is not valid after typecasting or 1 if it is valid.

    Definition Classes
    SparkFunctions
  120. def is_valid(input: Column, isNullable: Boolean, formatInfo: Option[Any]): Column
    Definition Classes
    SparkFunctions
  121. def is_valid(input: Column, formatInfo: Option[Any], len: Option[Seq[Int]]): Column
    Definition Classes
    SparkFunctions
  122. def is_valid(input: Column, formatInfo: Option[Any]): Column
    Definition Classes
    SparkFunctions
  123. def is_valid(input: Column, isNullable: Boolean): Column
    Definition Classes
    SparkFunctions
  124. def is_valid(input: Column): Column
    Definition Classes
    SparkFunctions
  125. def is_valid_date(dateFormat: String, inDate: Column): Column

    Validates date against a input format

    Validates date against a input format

    dateFormat

    A pattern such as yyyy-MM-dd or yyyy-MM-dd HH:mm:ss.SSSS or dd.MM.yyyy

    inDate

    Input date to be validated

    returns

    true if the input date is valid otherwise false

    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  126. def is_valid_python_bridge(input: Column, isNullable: Boolean, formatSerialized: Option[Any], len: Option[Seq[Int]]): Column
    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  127. def lastElementInCurrentWindow(input: Column): Column
    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  128. def loadBinaryFileAsBinaryDataFrame(filePath: String, lineDelimiter: String = "\n", minPartition: Int = 1, rowName: String = "line", spark: SparkSession): DataFrame
    Definition Classes
    DataHelpers
    Annotations
    @Py4JWhitelist()
  129. def loadBinaryFileAsStringDataFrame(filePath: String, lineDelimiter: String = "\n", charSetEncoding: String = "Cp1047", minPartition: Int = 1, rowName: String = "line", spark: SparkSession): DataFrame
    Definition Classes
    DataHelpers
    Annotations
    @Py4JWhitelist()
  130. def loadFixedWindowBinaryFileAsDataFrame(filePath: String, lineLength: Int, minPartition: Int = 1, rowName: String = "line", spark: SparkSession): DataFrame
    Definition Classes
    DataHelpers
    Annotations
    @Py4JWhitelist()
  131. lazy val logger: Logger
    Attributes
    protected
    Definition Classes
    LazyLogging
    Annotations
    @transient()
  132. def lookup(lookupName: String, cols: Column*): Column

    By default returns only the first matching record

    By default returns only the first matching record

    Definition Classes
    UDFUtils
  133. def lookup_count(lookupName: String, cols: Column*): Column
    Definition Classes
    UDFUtils
  134. def lookup_last(lookupName: String, cols: Column*): Column

    Returns the last matching record

    Returns the last matching record

    Definition Classes
    UDFUtils
  135. def lookup_match(lookupName: String, cols: Column*): Column

    returns

    Boolean Column

    Definition Classes
    UDFUtils
  136. def lookup_nth(lookupName: String, cols: Column*): Column
    Definition Classes
    UDFUtils
  137. def lookup_range(lookupName: String, input: Column): Column
    Definition Classes
    UDFUtils
  138. def lookup_row(lookupName: String, cols: Column*): Column
    Definition Classes
    UDFUtils
  139. def lookup_row_reverse(lookupName: String, cols: Column*): Column
    Definition Classes
    UDFUtils
  140. lazy val make_byte_flags: UserDefinedFunction

    UDF to return a flag for each character if it is present or not in input String.

    UDF to return a flag for each character if it is present or not in input String.

    Definition Classes
    SparkFunctions
  141. def make_constant_vector(size: Int, seedVal: Int): Array[Int]

    Method to create array of size "size" containing seedVal as each entry

    Method to create array of size "size" containing seedVal as each entry

    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  142. def make_constant_vector(size: Int, seedVal: Column): Column

    Method to create array of size "size" containing seedVal as each entry

    Method to create array of size "size" containing seedVal as each entry

    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  143. def math_min(input: Column*): Column
    Definition Classes
    SparkFunctions
  144. def measure[T](fn: ⇒ T)(caller: String = findCaller()): T
    Definition Classes
    UDFUtils
  145. lazy val multi_regex_match: UserDefinedFunction
    Definition Classes
    SparkFunctions
  146. def multi_regex_replace_with_char_conversion(input: Column, charSet: Column, replaceStr: Column, replacement0: String, replacement1: String, pattern: String*): Column
    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  147. lazy val multifile_information: UserDefinedFunction

    UDF to get multifile information for passed input file path.

    UDF to get multifile information for passed input file path.

    Definition Classes
    SparkFunctions
  148. lazy val murmur: UserDefinedFunction

    UDF for murmur hash generation for any column type

    UDF for murmur hash generation for any column type

    Definition Classes
    SparkFunctions
  149. def now(): Column

    Method to get current timestamp.

    Method to get current timestamp.

    returns

    current timestamp in YYYYMMddHHmmssSSSSSS format.

    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  150. def numberOfPartitions(in: DataFrame): Column
    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  151. lazy val number_grouping: UserDefinedFunction

    udf to group input decimal into multiple groups separated by separator

    udf to group input decimal into multiple groups separated by separator

    Definition Classes
    SparkFunctions
  152. lazy val packedBytesStringToDecimal: UserDefinedFunction
    Definition Classes
    SparkFunctions
  153. lazy val packedBytesToDecimal: UserDefinedFunction
    Definition Classes
    SparkFunctions
  154. lazy val re_get_match: UserDefinedFunction

    Returns the first string in a target string that matches a regular expression.

    Returns the first string in a target string that matches a regular expression.

    Definition Classes
    SparkFunctions
  155. lazy val re_get_match_with_index: UserDefinedFunction
    Definition Classes
    SparkFunctions
  156. lazy val re_get_matches: UserDefinedFunction
    Definition Classes
    SparkFunctions
  157. lazy val re_get_matches_with_offset: UserDefinedFunction
    Definition Classes
    SparkFunctions
  158. lazy val re_index: UserDefinedFunction

    UDF wrapper over re_index function.

    UDF wrapper over re_index function.

    Definition Classes
    SparkFunctions
  159. lazy val re_index_with_offset: UserDefinedFunction

    Returns the first string in a target string that matches a regular expression.

    Returns the first string in a target string that matches a regular expression.

    Definition Classes
    SparkFunctions
  160. def re_replace(target: Column, pattern: String, replacement: String, offset: Int = 0): Column

    Replaces all substrings in a target string that match a specified regular expression.

    Replaces all substrings in a target string that match a specified regular expression.

    target

    A string that the function searches for a substring that matches pattern_expr.

    pattern

    regular expression

    replacement

    replacement string

    offset

    Number of characters, from the beginning of str, to skip before searching.

    returns

    a replaced string in which all substrings, which matches a specified regular expression, are replaced.

    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  161. def re_replace_first(target: Column, pattern: String, replacement: String, offset: Column = lit(0)): Column

    Replaces only the first regex matching occurrence in the target string.

    Replaces only the first regex matching occurrence in the target string.

    target

    A string that the function searches for a substring that matches pattern_expr.

    pattern

    regular expression

    replacement

    replacement string

    returns

    a replaced string in which first substring, which matches a specified regular expression, is replaced.

    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  162. lazy val re_split_no_empty: UserDefinedFunction

    UDF to split input string via pattern string and remove all empty subtrings.

    UDF to split input string via pattern string and remove all empty subtrings.

    Definition Classes
    SparkFunctions
  163. lazy val readBytesIntoInteger: UserDefinedFunction
    Definition Classes
    SparkFunctions
  164. lazy val readBytesIntoLong: UserDefinedFunction
    Definition Classes
    SparkFunctions
  165. lazy val readBytesStringIntoInteger: UserDefinedFunction
    Definition Classes
    SparkFunctions
  166. lazy val readBytesStringIntoLong: UserDefinedFunction
    Definition Classes
    SparkFunctions
  167. def readDFSFile(inputPath: String): String
    Definition Classes
    DataHelpers
  168. def readHiveTable(spark: SparkSession, database: String, table: String, partition: String = ""): DataFrame

    Method to read data from hive table.

    Method to read data from hive table.

    spark

    spark session

    database

    hive database

    table

    hive table.

    partition

    hive table partition to read data specifically from if provided.

    returns

    dataframe with data read from Hive Table.

    Definition Classes
    DataHelpers
  169. def readHiveTableInChunks(spark: SparkSession, database: String, table: String, partitionKey: String, partitionValue: String): DataFrame

    Reads a full hive table partition, by reading every subpartition separately and performing a union on all the final DataFrames

    Reads a full hive table partition, by reading every subpartition separately and performing a union on all the final DataFrames

    This function is meant to temporarily solve the problem with Hive metastore crashing when querying too many partitions at the same time.

    spark

    spark session

    database

    hive database name

    table

    hive table name

    partitionKey

    top-level partition's key

    partitionValue

    top-level partition's value

    returns

    A complete DataFrame with the selected hive table partition

    Definition Classes
    DataHelpers
  170. def readInputFile(spark: SparkSession, inputPath: String, ffSchema: Option[FFSchemaRecord], formatType: String = "fixedFormat", delimiter: String = ",", sparkSchema: String = ""): DataFrame
    Definition Classes
    DataHelpers
  171. def readInputFile(spark: SparkSession, inputPath: String, ffSchemaString: String, formatType: String, delimiter: String, sparkSchema: String): DataFrame
    Definition Classes
    DataHelpers
    Annotations
    @Py4JWhitelist()
  172. lazy val read_file: UserDefinedFunction
    Definition Classes
    SparkFunctions
  173. lazy val record_info: UserDefinedFunction
    Definition Classes
    SparkFunctions
  174. lazy val record_info_with_includes: UserDefinedFunction
    Definition Classes
    SparkFunctions
  175. def registerAllUDFs(spark: SparkSession): Unit
    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  176. def registerProphecyUdfs(spark: SparkSession): Unit
    Definition Classes
    UDFUtils
  177. def register_output_schema(portName: String, schema: StructType): Unit
    Definition Classes
    Component
  178. def remove_non_digit(input: Column): Column

    Method removes any non-digit characters from the specified string column.

    Method removes any non-digit characters from the specified string column.

    input

    input String Column

    returns

    Cleaned string column or null

    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  179. def replaceBlankColumnWithNull(input: Column): Column

    Method to replace String Columns with Empty value to Null.

    Method to replace String Columns with Empty value to Null.

    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  180. def replaceNullWithDefaultValues(input: Column, schema: String): Column
    Definition Classes
    SparkFunctions
  181. def replaceString(sparkSession: SparkSession, df: DataFrame, outputCol: String, inputCol: String, replaceWith: String, value: String, values: String*): DataFrame

    Function to add new column in passed dataframe.

    Function to add new column in passed dataframe. Newly added column value is decided by the presence of value corresponding to inputCol in array comprised of value and values. If inputCol is found then value of replaceWith is added in new column otherwise inputCol value is added.

    sparkSession

    spark session.

    df

    input dataframe.

    outputCol

    name of new column to be added.

    inputCol

    column name whose value is searched.

    replaceWith

    value with which to replace searched value if found.

    value

    element to be combined in array column

    values

    all values to be combined in array column for searching purpose.

    returns

    dataframe with new column with column name outputCol

    Definition Classes
    UDFUtils
  182. def replaceStringNull(sparkSession: SparkSession, df: DataFrame, outputCol: String, inputCol: String, replaceWith: String, value: String, values: String*): DataFrame

    Function to add new column in passed dataframe.

    Function to add new column in passed dataframe. Newly added column value is decided by the presence of value corresponding to inputCol in array comprised of value and values and null. If inputCol is found then value of replaceWith is added in new column otherwise inputCol value is added.

    sparkSession

    spark session.

    df

    input dataframe.

    outputCol

    name of new column to be added.

    inputCol

    column name whose value is searched.

    replaceWith

    value with which to replace searched value if found.

    value

    element to be combined in array column

    values

    all values to be combined in array column for searching purpose.

    returns

    dataframe with new column with column name outputCol

    Definition Classes
    UDFUtils
  183. def replaceStringWithNull(sparkSession: SparkSession, df: DataFrame, outputCol: String, inputCol: String, value: String, values: String*): DataFrame

    Function to add new column in passed dataframe.

    Function to add new column in passed dataframe. Newly added column value is decided by the presence of value corresponding to inputCol in array comprised of value and values and null. If inputCol is found then value of null is added in new column otherwise inputCol value is added.

    sparkSession

    spark session.

    df

    input dataframe.

    outputCol

    name of new Column to be added.

    inputCol

    column name whose value is searched.

    value

    element to be combined in array column.

    values

    all values to be combined in array column for searching purpose.

    returns

    dataframe with new column with column name outputCol

    Definition Classes
    UDFUtils
  184. def replace_null_with_blank(input: Column): Column
    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  185. lazy val replace_string: UserDefinedFunction

    UDF to find str in input sequence toBeReplaced and return replace if found.

    UDF to find str in input sequence toBeReplaced and return replace if found. Otherwise str is returned.

    Definition Classes
    UDFUtils
  186. lazy val replace_string_with_null: UserDefinedFunction

    UDF to find str in input sequence toBeReplaced and return null if found.

    UDF to find str in input sequence toBeReplaced and return null if found. Otherwise str is returned.

    Definition Classes
    UDFUtils
  187. def scanf_double(format: Column, value: Column): Column
    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  188. def scanf_long(format: Column, value: Column): Column
    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  189. def schemaRowCompareResult(row1: StructType, row2: StructType): Column
    Definition Classes
    SparkFunctions
  190. def sign_explicit(c: Column): Column

    Adds an explicit sign to the number.

    Adds an explicit sign to the number. E.g. 2 -> +2; -004 -> -004; 0 -> +0

    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  191. lazy val sign_explicit_Udf: UserDefinedFunction
    Definition Classes
    SparkFunctions
  192. def sign_reserved(c: Column): Column
    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  193. lazy val sign_reserved_Udf: UserDefinedFunction
    Definition Classes
    SparkFunctions
  194. def splitFile(spark: SparkSession, inputFileOrDirPath: String, splitParts: Int, rowLength: Int, tmpDir: String, bufferSize: Int = 4096): String
    Definition Classes
    DataHelpers
  195. def splitIntoMultipleColumns(sparkSession: SparkSession, df: DataFrame, colName: String, pattern: String, prefix: String = null): DataFrame

    Function to split column with colName in input dataframe using split pattern into multiple columns.

    Function to split column with colName in input dataframe using split pattern into multiple columns. If prefix name is provided each new generated column is prefixed with prefix followed by column number, otherwise original column name is used.

    sparkSession

    spark session.

    df

    input dataframe.

    colName

    column in dataframe which needs to be split into multiple columns.

    pattern

    regex with which column in input dataframe will be split into multiple columns.

    prefix

    column prefix to be used with all newly generated columns.

    returns

    new dataframe with new columns where new column values are generated after splitting original column colName.

    Definition Classes
    UDFUtils
  196. lazy val splitIntoMultipleColumnsUdf: UserDefinedFunction

    UDF to break input string into multiple string via delimiter.

    UDF to break input string into multiple string via delimiter. Number of strings after split are adjusted as per passed width parameter. If number of strings are less then empty strings are added otherwise in case of more number of strings, first width number of entries are picked and remaining are discarded.

    Definition Classes
    SparkFunctions
  197. def starts_with(input: Column, prefix: String): Column

    Returns true if string columns starts with given prefix

    Returns true if string columns starts with given prefix

    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  198. def string_char(inputStr: Column, index: Int): Column

    Method to return character code of character at index position in inputStr string.

    Method to return character code of character at index position in inputStr string.

    inputStr

    input string

    index

    location of character to get code.

    returns

    integer column.

    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  199. lazy val string_cleanse: UserDefinedFunction

    This implementation is incorrect.

    This implementation is incorrect.

    Definition Classes
    SparkFunctions
  200. def string_compare(input1: Column, input2: Column): Column
    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  201. lazy val string_concat_in_loop: UserDefinedFunction
    Definition Classes
    SparkFunctions
  202. lazy val string_convert_explicit: UserDefinedFunction

    Converts a string from one character set to another, replacing inconvertible characters with a specified string.

    Converts a string from one character set to another, replacing inconvertible characters with a specified string.

    Definition Classes
    SparkFunctions
  203. lazy val string_filter: UserDefinedFunction

    Method which returns string of characters present in both of the strings in the same order as appearing in first string

    Method which returns string of characters present in both of the strings in the same order as appearing in first string

    Definition Classes
    SparkFunctions
  204. lazy val string_filter_out: UserDefinedFunction

    Compares two input strings, then returns characters that appear in one string but not in the other.

    Compares two input strings, then returns characters that appear in one string but not in the other.

    Definition Classes
    SparkFunctions
  205. lazy val string_from_hex: UserDefinedFunction
    Definition Classes
    SparkFunctions
  206. lazy val string_index: UserDefinedFunction

    UDF to find index of seekStr in inputStr.

    UDF to find index of seekStr in inputStr. Returned index will be 1 based index.

    Definition Classes
    SparkFunctions
  207. lazy val string_index_with_offset: UserDefinedFunction

    UDF to find index of seekStr in inputStr from offset index onwards.

    UDF to find index of seekStr in inputStr from offset index onwards. Returned string position is 1 based position.

    Definition Classes
    SparkFunctions
  208. def string_is_alphabetic(input: Column): Column

    Method which returns true if input string contains all alphabetic characters, or false otherwise.

    Method which returns true if input string contains all alphabetic characters, or false otherwise.

    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  209. def string_is_numeric(input: Column): Column

    Method which returns true if input string contains all numeric characters, or false otherwise.

    Method which returns true if input string contains all numeric characters, or false otherwise.

    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  210. def string_join(column: Column, delimiter: String): Column

    Concatenates the elements of column using the delimiter.

    Concatenates the elements of column using the delimiter.

    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  211. def string_length(input: Column): Column
    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  212. lazy val string_like: UserDefinedFunction

    Method to test whether a string matches a specified pattern.

    Method to test whether a string matches a specified pattern. This function returns 1 if the input string matches a specified pattern, and 0 if the string does not match the pattern.

    In abinitio version % character in pattern means to match zero or more characters and _ character means matches a single character.

    Definition Classes
    SparkFunctions
  213. def string_lpad(input: Column, len: Int, pad_char: String = " "): Column

    Left-pad the input string column with pad_char to a length of len.

    Left-pad the input string column with pad_char to a length of len. If length of input column is more than len then returns input column unmodified.

    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  214. def string_lrepad(input: Column, len: Int, char_to_pad_with: String = " "): Column

    function trims the string and then pad the string with given character upto given length.

    function trims the string and then pad the string with given character upto given length. if the length of trimmed string is equal to or greater than given length than it return input string

    input

    input string

    len

    length in number of characters.

    char_to_pad_with

    A character used to pad input string to length len.

    returns

    string of a specified length, trimmed of leading and trailing blanks and left-padded with a given character.

    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  215. def string_pad(input: Column, len: Int, char_to_pad_with: String = " "): Column

    function pads input on the right with the character char_to_pad_with to make the string length len.

    function pads input on the right with the character char_to_pad_with to make the string length len. If str is already len or more characters long, the function returns input unmodified.

    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  216. lazy val string_pad: UserDefinedFunction
    Definition Classes
    SparkFunctions
  217. lazy val string_pad_with_char: UserDefinedFunction
    Definition Classes
    SparkFunctions
  218. def string_prefix(input: Column, length: Column): Column
    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  219. def string_repad(input: Column, len: Int, char_to_pad_with: String = " "): Column

    function trims the string and then pad the string on right side with given character upto given length.

    function trims the string and then pad the string on right side with given character upto given length. if the length of trimmed string is equal to or greater than given length than it return input string

    input

    input string

    len

    length in number of characters.

    char_to_pad_with

    A character used to pad input string to length len.

    returns

    string of a specified length, trimmed of leading and trailing blanks and left-padded with a given character.

    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  220. def string_replace(input: Column, seekStr: Column, newStr: Column, offset: Column = lit(0)): Column

    Function to replace occurrence of seekStr with newStr string in input string after offset characters from first character.

    Function to replace occurrence of seekStr with newStr string in input string after offset characters from first character.

    input

    input string on which to perform replace operation.

    seekStr

    string to be replaced in input string.

    newStr

    string to be used instead of seekStr in input string.

    offset

    number of characters to skip from begining in input string before performing string_replace operation.

    returns

    modified string where seekStr is replaced with newStr in input string.

    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  221. lazy val string_replace_first: UserDefinedFunction
    Definition Classes
    SparkFunctions
  222. lazy val string_replace_first_in_loop: UserDefinedFunction
    Definition Classes
    SparkFunctions
  223. lazy val string_replace_in_loop: UserDefinedFunction
    Definition Classes
    SparkFunctions
  224. lazy val string_representation: UserDefinedFunction
    Definition Classes
    SparkFunctions
  225. lazy val string_rindex: UserDefinedFunction

    Returns the index of the first character of the last occurrence of a seek string within another input string.

    Returns the index of the first character of the last occurrence of a seek string within another input string. Returned index is 1 based.

    Definition Classes
    SparkFunctions
  226. lazy val string_rindex_with_offset: UserDefinedFunction

    UDF to find index of seekStr in inputStr from end of inputStr skipping offset number of characters from end.

    UDF to find index of seekStr in inputStr from end of inputStr skipping offset number of characters from end. Offset index is number of characters, from the end of str, to skip before searching. Returned string position is 1 based position.

    Definition Classes
    SparkFunctions
  227. lazy val string_run_length_split: UserDefinedFunction
    Definition Classes
    SparkFunctions
  228. lazy val string_split: UserDefinedFunction

    UDF to split input string via delimiter string.

    UDF to split input string via delimiter string.

    Definition Classes
    SparkFunctions
  229. lazy val string_split_no_empty: UserDefinedFunction

    UDF to split input string via delimiter string and remove all empty subtrings.

    UDF to split input string via delimiter string and remove all empty subtrings.

    Definition Classes
    SparkFunctions
  230. def string_substring(input: Column, start_position: Column, length: Column = lit(Int.MaxValue)): Column

    Method to find substring of input string.

    Method to find substring of input string.

    input

    string on which to find substring.

    start_position

    1 based starting position to find substring from.

    length

    total length of substring to be found.

    returns

    substring of input string

    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  231. def string_suffix(input: Column, len: Int): Column
    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  232. lazy val string_suffix: UserDefinedFunction
    Definition Classes
    SparkFunctions
  233. lazy val string_to_hex: UserDefinedFunction
    Definition Classes
    SparkFunctions
  234. lazy val take_last_nth: UserDefinedFunction

    UDF to return nth element from last in passed array of elements.

    UDF to return nth element from last in passed array of elements. In case input sequence has less number of elements than n then first element is returned.

    Definition Classes
    UDFUtils
  235. lazy val take_nth: UserDefinedFunction

    UDF to take Nth element from beginning.

    UDF to take Nth element from beginning. In case input sequence has less element than N then exception is thrown.

    Definition Classes
    UDFUtils
  236. lazy val testUDFF: UserDefinedFunction
    Definition Classes
    SparkFunctions
  237. lazy val test_characters_all: UserDefinedFunction

    UDF to identify the number of characters in inputStr which are present in charFlag

    UDF to identify the number of characters in inputStr which are present in charFlag

    Definition Classes
    SparkFunctions
  238. def timezone_to_utc(timezone: String, time: Column): Column

    Method to convert

    Method to convert

    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  239. def toString(): String
    Definition Classes
    Any
  240. def today(): Column

    Method to return integer value representing number of days to today from “1-1-1990”.

    Method to return integer value representing number of days to today from “1-1-1990”.

    returns

    integer value

    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  241. lazy val translate_bytes: UserDefinedFunction

    UDF to return a string in the native character set made up of bytes from the given map.

    UDF to return a string in the native character set made up of bytes from the given map. Each byte of the result is the value of map indexed by the character code of the corresponding byte of the input string str. The function returns NULL if any argument is NULL.

    Definition Classes
    SparkFunctions
  242. lazy val truncateMicroSeconds: UserDefinedFunction

    UDF to truncate microseconds part of timestamp.

    UDF to truncate microseconds part of timestamp. This is needed as abinitio and spark has some incompatibility in microseconds part of timestamp format.

    Definition Classes
    SparkFunctions
  243. lazy val type_info: UserDefinedFunction
    Definition Classes
    SparkFunctions
  244. lazy val type_info_with_includes: UserDefinedFunction
    Definition Classes
    SparkFunctions
  245. def unionAll(df: DataFrame*): DataFrame

    Method to take union of all passed dataframes.

    Method to take union of all passed dataframes.

    df

    list of dataframes for which to take union of.

    returns

    union of all passed input dataframes.

    Definition Classes
    DataHelpers
  246. lazy val unique_identifier: UserDefinedFunction
    Definition Classes
    SparkFunctions
  247. lazy val url_encode_escapes: UserDefinedFunction
    Definition Classes
    SparkFunctions
  248. lazy val vector_avg: UserDefinedFunction
    Definition Classes
    SparkFunctions
  249. lazy val vector_stdev: UserDefinedFunction
    Definition Classes
    SparkFunctions
  250. def warning(errorMsg: String): Column
    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  251. def windowSpec(partitionByExpr: Column = lit(1)): WindowSpec
    Definition Classes
    SparkFunctions
  252. def windowSpecPrevRow(partitionByExpr: Column = lit(1)): WindowSpec
    Definition Classes
    SparkFunctions
  253. def withSparkProperty[T](key: String, value: String, spark: SparkSession)(body: ⇒ T): T
  254. def withSubgraphName[T](value: String, spark: SparkSession)(body: ⇒ T): T
  255. def withTargetId[T](value: String, spark: SparkSession)(body: ⇒ T): T
  256. def writeDataFrame(df: DataFrame, path: String, spark: SparkSession, props: Map[String, String], format: String, partitionColumns: List[String] = Nil, bucketColumns: List[String] = Nil, numBuckets: Option[Int] = None, sortColumns: List[String] = Nil, tableName: Option[String] = None, databaseName: Option[String] = None): Unit

    Method to write data passed in dataframe in specific file format.

    Method to write data passed in dataframe in specific file format.

    df

    dataframe containing data.

    path

    path to write data to.

    spark

    spark session.

    props

    underlying data source specific properties.

    format

    file format in which to persist data. Supported file formats are csv, text, json, parquet, orc

    partitionColumns

    columns to be used for partitioning.

    bucketColumns

    used to bucket the output by the given columns. If specified, the output is laid out on the file-system similar to Hive's bucketing scheme.

    numBuckets

    number of buckets to be used.

    sortColumns

    columns on which to order data while persisting.

    tableName

    table name for persisting data.

    databaseName

    database name for persisting data.

    Definition Classes
    DataHelpers
  257. lazy val writeIntegerToBytes: UserDefinedFunction
    Definition Classes
    SparkFunctions
  258. lazy val writeLongToBytes: UserDefinedFunction
    Definition Classes
    SparkFunctions
  259. lazy val write_to_log: UserDefinedFunction

    UDF to write logging parameters to log port.

    UDF to write logging parameters to log port.

    Definition Classes
    DataHelpers
  260. def xmlStringToJsString(input: String): String
    Definition Classes
    SparkFunctions
  261. lazy val xmlToJSON: UserDefinedFunction
    Definition Classes
    SparkFunctions
  262. def yyyyMMdd_to_YYYYJJJ(in_date: Column): Column

    Converts yyyyyMMdd to YYYYJJJ

    Converts yyyyyMMdd to YYYYJJJ

    in_date

    date in yyyyMMdd format

    returns

    a date converted to YYYYJJJ

    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  263. def zip_eventInfo_arrays(column1: Column, column2: Column): Column

    Method to zip two arrays with first one having event_type and second one having event_text

    Method to zip two arrays with first one having event_type and second one having event_text

    Definition Classes
    SparkFunctions
    Annotations
    @Py4JWhitelist()
  264. object AbinitioDMLs
  265. object CDC

    Column Dependency Calculator for two:

  266. object Component
  267. object ConfigBase extends LazyLogging
  268. object DataFrameValidator
  269. object DataHelpers
  270. object FTPHelpers extends LazyLogging
  271. object FixedFileFormat extends Serializable
  272. object FixedFileFormatImplicits
  273. object FixedFormatHelper
  274. object FixedFormatSchemaImplicits
  275. object ProphecySparkAnalysisException extends Serializable
  276. object RestAPIUtils
  277. object SchemaUtils
  278. object SparkFunctions
  279. object LongSequence
    Definition Classes
    SparkFunctions
  280. object RecordIterator extends Serializable
    Definition Classes
    SparkFunctions
  281. object SparkTestingUtils
  282. object VaultUtils extends LazyLogging
  283. object Version
  284. object SecretManager extends LazyLogging
    Definition Classes
    ProphecySecrets
  285. object SecretValue extends Serializable
    Definition Classes
    ProphecySecrets

Inherited from ProphecySecrets

Inherited from FutureExtensions

Inherited from ExceptionExtensions

Inherited from StringExtensions

Inherited from TraversableExtensions

Inherited from Extension

Inherited from FixedFileFormatImplicits

Inherited from SparkFunctions

Inherited from DataHelpers

Inherited from LazyLogging

Inherited from Component

Inherited from UDFUtils

Inherited from Serializable

Inherited from Serializable

Inherited from RestAPIUtils

Inherited from ProphecyDataFrame

Inherited from AnyRef

Inherited from Any

Ungrouped