Class/Object

io.smartdatalake.app

GlobalConfig

Related Docs: object GlobalConfig | package app

Permalink

case class GlobalConfig(kryoClasses: Option[Seq[String]] = None, sparkOptions: Option[Map[String, String]] = None, enableHive: Boolean = true, memoryLogTimer: Option[MemoryLogTimerConfig] = None, shutdownHookLogger: Boolean = false, stateListeners: Seq[StateListenerConfig] = Seq(), sparkUDFs: Option[Map[String, SparkUDFCreatorConfig]] = None, pythonUDFs: Option[Map[String, PythonUDFCreatorConfig]] = None, secretProviders: Option[Map[String, SecretProviderConfig]] = None, allowOverwriteAllPartitionsWithoutPartitionValues: Seq[DataObjectId] = Seq(), runtimeDataNumberOfExecutionsToKeep: Int = 10, synchronousStreamingTriggerIntervalSec: Int = 60) extends SmartDataLakeLogger with Product with Serializable

Global configuration options

kryoClasses

classes to register for spark kryo serialization

sparkOptions

spark options

enableHive

enable hive for spark session

memoryLogTimer

enable periodic memory usage logging, see detailed configuration MemoryLogTimerConfig

shutdownHookLogger

enable shutdown hook logger to trace shutdown cause

stateListeners

Define state listeners to be registered for receiving events of the execution of SmartDataLake job

sparkUDFs

Define UDFs to be registered in spark session. The registered UDFs are available in Spark SQL transformations and expression evaluation, e.g. configuration of ExecutionModes.

pythonUDFs

Define UDFs in python to be registered in spark session. The registered UDFs are available in Spark SQL transformations but not for expression evaluation.

secretProviders

Define SecretProvider's to be registered.

allowOverwriteAllPartitionsWithoutPartitionValues

Configure a list of exceptions for partitioned DataObject id's, which are allowed to overwrite the all partitions of a table if no partition values are set. This is used to override/avoid a protective error when using SDLSaveMode.OverwriteOptimized|OverwritePreserveDirectories. Define it as a list of DataObject id's.

runtimeDataNumberOfExecutionsToKeep

Number of Executions to keep runtime data for in streaming mode (default = 10). Must be bigger than 1.

synchronousStreamingTriggerIntervalSec

Trigger interval for synchronous actions in streaming mode in seconds (default = 60 seconds) The synchronous actions of the DAG will be executed with this interval if possile. Note that for asynchronous actions there are separate settings, e.g. SparkStreamingMode.triggerInterval.

Linear Supertypes
Serializable, Serializable, Product, Equals, SmartDataLakeLogger, AnyRef, Any
Ordering
  1. Alphabetic
  2. By Inheritance
Inherited
  1. GlobalConfig
  2. Serializable
  3. Serializable
  4. Product
  5. Equals
  6. SmartDataLakeLogger
  7. AnyRef
  8. Any
  1. Hide All
  2. Show All
Visibility
  1. Public
  2. All

Instance Constructors

  1. new GlobalConfig(kryoClasses: Option[Seq[String]] = None, sparkOptions: Option[Map[String, String]] = None, enableHive: Boolean = true, memoryLogTimer: Option[MemoryLogTimerConfig] = None, shutdownHookLogger: Boolean = false, stateListeners: Seq[StateListenerConfig] = Seq(), sparkUDFs: Option[Map[String, SparkUDFCreatorConfig]] = None, pythonUDFs: Option[Map[String, PythonUDFCreatorConfig]] = None, secretProviders: Option[Map[String, SecretProviderConfig]] = None, allowOverwriteAllPartitionsWithoutPartitionValues: Seq[DataObjectId] = Seq(), runtimeDataNumberOfExecutionsToKeep: Int = 10, synchronousStreamingTriggerIntervalSec: Int = 60)

    Permalink

    kryoClasses

    classes to register for spark kryo serialization

    sparkOptions

    spark options

    enableHive

    enable hive for spark session

    memoryLogTimer

    enable periodic memory usage logging, see detailed configuration MemoryLogTimerConfig

    shutdownHookLogger

    enable shutdown hook logger to trace shutdown cause

    stateListeners

    Define state listeners to be registered for receiving events of the execution of SmartDataLake job

    sparkUDFs

    Define UDFs to be registered in spark session. The registered UDFs are available in Spark SQL transformations and expression evaluation, e.g. configuration of ExecutionModes.

    pythonUDFs

    Define UDFs in python to be registered in spark session. The registered UDFs are available in Spark SQL transformations but not for expression evaluation.

    secretProviders

    Define SecretProvider's to be registered.

    allowOverwriteAllPartitionsWithoutPartitionValues

    Configure a list of exceptions for partitioned DataObject id's, which are allowed to overwrite the all partitions of a table if no partition values are set. This is used to override/avoid a protective error when using SDLSaveMode.OverwriteOptimized|OverwritePreserveDirectories. Define it as a list of DataObject id's.

    runtimeDataNumberOfExecutionsToKeep

    Number of Executions to keep runtime data for in streaming mode (default = 10). Must be bigger than 1.

    synchronousStreamingTriggerIntervalSec

    Trigger interval for synchronous actions in streaming mode in seconds (default = 60 seconds) The synchronous actions of the DAG will be executed with this interval if possile. Note that for asynchronous actions there are separate settings, e.g. SparkStreamingMode.triggerInterval.

Value Members

  1. final def !=(arg0: Any): Boolean

    Permalink
    Definition Classes
    AnyRef → Any
  2. final def ##(): Int

    Permalink
    Definition Classes
    AnyRef → Any
  3. final def ==(arg0: Any): Boolean

    Permalink
    Definition Classes
    AnyRef → Any
  4. val allowOverwriteAllPartitionsWithoutPartitionValues: Seq[DataObjectId]

    Permalink

    Configure a list of exceptions for partitioned DataObject id's, which are allowed to overwrite the all partitions of a table if no partition values are set.

    Configure a list of exceptions for partitioned DataObject id's, which are allowed to overwrite the all partitions of a table if no partition values are set. This is used to override/avoid a protective error when using SDLSaveMode.OverwriteOptimized|OverwritePreserveDirectories. Define it as a list of DataObject id's.

  5. final def asInstanceOf[T0]: T0

    Permalink
    Definition Classes
    Any
  6. def clone(): AnyRef

    Permalink
    Attributes
    protected[java.lang]
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  7. def createSparkSession(appName: String, master: Option[String], deployMode: Option[String] = None): SparkSession

    Permalink

    Create a spark session using settings from this global config

  8. val enableHive: Boolean

    Permalink

    enable hive for spark session

  9. final def eq(arg0: AnyRef): Boolean

    Permalink
    Definition Classes
    AnyRef
  10. def finalize(): Unit

    Permalink
    Attributes
    protected[java.lang]
    Definition Classes
    AnyRef
    Annotations
    @throws( classOf[java.lang.Throwable] )
  11. final def getClass(): Class[_]

    Permalink
    Definition Classes
    AnyRef → Any
  12. final def isInstanceOf[T0]: Boolean

    Permalink
    Definition Classes
    Any
  13. val kryoClasses: Option[Seq[String]]

    Permalink

    classes to register for spark kryo serialization

  14. lazy val logger: Logger

    Permalink
    Attributes
    protected
    Definition Classes
    SmartDataLakeLogger
  15. val memoryLogTimer: Option[MemoryLogTimerConfig]

    Permalink

    enable periodic memory usage logging, see detailed configuration MemoryLogTimerConfig

  16. final def ne(arg0: AnyRef): Boolean

    Permalink
    Definition Classes
    AnyRef
  17. final def notify(): Unit

    Permalink
    Definition Classes
    AnyRef
  18. final def notifyAll(): Unit

    Permalink
    Definition Classes
    AnyRef
  19. val pythonUDFs: Option[Map[String, PythonUDFCreatorConfig]]

    Permalink

    Define UDFs in python to be registered in spark session.

    Define UDFs in python to be registered in spark session. The registered UDFs are available in Spark SQL transformations but not for expression evaluation.

  20. val runtimeDataNumberOfExecutionsToKeep: Int

    Permalink

    Number of Executions to keep runtime data for in streaming mode (default = 10).

    Number of Executions to keep runtime data for in streaming mode (default = 10). Must be bigger than 1.

  21. val secretProviders: Option[Map[String, SecretProviderConfig]]

    Permalink

    Define SecretProvider's to be registered.

  22. val shutdownHookLogger: Boolean

    Permalink

    enable shutdown hook logger to trace shutdown cause

  23. val sparkOptions: Option[Map[String, String]]

    Permalink

    spark options

  24. val sparkUDFs: Option[Map[String, SparkUDFCreatorConfig]]

    Permalink

    Define UDFs to be registered in spark session.

    Define UDFs to be registered in spark session. The registered UDFs are available in Spark SQL transformations and expression evaluation, e.g. configuration of ExecutionModes.

  25. val stateListeners: Seq[StateListenerConfig]

    Permalink

    Define state listeners to be registered for receiving events of the execution of SmartDataLake job

  26. final def synchronized[T0](arg0: ⇒ T0): T0

    Permalink
    Definition Classes
    AnyRef
  27. val synchronousStreamingTriggerIntervalSec: Int

    Permalink

    Trigger interval for synchronous actions in streaming mode in seconds (default = 60 seconds) The synchronous actions of the DAG will be executed with this interval if possile.

    Trigger interval for synchronous actions in streaming mode in seconds (default = 60 seconds) The synchronous actions of the DAG will be executed with this interval if possile. Note that for asynchronous actions there are separate settings, e.g. SparkStreamingMode.triggerInterval.

  28. final def wait(): Unit

    Permalink
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  29. final def wait(arg0: Long, arg1: Int): Unit

    Permalink
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  30. final def wait(arg0: Long): Unit

    Permalink
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )

Inherited from Serializable

Inherited from Serializable

Inherited from Product

Inherited from Equals

Inherited from SmartDataLakeLogger

Inherited from AnyRef

Inherited from Any

Ungrouped