Object

org.apache.spark.sql.collection

Utils

Related Doc: package collection

Permalink

object Utils extends Logging

Linear Supertypes
Logging, AnyRef, Any
Ordering
  1. Alphabetic
  2. By Inheritance
Inherited
  1. Utils
  2. Logging
  3. AnyRef
  4. Any
  1. Hide All
  2. Show All
Visibility
  1. Public
  2. All

Value Members

  1. final def !=(arg0: Any): Boolean

    Permalink
    Definition Classes
    AnyRef → Any
  2. final def ##(): Int

    Permalink
    Definition Classes
    AnyRef → Any
  3. final def ==(arg0: Any): Boolean

    Permalink
    Definition Classes
    AnyRef → Any
  4. final val EMPTY_STRING_ARRAY: Array[String]

    Permalink
  5. def ERROR_NO_QCS(module: String): String

    Permalink
  6. final val SKIP_ANALYSIS_PREFIX: String("sample_")

    Permalink
  7. final val WEIGHTAGE_COLUMN_NAME: String("snappy_sampler_weightage")

    Permalink
  8. final val Z95Percent: Double

    Permalink
  9. final val Z95Squared: Double

    Permalink
  10. def analysisException(msg: String, cause: Option[Throwable] = None): AnalysisException

    Permalink
  11. final def asInstanceOf[T0]: T0

    Permalink
    Definition Classes
    Any
  12. def charMetadata(size: Int, md: Metadata): Metadata

    Permalink

    Utility function to return a metadata for a StructField of StringType, to ensure that the field is stored (and rendered) as CHAR by SnappyStore.

    Utility function to return a metadata for a StructField of StringType, to ensure that the field is stored (and rendered) as CHAR by SnappyStore.

    size

    the size parameter of the CHAR() column type

    md

    optional Metadata object to be merged into the result

    returns

    the result Metadata object to use for StructField

  13. def charMetadata(size: Int): Metadata

    Permalink

    Utility function to return a metadata for a StructField of StringType, to ensure that the field is stored (and rendered) as CHAR by SnappyStore.

    Utility function to return a metadata for a StructField of StringType, to ensure that the field is stored (and rendered) as CHAR by SnappyStore.

    size

    the size parameter of the CHAR() column type

    returns

    the result Metadata object to use for StructField

  14. def charMetadata(): Metadata

    Permalink

    Utility function to return a metadata for a StructField of StringType, to ensure that the field is stored (and rendered) as CHAR by SnappyStore.

    Utility function to return a metadata for a StructField of StringType, to ensure that the field is stored (and rendered) as CHAR by SnappyStore.

    returns

    the result Metadata object to use for StructField

  15. def classForName(className: String): Class[_]

    Permalink
  16. def clearDefaultSerializerAndCodec(): Unit

    Permalink
  17. def clone(): AnyRef

    Permalink
    Attributes
    protected[java.lang]
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  18. def closeJsonGenerator(gen: AnyRef): Unit

    Permalink
  19. def columnIndex(col: String, cols: Array[String], module: String): Int

    Permalink
  20. def compactMap[K, V](map: Map[K, V]): Object2ObjectOpenHashMap[K, V]

    Permalink
  21. def createCatalystConverter(dataType: DataType): (Any) ⇒ Any

    Permalink
  22. def createScalaConverter(dataType: DataType): (Any) ⇒ Any

    Permalink
  23. final def eq(arg0: AnyRef): Boolean

    Permalink
    Definition Classes
    AnyRef
  24. def equals(arg0: Any): Boolean

    Permalink
    Definition Classes
    AnyRef → Any
  25. def executorsListener(sc: SparkContext): Option[ExecutorsListener]

    Permalink
  26. def fieldIndex(relationOutput: Seq[Attribute], columnName: String, caseSensitive: Boolean): Int

    Permalink
  27. def fillArray[T](a: Array[_ >: T], v: T, start: Int, endP1: Int): Unit

    Permalink
  28. def finalize(): Unit

    Permalink
    Attributes
    protected[java.lang]
    Definition Classes
    AnyRef
    Annotations
    @throws( classOf[java.lang.Throwable] )
  29. def genTaskContextFunction(ctx: CodegenContext): String

    Permalink
  30. def generateJson(gen: AnyRef, row: InternalRow, columnIndex: Int, columnType: DataType): Unit

    Permalink
  31. def getActiveSession: Option[SparkSession]

    Permalink
  32. def getAllExecutorsMemoryStatus(sc: SparkContext): Map[BlockManagerId, (Long, Long)]

    Permalink
  33. final def getClass(): Class[_]

    Permalink
    Definition Classes
    AnyRef → Any
  34. def getClientHostPort(netServer: String): String

    Permalink
  35. def getDriverClassName(url: String): String

    Permalink
  36. def getFields(o: Any): Map[String, Any]

    Permalink
  37. def getFixedPartitionRDD[T](sc: SparkContext, f: (TaskContext, Partition) ⇒ Iterator[T], partitioner: Partitioner, numPartitions: Int)(implicit arg0: ClassTag[T]): RDD[T]

    Permalink
  38. def getGenericRowValues(row: GenericRow): Array[Any]

    Permalink
  39. def getHostExecutorId(blockId: BlockManagerId): String

    Permalink
  40. def getInternalSparkConf(sc: SparkContext): SparkConf

    Permalink
  41. def getInternalType(dataType: DataType): Class[_]

    Permalink
  42. def getJsonGenerator(dataType: DataType, columnName: String, writer: Writer): AnyRef

    Permalink
  43. def getLocalProperties(sc: SparkContext): Properties

    Permalink
  44. def getNumColumns(partitioning: Partitioning): Int

    Permalink
  45. def getPartitionData(blockId: BlockId, bm: BlockManager): ByteBuffer

    Permalink
  46. def getPartitions(region: Any, bucketId: Int): Array[Partition]

    Permalink
  47. def getPrunedPartition(partitionColumns: Seq[String], filters: Array[Expression], schema: StructType, numBuckets: Int, partitionColumnCount: Int): Int

    Permalink
  48. def getSQLDataType(dataType: DataType): DataType

    Permalink
  49. def getSchemaAndPlanFromBase(schemaOpt: Option[StructType], baseTableOpt: Option[String], catalog: SnappySessionCatalog, asSelect: Boolean, table: String, tableType: String): (StructType, Option[LogicalPlan])

    Permalink

    Get the result schema given an optional explicit schema and base table.

    Get the result schema given an optional explicit schema and base table. In case both are specified, then check compatibility between the two.

  50. def getTaskContext: TaskContext

    Permalink

    Get either the normal TaskContext or the one created by withTempTaskContextIfAbsent.

  51. def getTempTaskContext: TaskContext

    Permalink

    Get the TaskContext when created by withTempTaskContextIfAbsent.

  52. def hashCode(): Int

    Permalink
    Definition Classes
    AnyRef → Any
  53. def immutableMap[A, B](m: Map[A, B]): Map[A, B]

    Permalink
  54. def initializeLogIfNecessary(): Unit

    Permalink
    Attributes
    protected
    Definition Classes
    Logging
  55. final def isDebugEnabled: Boolean

    Permalink
    Definition Classes
    Logging
  56. final def isInfoEnabled: Boolean

    Permalink
    Definition Classes
    Logging
  57. final def isInstanceOf[T0]: Boolean

    Permalink
    Definition Classes
    Any
  58. final def isLoner(sc: SparkContext): Boolean

    Permalink
  59. def isSmartConnectorMode(sc: SparkContext): Boolean

    Permalink
  60. final def isTraceEnabled: Boolean

    Permalink
    Definition Classes
    Logging
  61. final var levelFlags: Int

    Permalink
    Attributes
    protected
    Definition Classes
    Logging
  62. def log: Logger

    Permalink
    Attributes
    protected
    Definition Classes
    Logging
  63. def logDebug(msg: ⇒ String, throwable: Throwable): Unit

    Permalink
    Definition Classes
    Logging
  64. def logDebug(msg: ⇒ String): Unit

    Permalink
    Definition Classes
    Logging
  65. def logError(msg: ⇒ String, throwable: Throwable): Unit

    Permalink
    Definition Classes
    Logging
  66. def logError(msg: ⇒ String): Unit

    Permalink
    Definition Classes
    Logging
  67. def logInfo(msg: ⇒ String, throwable: Throwable): Unit

    Permalink
    Definition Classes
    Logging
  68. def logInfo(msg: ⇒ String): Unit

    Permalink
    Definition Classes
    Logging
  69. def logName: String

    Permalink
    Attributes
    protected
    Definition Classes
    Logging
  70. def logTrace(msg: ⇒ String, throwable: Throwable): Unit

    Permalink
    Definition Classes
    Logging
  71. def logTrace(msg: ⇒ String): Unit

    Permalink
    Definition Classes
    Logging
  72. def logWarning(msg: ⇒ String, throwable: Throwable): Unit

    Permalink
    Definition Classes
    Logging
  73. def logWarning(msg: ⇒ String): Unit

    Permalink
    Definition Classes
    Logging
  74. final var log_: Logger

    Permalink
    Attributes
    protected
    Definition Classes
    Logging
  75. def mapExecutors[T](sc: SparkContext, f: (TaskContext, ExecutorLocalPartition) ⇒ Iterator[T], maxTries: Int)(implicit arg0: ClassTag[T]): Array[T]

    Permalink
  76. def mapExecutors[T](sc: SparkContext, f: () ⇒ Iterator[T], maxTries: Int = 30, blockManagerIds: Seq[BlockManagerId] = Nil)(implicit arg0: ClassTag[T]): Array[T]

    Permalink
  77. def matchOption(optName: String, options: Map[String, Any]): Option[(String, Any)]

    Permalink
  78. def metricMethods: ((String) ⇒ String, (String) ⇒ String)

    Permalink
  79. def millisToDays(millisUtc: Long, tz: TimeZone): Int

    Permalink
  80. final def ne(arg0: AnyRef): Boolean

    Permalink
    Definition Classes
    AnyRef
  81. def newChunkedByteBuffer(chunks: Array[ByteBuffer]): ChunkedByteBuffer

    Permalink
  82. def newClusterSparkConf(props: Map[AnyRef, AnyRef]): SparkConf

    Permalink
  83. def newClusterSparkConf(): SparkConf

    Permalink
  84. final def notify(): Unit

    Permalink
    Definition Classes
    AnyRef
  85. final def notifyAll(): Unit

    Permalink
    Definition Classes
    AnyRef
  86. def parseCSVList(s: String, parser: SnappyParser): Seq[String]

    Permalink
  87. def parseColumn(cv: Any, cols: Array[String], module: String, option: String): Int

    Permalink
  88. def parseColumnsAsClob(s: String, session: SnappySession): (Boolean, Set[String])

    Permalink
  89. def parseDouble(v: Any, module: String, option: String, min: Double, max: Double, exclusive: Boolean = true): Double

    Permalink
  90. def parseInteger(v: Any, module: String, option: String, min: Int = 1, max: Int = Int.MaxValue): Int

    Permalink
  91. def parseTimeInterval(optV: Any, module: String): Long

    Permalink

    Parse the given time interval value as long milliseconds.

    Parse the given time interval value as long milliseconds.

    See also

    timeIntervalSpec for the allowed string specification

  92. def parseTimestamp(ts: String, module: String, col: String): Long

    Permalink
  93. def qcsOf(qa: Array[String], cols: Array[String], module: String): (Array[Int], Array[String])

    Permalink
  94. def registerDriver(driver: String): Unit

    Permalink

    Register given driver class with Spark's loader.

  95. def registerDriverUrl(url: String): String

    Permalink

    Register driver for given JDBC URL and return the driver class name.

  96. def resetLogger(): Unit

    Permalink
    Attributes
    protected
    Definition Classes
    Logging
  97. def resolveQCS(options: Map[String, Any], fieldNames: Array[String], module: String): (Array[Int], Array[String])

    Permalink
  98. def resolveQCS(qcsV: Option[Any], fieldNames: Array[String], module: String): (Array[Int], Array[String])

    Permalink
  99. def resultSetToSparkInternalRows(resultSet: ResultSet, schema: StructType, inputMetrics: InputMetrics = new InputMetrics): Iterator[InternalRow]

    Permalink
  100. def schemaFields(schema: StructType): Map[String, StructField]

    Permalink
  101. def setCurrentSchema(session: SnappySession, schema: String, createIfNotExists: Boolean): Unit

    Permalink
  102. def setDefaultConfProperty(conf: SparkConf, name: String, default: String): Unit

    Permalink
  103. def setDefaultSerializerAndCodec(conf: SparkConf): Unit

    Permalink
  104. def sqlInternal(snappy: SnappySession, sqlText: String): CachedDataFrame

    Permalink
  105. def stringMetadata(md: Metadata = Metadata.empty): Metadata

    Permalink

    Utility function to return a metadata for a StructField of StringType, to ensure that the field is rendered as CLOB by SnappyStore.

    Utility function to return a metadata for a StructField of StringType, to ensure that the field is rendered as CLOB by SnappyStore.

    md

    optional Metadata object to be merged into the result

    returns

    the result Metadata object to use for StructField

  106. final def synchronized[T0](arg0: ⇒ T0): T0

    Permalink
    Definition Classes
    AnyRef
  107. def toLowerCase(k: String): String

    Permalink
  108. def toString(): String

    Permalink
    Definition Classes
    AnyRef → Any
  109. def toUpperCase(k: String): String

    Permalink
  110. def unAlias(e: Expression, childClass: Class[_] = null): Expression

    Permalink
    Annotations
    @tailrec()
  111. lazy val usingEnhancedSpark: Boolean

    Permalink
  112. def varcharMetadata(size: Int, md: Metadata): Metadata

    Permalink

    Utility function to return a metadata for a StructField of StringType, to ensure that the field is stored (and rendered) as VARCHAR by SnappyStore.

    Utility function to return a metadata for a StructField of StringType, to ensure that the field is stored (and rendered) as VARCHAR by SnappyStore.

    size

    the size parameter of the VARCHAR() column type

    md

    optional Metadata object to be merged into the result

    returns

    the result Metadata object to use for StructField

  113. def varcharMetadata(size: Int): Metadata

    Permalink

    Utility function to return a metadata for a StructField of StringType, to ensure that the field is stored (and rendered) as VARCHAR by SnappyStore.

    Utility function to return a metadata for a StructField of StringType, to ensure that the field is stored (and rendered) as VARCHAR by SnappyStore.

    size

    the size parameter of the VARCHAR() column type

    returns

    the result Metadata object to use for StructField

  114. def varcharMetadata(): Metadata

    Permalink

    Utility function to return a metadata for a StructField of StringType, to ensure that the field is stored (and rendered) as VARCHAR by SnappyStore.

    Utility function to return a metadata for a StructField of StringType, to ensure that the field is stored (and rendered) as VARCHAR by SnappyStore.

    returns

    the result Metadata object to use for StructField

  115. final def wait(): Unit

    Permalink
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  116. final def wait(arg0: Long, arg1: Int): Unit

    Permalink
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  117. final def wait(arg0: Long): Unit

    Permalink
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  118. def withTempTaskContextIfAbsent[T](f: ⇒ T): T

    Permalink

    Run a plan with an empty TaskContext which can be obtained using getTaskContext if it is in use.

    Run a plan with an empty TaskContext which can be obtained using getTaskContext if it is in use. Currently only SnapshotConnectionListener makes use of it to commit/rollback/close connection at the end but can be used by others too. Note that this cannot be directly put inside TaskContext since it has no TaskMemoryManager among other missing things (a valid taskId, partitionId, metrics) so callers should use this only if they do not depend on any of those.

  119. def withThreadLocalTransactionForBucket[T](bucketId: Int, pr: PartitionedRegion, f: (LanguageConnectionContext) ⇒ T): T

    Permalink

Inherited from Logging

Inherited from AnyRef

Inherited from Any

Ungrouped