object
SparkWorkflowUtils extends Logging
Value Members
-
final
def
!=(arg0: AnyRef): Boolean
-
final
def
!=(arg0: Any): Boolean
-
final
def
##(): Int
-
final
def
==(arg0: AnyRef): Boolean
-
final
def
==(arg0: Any): Boolean
-
final
def
asInstanceOf[T0]: T0
-
def
clone(): AnyRef
-
def
debug(msg: ⇒ Any, t: ⇒ Throwable): Unit
-
def
debug(msg: ⇒ Any): Unit
-
final
def
eq(arg0: AnyRef): Boolean
-
def
equals(arg0: Any): Boolean
-
def
error(msg: ⇒ Any, t: ⇒ Throwable): Unit
-
def
error(msg: ⇒ Any): Unit
-
def
finalize(): Unit
-
final
def
getClass(): Class[_]
-
def
getPersistentModel[AP <: Params, M](pmm: PersistentModelManifest, runId: String, params: AP, sc: Option[SparkContext], cl: ClassLoader): M
-
def
hashCode(): Int
-
def
info(msg: ⇒ Any, t: ⇒ Throwable): Unit
-
def
info(msg: ⇒ Any): Unit
-
def
isDebugEnabled: Boolean
-
def
isErrorEnabled: Boolean
-
def
isInfoEnabled: Boolean
-
final
def
isInstanceOf[T0]: Boolean
-
def
isTraceEnabled: Boolean
-
def
isWarnEnabled: Boolean
-
def
logger: Logger
-
def
loggerName: String
-
final
def
ne(arg0: AnyRef): Boolean
-
final
def
notify(): Unit
-
final
def
notifyAll(): Unit
-
final
def
synchronized[T0](arg0: ⇒ T0): T0
-
def
toString(): String
-
def
trace(msg: ⇒ Any, t: ⇒ Throwable): Unit
-
def
trace(msg: ⇒ Any): Unit
-
final
def
wait(): Unit
-
final
def
wait(arg0: Long, arg1: Int): Unit
-
final
def
wait(arg0: Long): Unit
-
def
warn(msg: ⇒ Any, t: ⇒ Throwable): Unit
-
def
warn(msg: ⇒ Any): Unit
Inherited from Logging
Inherited from AnyRef
Inherited from Any
Collection of reusable workflow related utilities that touch on Apache Spark. They are separated to avoid compilation problems with certain code.