org.apache.hadoop.mapred

SparkHadoopWriter

class SparkHadoopWriter extends Logging with SparkHadoopMapRedUtil with Serializable

Internal helper class that saves an RDD using a Hadoop OutputFormat. This is only public because we need to access this class from the spark package to use some package-private Hadoop functions, but this class should not be used directly by users.

Saves the RDD using a JobConf, which should contain an output key class, an output value class, a filename to write to, etc, exactly like in a Hadoop MapReduce job.

Linear Supertypes
Serializable, Serializable, SparkHadoopMapRedUtil, Logging, AnyRef, Any
Ordering
  1. Alphabetic
  2. By inheritance
Inherited
  1. Hide All
  2. Show all
  1. SparkHadoopWriter
  2. Serializable
  3. Serializable
  4. SparkHadoopMapRedUtil
  5. Logging
  6. AnyRef
  7. Any
Visibility
  1. Public
  2. All

Instance Constructors

  1. new SparkHadoopWriter(jobConf: JobConf)

Value Members

  1. final def !=(arg0: AnyRef): Boolean

    Definition Classes
    AnyRef
  2. final def !=(arg0: Any): Boolean

    Definition Classes
    Any
  3. final def ##(): Int

    Definition Classes
    AnyRef → Any
  4. final def ==(arg0: AnyRef): Boolean

    Definition Classes
    AnyRef
  5. final def ==(arg0: Any): Boolean

    Definition Classes
    Any
  6. final def asInstanceOf[T0]: T0

    Definition Classes
    Any
  7. def cleanup(): Unit

  8. def clone(): AnyRef

    Attributes
    protected[lang]
    Definition Classes
    AnyRef
    Annotations
    @throws()
  9. def close(): Unit

  10. def commit(): Unit

  11. def commitJob(): Unit

  12. final def eq(arg0: AnyRef): Boolean

    Definition Classes
    AnyRef
  13. def equals(arg0: Any): Boolean

    Definition Classes
    AnyRef → Any
  14. def finalize(): Unit

    Attributes
    protected[lang]
    Definition Classes
    AnyRef
    Annotations
    @throws()
  15. final def getClass(): java.lang.Class[_]

    Definition Classes
    AnyRef → Any
  16. def hashCode(): Int

    Definition Classes
    AnyRef → Any
  17. def initLogging(): Unit

    Attributes
    protected
    Definition Classes
    Logging
  18. final def isInstanceOf[T0]: Boolean

    Definition Classes
    Any
  19. def isTraceEnabled(): Boolean

    Attributes
    protected
    Definition Classes
    Logging
  20. def log: Logger

    Attributes
    protected
    Definition Classes
    Logging
  21. def logDebug(msg: ⇒ String, throwable: Throwable): Unit

    Attributes
    protected
    Definition Classes
    Logging
  22. def logDebug(msg: ⇒ String): Unit

    Attributes
    protected
    Definition Classes
    Logging
  23. def logError(msg: ⇒ String, throwable: Throwable): Unit

    Attributes
    protected
    Definition Classes
    Logging
  24. def logError(msg: ⇒ String): Unit

    Attributes
    protected
    Definition Classes
    Logging
  25. def logInfo(msg: ⇒ String, throwable: Throwable): Unit

    Attributes
    protected
    Definition Classes
    Logging
  26. def logInfo(msg: ⇒ String): Unit

    Attributes
    protected
    Definition Classes
    Logging
  27. def logTrace(msg: ⇒ String, throwable: Throwable): Unit

    Attributes
    protected
    Definition Classes
    Logging
  28. def logTrace(msg: ⇒ String): Unit

    Attributes
    protected
    Definition Classes
    Logging
  29. def logWarning(msg: ⇒ String, throwable: Throwable): Unit

    Attributes
    protected
    Definition Classes
    Logging
  30. def logWarning(msg: ⇒ String): Unit

    Attributes
    protected
    Definition Classes
    Logging
  31. final def ne(arg0: AnyRef): Boolean

    Definition Classes
    AnyRef
  32. def newJobContext(conf: JobConf, jobId: JobID): JobContext

    Definition Classes
    SparkHadoopMapRedUtil
  33. def newTaskAttemptContext(conf: JobConf, attemptId: TaskAttemptID): TaskAttemptContext

    Definition Classes
    SparkHadoopMapRedUtil
  34. def newTaskAttemptID(jtIdentifier: String, jobId: Int, isMap: Boolean, taskId: Int, attemptId: Int): TaskAttemptID

    Definition Classes
    SparkHadoopMapRedUtil
  35. final def notify(): Unit

    Definition Classes
    AnyRef
  36. final def notifyAll(): Unit

    Definition Classes
    AnyRef
  37. def open(): Unit

  38. def preSetup(): Unit

  39. def setup(jobid: Int, splitid: Int, attemptid: Int): Unit

  40. final def synchronized[T0](arg0: ⇒ T0): T0

    Definition Classes
    AnyRef
  41. def toString(): String

    Definition Classes
    AnyRef → Any
  42. final def wait(): Unit

    Definition Classes
    AnyRef
    Annotations
    @throws()
  43. final def wait(arg0: Long, arg1: Int): Unit

    Definition Classes
    AnyRef
    Annotations
    @throws()
  44. final def wait(arg0: Long): Unit

    Definition Classes
    AnyRef
    Annotations
    @throws()
  45. def write(key: AnyRef, value: AnyRef): Unit

Inherited from Serializable

Inherited from Serializable

Inherited from SparkHadoopMapRedUtil

Inherited from Logging

Inherited from AnyRef

Inherited from Any