Packages

trait V1Scan extends Scan

A trait that should be implemented by V1 DataSources that would like to leverage the DataSource V2 read code paths.

This interface is designed to provide Spark DataSources time to migrate to DataSource V2 and will be removed in a future Spark release.

Annotations
@Unstable()
Source
V1Scan.java
Since

3.0.0

Linear Supertypes
Scan, AnyRef, Any
Ordering
  1. Alphabetic
  2. By Inheritance
Inherited
  1. V1Scan
  2. Scan
  3. AnyRef
  4. Any
  1. Hide All
  2. Show All
Visibility
  1. Public
  2. Protected

Abstract Value Members

  1. abstract def readSchema(): StructType

    Returns the actual schema of this data source scan, which may be different from the physical schema of the underlying storage, as column pruning or other optimizations may happen.

    Returns the actual schema of this data source scan, which may be different from the physical schema of the underlying storage, as column pruning or other optimizations may happen.

    Definition Classes
    Scan
  2. abstract def toV1TableScan[T <: BaseRelation with TableScan](context: SQLContext): T

    Create an BaseRelation with TableScan that can scan data from DataSource v1 to RDD[Row].

    Create an BaseRelation with TableScan that can scan data from DataSource v1 to RDD[Row].

    Since

    3.0.0

Concrete Value Members

  1. final def !=(arg0: Any): Boolean
    Definition Classes
    AnyRef → Any
  2. final def ##: Int
    Definition Classes
    AnyRef → Any
  3. final def ==(arg0: Any): Boolean
    Definition Classes
    AnyRef → Any
  4. final def asInstanceOf[T0]: T0
    Definition Classes
    Any
  5. def clone(): AnyRef
    Attributes
    protected[lang]
    Definition Classes
    AnyRef
    Annotations
    @throws(classOf[java.lang.CloneNotSupportedException]) @IntrinsicCandidate() @native()
  6. def columnarSupportMode(): ColumnarSupportMode

    Subclasses can implement this method to indicate if the support for columnar data should be determined by each partition or is set as a default for the whole scan.

    Subclasses can implement this method to indicate if the support for columnar data should be determined by each partition or is set as a default for the whole scan.

    Definition Classes
    Scan
    Since

    3.5.0

  7. def description(): String

    A description string of this scan, which may includes information like: what filters are configured for this scan, what's the value of some important options like path, etc.

    A description string of this scan, which may includes information like: what filters are configured for this scan, what's the value of some important options like path, etc. The description doesn't need to include #readSchema(), as Spark already knows it.

    By default this returns the class name of the implementation. Please override it to provide a meaningful description.

    Definition Classes
    Scan
  8. final def eq(arg0: AnyRef): Boolean
    Definition Classes
    AnyRef
  9. def equals(arg0: AnyRef): Boolean
    Definition Classes
    AnyRef → Any
  10. final def getClass(): Class[_ <: AnyRef]
    Definition Classes
    AnyRef → Any
    Annotations
    @IntrinsicCandidate() @native()
  11. def hashCode(): Int
    Definition Classes
    AnyRef → Any
    Annotations
    @IntrinsicCandidate() @native()
  12. final def isInstanceOf[T0]: Boolean
    Definition Classes
    Any
  13. final def ne(arg0: AnyRef): Boolean
    Definition Classes
    AnyRef
  14. final def notify(): Unit
    Definition Classes
    AnyRef
    Annotations
    @IntrinsicCandidate() @native()
  15. final def notifyAll(): Unit
    Definition Classes
    AnyRef
    Annotations
    @IntrinsicCandidate() @native()
  16. def reportDriverMetrics(): Array[CustomTaskMetric]

    Returns an array of custom metrics which are collected with values at the driver side only.

    Returns an array of custom metrics which are collected with values at the driver side only. Note that these metrics must be included in the supported custom metrics reported by supportedCustomMetrics.

    Definition Classes
    Scan
    Since

    3.4.0

  17. def supportedCustomMetrics(): Array[CustomMetric]

    Returns an array of supported custom metrics with name and description.

    Returns an array of supported custom metrics with name and description. By default it returns empty array.

    Definition Classes
    Scan
  18. final def synchronized[T0](arg0: => T0): T0
    Definition Classes
    AnyRef
  19. def toBatch(): Batch

    Returns the physical representation of this scan for batch query.

    Returns the physical representation of this scan for batch query. By default this method throws exception, data sources must overwrite this method to provide an implementation, if the Table that creates this scan returns TableCapability#BATCH_READ support in its Table#capabilities().

    If the scan supports runtime filtering and implements SupportsRuntimeFiltering, this method may be called multiple times. Therefore, implementations can cache some state to avoid planning the job twice.

    Definition Classes
    Scan
    Exceptions thrown
  20. def toContinuousStream(checkpointLocation: String): ContinuousStream

    Returns the physical representation of this scan for streaming query with continuous mode.

    Returns the physical representation of this scan for streaming query with continuous mode. By default this method throws exception, data sources must overwrite this method to provide an implementation, if the Table that creates this scan returns TableCapability#CONTINUOUS_READ support in its Table#capabilities().

    checkpointLocation

    a path to Hadoop FS scratch space that can be used for failure recovery. Data streams for the same logical source in the same query will be given the same checkpointLocation.

    Definition Classes
    Scan
    Exceptions thrown
  21. def toMicroBatchStream(checkpointLocation: String): MicroBatchStream

    Returns the physical representation of this scan for streaming query with micro-batch mode.

    Returns the physical representation of this scan for streaming query with micro-batch mode. By default this method throws exception, data sources must overwrite this method to provide an implementation, if the Table that creates this scan returns TableCapability#MICRO_BATCH_READ support in its Table#capabilities().

    checkpointLocation

    a path to Hadoop FS scratch space that can be used for failure recovery. Data streams for the same logical source in the same query will be given the same checkpointLocation.

    Definition Classes
    Scan
    Exceptions thrown
  22. def toString(): String
    Definition Classes
    AnyRef → Any
  23. final def wait(arg0: Long, arg1: Int): Unit
    Definition Classes
    AnyRef
    Annotations
    @throws(classOf[java.lang.InterruptedException])
  24. final def wait(arg0: Long): Unit
    Definition Classes
    AnyRef
    Annotations
    @throws(classOf[java.lang.InterruptedException]) @native()
  25. final def wait(): Unit
    Definition Classes
    AnyRef
    Annotations
    @throws(classOf[java.lang.InterruptedException])

Deprecated Value Members

  1. def finalize(): Unit
    Attributes
    protected[lang]
    Definition Classes
    AnyRef
    Annotations
    @throws(classOf[java.lang.Throwable]) @Deprecated
    Deprecated

    (Since version 9)

Inherited from Scan

Inherited from AnyRef

Inherited from Any

Ungrouped