org.apache.spark.ml.source.libsvm

DefaultSource

class DefaultSource extends RelationProvider with DataSourceRegister

libsvm package implements Spark SQL data source API for loading LIBSVM data as DataFrame. The loaded DataFrame has two columns: label containing labels stored as doubles and features containing feature vectors stored as Vectors.

To use LIBSVM data source, you need to set "libsvm" as the format in DataFrameReader and optionally specify options, for example:

// Scala
val df = sqlContext.read.format("libsvm")
  .option("numFeatures", "780")
  .load("data/mllib/sample_libsvm_data.txt")

// Java
DataFrame df = sqlContext.read().format("libsvm")
  .option("numFeatures, "780")
  .load("data/mllib/sample_libsvm_data.txt");

LIBSVM data source supports the following options:

Annotations
@Since( "1.6.0" )
Source
LibSVMRelation.scala
See also

LIBSVM datasets

Linear Supertypes
Ordering
  1. Alphabetic
  2. By inheritance
Inherited
  1. DefaultSource
  2. DataSourceRegister
  3. RelationProvider
  4. AnyRef
  5. Any
  1. Hide All
  2. Show all
Learn more about member selection
Visibility
  1. Public
  2. All

Instance Constructors

  1. new DefaultSource()

Value Members

  1. final def !=(arg0: AnyRef): Boolean

    Definition Classes
    AnyRef
  2. final def !=(arg0: Any): Boolean

    Definition Classes
    Any
  3. final def ##(): Int

    Definition Classes
    AnyRef → Any
  4. final def ==(arg0: AnyRef): Boolean

    Definition Classes
    AnyRef
  5. final def ==(arg0: Any): Boolean

    Definition Classes
    Any
  6. final def asInstanceOf[T0]: T0

    Definition Classes
    Any
  7. def clone(): AnyRef

    Attributes
    protected[java.lang]
    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  8. def createRelation(sqlContext: SQLContext, parameters: Map[String, String]): BaseRelation

    Returns a new base relation with the given parameters.

    Returns a new base relation with the given parameters. Note: the parameters' keywords are case insensitive and this insensitivity is enforced by the Map that is passed to the function.

    Definition Classes
    DefaultSourceRelationProvider
    Annotations
    @Since( "1.6.0" )
  9. final def eq(arg0: AnyRef): Boolean

    Definition Classes
    AnyRef
  10. def equals(arg0: Any): Boolean

    Definition Classes
    AnyRef → Any
  11. def finalize(): Unit

    Attributes
    protected[java.lang]
    Definition Classes
    AnyRef
    Annotations
    @throws( classOf[java.lang.Throwable] )
  12. final def getClass(): Class[_]

    Definition Classes
    AnyRef → Any
  13. def hashCode(): Int

    Definition Classes
    AnyRef → Any
  14. final def isInstanceOf[T0]: Boolean

    Definition Classes
    Any
  15. final def ne(arg0: AnyRef): Boolean

    Definition Classes
    AnyRef
  16. final def notify(): Unit

    Definition Classes
    AnyRef
  17. final def notifyAll(): Unit

    Definition Classes
    AnyRef
  18. def shortName(): String

    The string that represents the format that this data source provider uses.

    The string that represents the format that this data source provider uses. This is overridden by children to provide a nice alias for the data source. For example:

    override def format(): String = "parquet"
    Definition Classes
    DefaultSourceDataSourceRegister
    Annotations
    @Since( "1.6.0" )
    Since

    1.5.0

  19. final def synchronized[T0](arg0: ⇒ T0): T0

    Definition Classes
    AnyRef
  20. def toString(): String

    Definition Classes
    AnyRef → Any
  21. final def wait(): Unit

    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  22. final def wait(arg0: Long, arg1: Int): Unit

    Definition Classes
    AnyRef
    Annotations
    @throws( ... )
  23. final def wait(arg0: Long): Unit

    Definition Classes
    AnyRef
    Annotations
    @throws( ... )

Inherited from DataSourceRegister

Inherited from RelationProvider

Inherited from AnyRef

Inherited from Any

Members