Packages

  • package root
    Definition Classes
    root
  • package org
    Definition Classes
    root
  • package apache
    Definition Classes
    org
  • package spark

    Core Spark functionality.

    Core Spark functionality. org.apache.spark.SparkContext serves as the main entry point to Spark, while org.apache.spark.rdd.RDD is the data type representing a distributed collection, and provides most parallel operations.

    In addition, org.apache.spark.rdd.PairRDDFunctions contains operations available only on RDDs of key-value pairs, such as groupByKey and join; org.apache.spark.rdd.DoubleRDDFunctions contains operations available only on RDDs of Doubles; and org.apache.spark.rdd.SequenceFileRDDFunctions contains operations available on RDDs that can be saved as SequenceFiles. These operations are automatically available on any RDD of the right type (e.g. RDD[(Int, Int)] through implicit conversions.

    Java programmers should reference the org.apache.spark.api.java package for Spark programming APIs in Java.

    Classes and methods marked with Experimental are user-facing features which have not been officially adopted by the Spark project. These are subject to change or removal in minor releases.

    Classes and methods marked with Developer API are intended for advanced users want to extend Spark through lower level interfaces. These are subject to changes or removal in minor releases.

    Definition Classes
    apache
  • package sql

    Allows the execution of relational queries, including those expressed in SQL using Spark.

    Allows the execution of relational queries, including those expressed in SQL using Spark.

    Definition Classes
    spark
  • package api

    Contains API classes that are specific to a single language (i.e.

    Contains API classes that are specific to a single language (i.e. Java).

    Definition Classes
    sql
  • package artifact
    Definition Classes
    sql
  • package avro
    Definition Classes
    sql
  • package catalog
    Definition Classes
    sql
  • package catalyst
    Definition Classes
    sql
  • package columnar
    Definition Classes
    sql
  • package connector
    Definition Classes
    sql
  • package expressions
    Definition Classes
    sql
  • package jdbc
    Definition Classes
    sql
  • package sources

    A set of APIs for adding data sources to Spark SQL.

    A set of APIs for adding data sources to Spark SQL.

    Definition Classes
    sql
  • package streaming
    Definition Classes
    sql
  • package types

    Contains a type system for attributes produced by relations, including complex types like structs, arrays and maps.

    Contains a type system for attributes produced by relations, including complex types like structs, arrays and maps.

    Definition Classes
    sql
  • package util
    Definition Classes
    sql
  • CaseInsensitiveStringMap
  • ExecutionListenerManager
  • MapperRowCounter
  • NumericHistogram
  • QueryExecutionListener
  • SQLOpenHashSet
  • package vectorized
    Definition Classes
    sql

package util

Ordering
  1. Alphabetic
Visibility
  1. Public
  2. Protected

Type Members

  1. class CaseInsensitiveStringMap extends Map[String, String]

    Case-insensitive map of string keys to string values.

    Case-insensitive map of string keys to string values.

    This is used to pass options to v2 implementations to ensure consistent case insensitivity.

    Methods that return keys in this map, like #entrySet() and #keySet(), return keys converted to lower case. This map doesn't allow null key.

    Annotations
    @Experimental()
    Since

    3.0.0

  2. class ExecutionListenerManager extends Logging

    Manager for QueryExecutionListener.

    Manager for QueryExecutionListener. See org.apache.spark.sql.SQLContext.listenerManager.

  3. class MapperRowCounter extends AccumulatorV2[Long, List[(Integer, Long)]]

    An AccumulatorV2 counter for collecting a list of (mapper index, row count).

    An AccumulatorV2 counter for collecting a list of (mapper index, row count).

    Since

    3.4.0

  4. class NumericHistogram extends AnyRef

    A generic, re-usable histogram class that supports partial aggregations.

    A generic, re-usable histogram class that supports partial aggregations. The algorithm is a heuristic adapted from the following paper: Yael Ben-Haim and Elad Tom-Tov, "A streaming parallel decision tree algorithm", J. Machine Learning Research 11 (2010), pp. 849--872. Although there are no approximation guarantees, it appears to work well with adequate data and a large (e.g., 20-80) number of histogram bins.

    Adapted from Hive's NumericHistogram. Can refer to https://github.com/apache/hive/blob/master/ql/src/ java/org/apache/hadoop/hive/ql/udf/generic/NumericHistogram.java

    Differences:

    1. Declaring Coord and it's variables as public types for easy access in the HistogramNumeric class. 2. Add method getNumBins() for serialize NumericHistogram in NumericHistogramSerializer. 3. Add method addBin() for deserialize NumericHistogram in NumericHistogramSerializer. 4. In Hive's code, the method pass a serialized histogram, in Spark, this method pass a deserialized histogram. Here we change the code about merge bins.
    Since

    3.3.0

  5. trait QueryExecutionListener extends AnyRef

    The interface of query execution listener that can be used to analyze execution metrics.

    The interface of query execution listener that can be used to analyze execution metrics.

    Note

    Implementations should guarantee thread-safety as they can be invoked by multiple different threads.

  6. class SQLOpenHashSet[T] extends AnyRef
    Annotations
    @Private()

Value Members

  1. object SQLOpenHashSet

Ungrouped