Core Classes#
| 
 | The entry point to programming Spark with the Dataset and DataFrame API. | 
| 
 | User-facing catalog API, accessible through SparkSession.catalog. | 
| 
 | A distributed collection of data grouped into named columns. | 
| 
 | A column in a DataFrame. | 
| 
 | Class to observe (named) metrics on a  | 
| 
 | A row in  | 
| 
 | A set of methods for aggregations on a  | 
| 
 | A logical grouping of two  | 
| Functionality for working with missing data in  | |
| Functionality for statistic functions with  | |
| 
 | Utility functions for defining window in DataFrames. | 
| 
 | Interface used to load a  | 
| 
 | Interface used to write a  | 
| 
 | Interface used to write a class:pyspark.sql.dataframe.DataFrame to external storage using the v2 API. | 
| 
 | Wrapper for user-defined function registration. | 
| 
 | Wrapper for user-defined table function registration. | 
| 
 | User defined function in Python | 
| 
 | User-defined table function in Python | 
| 
 | A base class for data sources. | 
| A base class for data source readers. | |
| A base class for streaming data source readers. | |
| A base class for data source writers. | |
| 
 | Wrapper for data source registration. | 
| 
 | A base class representing an input partition returned by the partitions() method of  | 
| A commit message returned by the  | |
| 
 | Interface for invoking table-valued functions in Spark SQL. | 
| 
 | A class to represent a Variant value in Python. | 
| Represents a table argument in PySpark. |