pyspark.sql.SparkSession.range

SparkSession.range(start: int, end: Optional[int] = None, step: int = 1, numPartitions: Optional[int] = None) → pyspark.sql.dataframe.DataFrame[source]

Create a DataFrame with single pyspark.sql.types.LongType column named id, containing elements in a range from start to end (exclusive) with step value step.

New in version 2.0.0.

Parameters
startint

the start value

endint, optional

the end value (exclusive)

stepint, optional

the incremental step (default: 1)

numPartitionsint, optional

the number of partitions of the DataFrame

Returns
DataFrame

Examples

>>> spark.range(1, 7, 2).collect()
[Row(id=1), Row(id=3), Row(id=5)]

If only one argument is specified, it will be used as the end value.

>>> spark.range(3).collect()
[Row(id=0), Row(id=1), Row(id=2)]