unpersist {SparkR}R Documentation

Unpersist

Description

Mark this SparkDataFrame as non-persistent, and remove all blocks for it from memory and disk.

Usage

unpersist(x, ...)

## S4 method for signature 'SparkDataFrame'
unpersist(x, blocking = TRUE)

Arguments

x

the SparkDataFrame to unpersist.

...

further arguments to be passed to or from other methods.

blocking

whether to block until all blocks are deleted.

Note

unpersist since 1.4.0

See Also

Other SparkDataFrame functions: SparkDataFrame-class, agg, alias, arrange, as.data.frame, attach,SparkDataFrame-method, broadcast, cache, checkpoint, coalesce, collect, colnames, coltypes, createOrReplaceTempView, crossJoin, cube, dapplyCollect, dapply, describe, dim, distinct, dropDuplicates, dropna, drop, dtypes, exceptAll, except, explain, filter, first, gapplyCollect, gapply, getNumPartitions, group_by, head, hint, histogram, insertInto, intersectAll, intersect, isLocal, isStreaming, join, limit, localCheckpoint, merge, mutate, ncol, nrow, persist, printSchema, randomSplit, rbind, rename, repartitionByRange, repartition, rollup, sample, saveAsTable, schema, selectExpr, select, showDF, show, storageLevel, str, subset, summary, take, toJSON, unionByName, union, withColumn, withWatermark, with, write.df, write.jdbc, write.json, write.orc, write.parquet, write.stream, write.text

Examples

## Not run: 
##D sparkR.session()
##D path <- "path/to/file.json"
##D df <- read.json(path)
##D persist(df, "MEMORY_AND_DISK")
##D unpersist(df)
## End(Not run)

[Package SparkR version 2.4.0 Index]