write.json {SparkR}R Documentation

Save the contents of SparkDataFrame as a JSON file

Description

Save the contents of a SparkDataFrame as a JSON file ( JSON Lines text format or newline-delimited JSON). Files written out with this method can be read back in as a SparkDataFrame using read.json().

Usage

write.json(x, path, ...)

## S4 method for signature 'SparkDataFrame,character'
write.json(x, path, mode = "error", ...)

Arguments

x

A SparkDataFrame

path

The directory where the file is saved

...

additional argument(s) passed to the method.

mode

one of 'append', 'overwrite', 'error', 'ignore' save mode (it is 'error' by default)

Note

write.json since 1.6.0

See Also

Other SparkDataFrame functions: SparkDataFrame-class, agg, arrange, as.data.frame, attach, cache, coalesce, collect, colnames, coltypes, createOrReplaceTempView, crossJoin, dapplyCollect, dapply, describe, dim, distinct, dropDuplicates, dropna, drop, dtypes, except, explain, filter, first, gapplyCollect, gapply, getNumPartitions, group_by, head, histogram, insertInto, intersect, isLocal, join, limit, merge, mutate, ncol, nrow, persist, printSchema, randomSplit, rbind, registerTempTable, rename, repartition, sample, saveAsTable, schema, selectExpr, select, showDF, show, storageLevel, str, subset, take, union, unpersist, withColumn, with, write.df, write.jdbc, write.orc, write.parquet, write.text

Examples

## Not run: 
##D sparkR.session()
##D path <- "path/to/file.json"
##D df <- read.json(path)
##D write.json(df, "/tmp/sparkr-tmp/")
## End(Not run)

[Package SparkR version 2.1.3 Index]