write.parquet {SparkR}R Documentation

Save the contents of SparkDataFrame as a Parquet file, preserving the schema.


Save the contents of a SparkDataFrame as a Parquet file, preserving the schema. Files written out with this method can be read back in as a SparkDataFrame using read.parquet().


write.parquet(x, path, ...)

saveAsParquetFile(x, path)

## S4 method for signature 'SparkDataFrame,character'
write.parquet(x, path, mode = "error",

## S4 method for signature 'SparkDataFrame,character'
saveAsParquetFile(x, path)



A SparkDataFrame


The directory where the file is saved


additional argument(s) passed to the method.


one of 'append', 'overwrite', 'error', 'errorifexists', 'ignore' save mode (it is 'error' by default)


write.parquet since 1.6.0

saveAsParquetFile since 1.4.0

See Also

Other SparkDataFrame functions: SparkDataFrame-class, agg, alias, arrange, as.data.frame, attach,SparkDataFrame-method, broadcast, cache, checkpoint, coalesce, collect, colnames, coltypes, createOrReplaceTempView, crossJoin, cube, dapplyCollect, dapply, describe, dim, distinct, dropDuplicates, dropna, drop, dtypes, except, explain, filter, first, gapplyCollect, gapply, getNumPartitions, group_by, head, hint, histogram, insertInto, intersect, isLocal, isStreaming, join, limit, localCheckpoint, merge, mutate, ncol, nrow, persist, printSchema, randomSplit, rbind, registerTempTable, rename, repartition, rollup, sample, saveAsTable, schema, selectExpr, select, showDF, show, storageLevel, str, subset, summary, take, toJSON, unionByName, union, unpersist, withColumn, withWatermark, with, write.df, write.jdbc, write.json, write.orc, write.stream, write.text


## Not run: 
##D sparkR.session()
##D path <- "path/to/file.json"
##D df <- read.json(path)
##D write.parquet(df, "/tmp/sparkr-tmp1/")
##D saveAsParquetFile(df, "/tmp/sparkr-tmp2/")
## End(Not run)

[Package SparkR version 2.3.0 Index]