union {SparkR}R Documentation

Return a new SparkDataFrame containing the union of rows


Return a new SparkDataFrame containing the union of rows in this SparkDataFrame and another SparkDataFrame. This is equivalent to UNION ALL in SQL. Input SparkDataFrames can have different schemas (names and data types).

unionAll is deprecated - use union instead


union(x, y)

unionAll(x, y)

## S4 method for signature 'SparkDataFrame,SparkDataFrame'
union(x, y)

## S4 method for signature 'SparkDataFrame,SparkDataFrame'
unionAll(x, y)



A SparkDataFrame


A SparkDataFrame


Note: This does not remove duplicate rows across the two SparkDataFrames. Also as standard in SQL, this function resolves columns by position (not by name).


A SparkDataFrame containing the result of the union.


union since 2.0.0

unionAll since 1.4.0

See Also


Other SparkDataFrame functions: SparkDataFrame-class, agg, arrange, as.data.frame, attach,SparkDataFrame-method, cache, checkpoint, coalesce, collect, colnames, coltypes, createOrReplaceTempView, crossJoin, dapplyCollect, dapply, describe, dim, distinct, dropDuplicates, dropna, drop, dtypes, except, explain, filter, first, gapplyCollect, gapply, getNumPartitions, group_by, head, hint, histogram, insertInto, intersect, isLocal, isStreaming, join, limit, merge, mutate, ncol, nrow, persist, printSchema, randomSplit, rbind, registerTempTable, rename, repartition, sample, saveAsTable, schema, selectExpr, select, showDF, show, storageLevel, str, subset, take, toJSON, unpersist, withColumn, with, write.df, write.jdbc, write.json, write.orc, write.parquet, write.stream, write.text


## Not run: 
##D sparkR.session()
##D df1 <- read.json(path)
##D df2 <- read.json(path2)
##D unioned <- union(df, df2)
##D unions <- rbind(df, df2, df3, df4)
## End(Not run)

[Package SparkR version 2.2.1 Index]