copy_to.spark_connection
From sparklyr v1.4.0
by Yitao Li
Copy an R Data Frame to Spark
Copy an R data.frame
to Spark, and return a reference to the
generated Spark DataFrame as a tbl_spark
. The returned object will
act as a dplyr
-compatible interface to the underlying Spark table.
Usage
# S3 method for spark_connection
copy_to(
dest,
df,
name = spark_table_name(substitute(df)),
overwrite = FALSE,
memory = TRUE,
repartition = 0L,
...
)
Arguments
- dest
A
spark_connection
.- df
An R
data.frame
.- name
The name to assign to the copied table in Spark.
- overwrite
Boolean; overwrite a pre-existing table with the name
name
if one already exists?- memory
Boolean; should the table be cached into memory?
- repartition
The number of partitions to use when distributing the table across the Spark cluster. The default (0) can be used to avoid partitioning.
- ...
Optional arguments; currently unused.
Value
A tbl_spark
, representing a dplyr
-compatible interface
to a Spark DataFrame.
Community examples
Looks like there are no examples yet.