spark-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From lix...@apache.org
Subject [41/61] [partial] spark-website git commit: Add docs for Spark 2.4.0 and update the latest link
Date Thu, 08 Nov 2018 14:48:29 GMT
http://git-wip-us.apache.org/repos/asf/spark-website/blob/52917ac4/site/docs/2.4.0/api/R/over.html
----------------------------------------------------------------------
diff --git a/site/docs/2.4.0/api/R/over.html b/site/docs/2.4.0/api/R/over.html
new file mode 100644
index 0000000..94a55e5
--- /dev/null
+++ b/site/docs/2.4.0/api/R/over.html
@@ -0,0 +1,79 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: over</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for over {SparkR}"><tr><td>over {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>over</h2>
+
+<h3>Description</h3>
+
+<p>Define a windowing column.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+over(x, window)
+
+## S4 method for signature 'Column,WindowSpec'
+over(x, window)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>x</code></td>
+<td>
+<p>a Column, usually one returned by window function(s).</p>
+</td></tr>
+<tr valign="top"><td><code>window</code></td>
+<td>
+<p>a WindowSpec object. Can be created by <code>windowPartitionBy</code> or
+<code>windowOrderBy</code> and configured by other WindowSpec methods.</p>
+</td></tr>
+</table>
+
+
+<h3>Note</h3>
+
+<p>over since 2.0.0
+</p>
+
+
+<h3>See Also</h3>
+
+<p>Other colum_func: <code><a href="alias.html">alias</a></code>,
+<code><a href="between.html">between</a></code>, <code><a href="cast.html">cast</a></code>,
+<code><a href="endsWith.html">endsWith</a></code>, <code><a href="otherwise.html">otherwise</a></code>,
+<code><a href="startsWith.html">startsWith</a></code>, <code><a href="substr.html">substr</a></code>
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D   df &lt;- createDataFrame(mtcars)
+##D 
+##D   # Partition by am (transmission) and order by hp (horsepower)
+##D   ws &lt;- orderBy(windowPartitionBy(&quot;am&quot;), &quot;hp&quot;)
+##D 
+##D   # Rank on hp within each partition
+##D   out &lt;- select(df, over(rank(), ws), df$hp, df$am)
+##D 
+##D   # Lag mpg values by 1 row on the partition-and-ordered table
+##D   out &lt;- select(df, over(lead(df$mpg), ws), df$mpg, df$hp, df$am)
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.0 <a href="00Index.html">Index</a>]</div>
+</body></html>

http://git-wip-us.apache.org/repos/asf/spark-website/blob/52917ac4/site/docs/2.4.0/api/R/partitionBy.html
----------------------------------------------------------------------
diff --git a/site/docs/2.4.0/api/R/partitionBy.html b/site/docs/2.4.0/api/R/partitionBy.html
new file mode 100644
index 0000000..9096895
--- /dev/null
+++ b/site/docs/2.4.0/api/R/partitionBy.html
@@ -0,0 +1,78 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: partitionBy</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for partitionBy {SparkR}"><tr><td>partitionBy {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>partitionBy</h2>
+
+<h3>Description</h3>
+
+<p>Defines the partitioning columns in a WindowSpec.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+partitionBy(x, ...)
+
+## S4 method for signature 'WindowSpec'
+partitionBy(x, col, ...)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>x</code></td>
+<td>
+<p>a WindowSpec.</p>
+</td></tr>
+<tr valign="top"><td><code>...</code></td>
+<td>
+<p>additional column(s) to partition on.</p>
+</td></tr>
+<tr valign="top"><td><code>col</code></td>
+<td>
+<p>a column to partition on (desribed by the name or Column).</p>
+</td></tr>
+</table>
+
+
+<h3>Value</h3>
+
+<p>A WindowSpec.
+</p>
+
+
+<h3>Note</h3>
+
+<p>partitionBy(WindowSpec) since 2.0.0
+</p>
+
+
+<h3>See Also</h3>
+
+<p>Other windowspec_method: <code><a href="orderBy.html">orderBy</a></code>,
+<code><a href="rangeBetween.html">rangeBetween</a></code>, <code><a href="rowsBetween.html">rowsBetween</a></code>
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D   partitionBy(ws, &quot;col1&quot;, &quot;col2&quot;)
+##D   partitionBy(ws, df$col1, df$col2)
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.0 <a href="00Index.html">Index</a>]</div>
+</body></html>

http://git-wip-us.apache.org/repos/asf/spark-website/blob/52917ac4/site/docs/2.4.0/api/R/persist.html
----------------------------------------------------------------------
diff --git a/site/docs/2.4.0/api/R/persist.html b/site/docs/2.4.0/api/R/persist.html
new file mode 100644
index 0000000..6039a7e
--- /dev/null
+++ b/site/docs/2.4.0/api/R/persist.html
@@ -0,0 +1,114 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: Persist</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for persist {SparkR}"><tr><td>persist {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>Persist</h2>
+
+<h3>Description</h3>
+
+<p>Persist this SparkDataFrame with the specified storage level. For details of the
+supported storage levels, refer to
+<a href="http://spark.apache.org/docs/latest/rdd-programming-guide.html#rdd-persistence">http://spark.apache.org/docs/latest/rdd-programming-guide.html#rdd-persistence</a>.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+persist(x, newLevel)
+
+## S4 method for signature 'SparkDataFrame,character'
+persist(x, newLevel)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>x</code></td>
+<td>
+<p>the SparkDataFrame to persist.</p>
+</td></tr>
+<tr valign="top"><td><code>newLevel</code></td>
+<td>
+<p>storage level chosen for the persistence. See available options in
+the description.</p>
+</td></tr>
+</table>
+
+
+<h3>Note</h3>
+
+<p>persist since 1.4.0
+</p>
+
+
+<h3>See Also</h3>
+
+<p>Other SparkDataFrame functions: <code><a href="SparkDataFrame.html">SparkDataFrame-class</a></code>,
+<code><a href="summarize.html">agg</a></code>, <code><a href="alias.html">alias</a></code>,
+<code><a href="arrange.html">arrange</a></code>, <code><a href="as.data.frame.html">as.data.frame</a></code>,
+<code><a href="attach.html">attach,SparkDataFrame-method</a></code>,
+<code><a href="broadcast.html">broadcast</a></code>, <code><a href="cache.html">cache</a></code>,
+<code><a href="checkpoint.html">checkpoint</a></code>, <code><a href="coalesce.html">coalesce</a></code>,
+<code><a href="collect.html">collect</a></code>, <code><a href="columns.html">colnames</a></code>,
+<code><a href="coltypes.html">coltypes</a></code>,
+<code><a href="createOrReplaceTempView.html">createOrReplaceTempView</a></code>,
+<code><a href="crossJoin.html">crossJoin</a></code>, <code><a href="cube.html">cube</a></code>,
+<code><a href="dapplyCollect.html">dapplyCollect</a></code>, <code><a href="dapply.html">dapply</a></code>,
+<code><a href="describe.html">describe</a></code>, <code><a href="dim.html">dim</a></code>,
+<code><a href="distinct.html">distinct</a></code>, <code><a href="dropDuplicates.html">dropDuplicates</a></code>,
+<code><a href="nafunctions.html">dropna</a></code>, <code><a href="drop.html">drop</a></code>,
+<code><a href="dtypes.html">dtypes</a></code>, <code><a href="exceptAll.html">exceptAll</a></code>,
+<code><a href="except.html">except</a></code>, <code><a href="explain.html">explain</a></code>,
+<code><a href="filter.html">filter</a></code>, <code><a href="first.html">first</a></code>,
+<code><a href="gapplyCollect.html">gapplyCollect</a></code>, <code><a href="gapply.html">gapply</a></code>,
+<code><a href="getNumPartitions.html">getNumPartitions</a></code>, <code><a href="groupBy.html">group_by</a></code>,
+<code><a href="head.html">head</a></code>, <code><a href="hint.html">hint</a></code>,
+<code><a href="histogram.html">histogram</a></code>, <code><a href="insertInto.html">insertInto</a></code>,
+<code><a href="intersectAll.html">intersectAll</a></code>, <code><a href="intersect.html">intersect</a></code>,
+<code><a href="isLocal.html">isLocal</a></code>, <code><a href="isStreaming.html">isStreaming</a></code>,
+<code><a href="join.html">join</a></code>, <code><a href="limit.html">limit</a></code>,
+<code><a href="localCheckpoint.html">localCheckpoint</a></code>, <code><a href="merge.html">merge</a></code>,
+<code><a href="mutate.html">mutate</a></code>, <code><a href="ncol.html">ncol</a></code>,
+<code><a href="nrow.html">nrow</a></code>, <code><a href="printSchema.html">printSchema</a></code>,
+<code><a href="randomSplit.html">randomSplit</a></code>, <code><a href="rbind.html">rbind</a></code>,
+<code><a href="rename.html">rename</a></code>, <code><a href="repartitionByRange.html">repartitionByRange</a></code>,
+<code><a href="repartition.html">repartition</a></code>, <code><a href="rollup.html">rollup</a></code>,
+<code><a href="sample.html">sample</a></code>, <code><a href="saveAsTable.html">saveAsTable</a></code>,
+<code><a href="schema.html">schema</a></code>, <code><a href="selectExpr.html">selectExpr</a></code>,
+<code><a href="select.html">select</a></code>, <code><a href="showDF.html">showDF</a></code>,
+<code><a href="show.html">show</a></code>, <code><a href="storageLevel.html">storageLevel</a></code>,
+<code><a href="str.html">str</a></code>, <code><a href="subset.html">subset</a></code>,
+<code><a href="summary.html">summary</a></code>, <code><a href="take.html">take</a></code>,
+<code><a href="toJSON.html">toJSON</a></code>, <code><a href="unionByName.html">unionByName</a></code>,
+<code><a href="union.html">union</a></code>, <code><a href="unpersist.html">unpersist</a></code>,
+<code><a href="withColumn.html">withColumn</a></code>, <code><a href="withWatermark.html">withWatermark</a></code>,
+<code><a href="with.html">with</a></code>, <code><a href="write.df.html">write.df</a></code>,
+<code><a href="write.jdbc.html">write.jdbc</a></code>, <code><a href="write.json.html">write.json</a></code>,
+<code><a href="write.orc.html">write.orc</a></code>, <code><a href="write.parquet.html">write.parquet</a></code>,
+<code><a href="write.stream.html">write.stream</a></code>, <code><a href="write.text.html">write.text</a></code>
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D sparkR.session()
+##D path &lt;- &quot;path/to/file.json&quot;
+##D df &lt;- read.json(path)
+##D persist(df, &quot;MEMORY_AND_DISK&quot;)
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.0 <a href="00Index.html">Index</a>]</div>
+</body></html>

http://git-wip-us.apache.org/repos/asf/spark-website/blob/52917ac4/site/docs/2.4.0/api/R/pivot.html
----------------------------------------------------------------------
diff --git a/site/docs/2.4.0/api/R/pivot.html b/site/docs/2.4.0/api/R/pivot.html
new file mode 100644
index 0000000..bd43b31
--- /dev/null
+++ b/site/docs/2.4.0/api/R/pivot.html
@@ -0,0 +1,80 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: Pivot a column of the GroupedData and perform the specified...</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for pivot {SparkR}"><tr><td>pivot {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>Pivot a column of the GroupedData and perform the specified aggregation.</h2>
+
+<h3>Description</h3>
+
+<p>Pivot a column of the GroupedData and perform the specified aggregation.
+There are two versions of pivot function: one that requires the caller to specify the list
+of distinct values to pivot on, and one that does not. The latter is more concise but less
+efficient, because Spark needs to first compute the list of distinct values internally.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+## S4 method for signature 'GroupedData,character'
+pivot(x, colname, values = list())
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>x</code></td>
+<td>
+<p>a GroupedData object</p>
+</td></tr>
+<tr valign="top"><td><code>colname</code></td>
+<td>
+<p>A column name</p>
+</td></tr>
+<tr valign="top"><td><code>values</code></td>
+<td>
+<p>A value or a list/vector of distinct values for the output columns.</p>
+</td></tr>
+</table>
+
+
+<h3>Value</h3>
+
+<p>GroupedData object
+</p>
+
+
+<h3>Note</h3>
+
+<p>pivot since 2.0.0
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D df &lt;- createDataFrame(data.frame(
+##D     earnings = c(10000, 10000, 11000, 15000, 12000, 20000, 21000, 22000),
+##D     course = c(&quot;R&quot;, &quot;Python&quot;, &quot;R&quot;, &quot;Python&quot;, &quot;R&quot;, &quot;Python&quot;, &quot;R&quot;, &quot;Python&quot;),
+##D     period = c(&quot;1H&quot;, &quot;1H&quot;, &quot;2H&quot;, &quot;2H&quot;, &quot;1H&quot;, &quot;1H&quot;, &quot;2H&quot;, &quot;2H&quot;),
+##D     year = c(2015, 2015, 2015, 2015, 2016, 2016, 2016, 2016)
+##D ))
+##D group_sum &lt;- sum(pivot(groupBy(df, &quot;year&quot;), &quot;course&quot;), &quot;earnings&quot;)
+##D group_min &lt;- min(pivot(groupBy(df, &quot;year&quot;), &quot;course&quot;, &quot;R&quot;), &quot;earnings&quot;)
+##D group_max &lt;- max(pivot(groupBy(df, &quot;year&quot;), &quot;course&quot;, c(&quot;Python&quot;, &quot;R&quot;)), &quot;earnings&quot;)
+##D group_mean &lt;- mean(pivot(groupBy(df, &quot;year&quot;), &quot;course&quot;, list(&quot;Python&quot;, &quot;R&quot;)), &quot;earnings&quot;)
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.0 <a href="00Index.html">Index</a>]</div>
+</body></html>

http://git-wip-us.apache.org/repos/asf/spark-website/blob/52917ac4/site/docs/2.4.0/api/R/predict.html
----------------------------------------------------------------------
diff --git a/site/docs/2.4.0/api/R/predict.html b/site/docs/2.4.0/api/R/predict.html
new file mode 100644
index 0000000..3378f3f
--- /dev/null
+++ b/site/docs/2.4.0/api/R/predict.html
@@ -0,0 +1,54 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: Makes predictions from a MLlib model</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+</head><body>
+
+<table width="100%" summary="page for predict {SparkR}"><tr><td>predict {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>Makes predictions from a MLlib model</h2>
+
+<h3>Description</h3>
+
+<p>Makes predictions from a MLlib model. For more information, see the specific
+MLlib model below.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+predict(object, ...)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>object</code></td>
+<td>
+<p>a fitted ML model object.</p>
+</td></tr>
+<tr valign="top"><td><code>...</code></td>
+<td>
+<p>additional argument(s) passed to the method.</p>
+</td></tr>
+</table>
+
+
+<h3>See Also</h3>
+
+<p><a href="spark.als.html">spark.als</a>, <a href="spark.bisectingKmeans.html">spark.bisectingKmeans</a>, <a href="spark.decisionTree.html">spark.decisionTree</a>,
+</p>
+<p><a href="spark.gaussianMixture.html">spark.gaussianMixture</a>, <a href="spark.gbt.html">spark.gbt</a>,
+</p>
+<p><a href="spark.glm.html">spark.glm</a>, <a href="glm.html">glm</a>, <a href="spark.isoreg.html">spark.isoreg</a>,
+</p>
+<p><a href="spark.kmeans.html">spark.kmeans</a>,
+</p>
+<p><a href="spark.logit.html">spark.logit</a>, <a href="spark.mlp.html">spark.mlp</a>, <a href="spark.naiveBayes.html">spark.naiveBayes</a>,
+</p>
+<p><a href="spark.randomForest.html">spark.randomForest</a>, <a href="spark.survreg.html">spark.survreg</a>, <a href="spark.svmLinear.html">spark.svmLinear</a>
+</p>
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.0 <a href="00Index.html">Index</a>]</div>
+</body></html>

http://git-wip-us.apache.org/repos/asf/spark-website/blob/52917ac4/site/docs/2.4.0/api/R/print.jobj.html
----------------------------------------------------------------------
diff --git a/site/docs/2.4.0/api/R/print.jobj.html b/site/docs/2.4.0/api/R/print.jobj.html
new file mode 100644
index 0000000..491aba1
--- /dev/null
+++ b/site/docs/2.4.0/api/R/print.jobj.html
@@ -0,0 +1,45 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: Print a JVM object reference.</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+</head><body>
+
+<table width="100%" summary="page for print.jobj {SparkR}"><tr><td>print.jobj {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>Print a JVM object reference.</h2>
+
+<h3>Description</h3>
+
+<p>This function prints the type and id for an object stored
+in the SparkR JVM backend.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+## S3 method for class 'jobj'
+print(x, ...)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>x</code></td>
+<td>
+<p>The JVM object reference</p>
+</td></tr>
+<tr valign="top"><td><code>...</code></td>
+<td>
+<p>further arguments passed to or from other methods</p>
+</td></tr>
+</table>
+
+
+<h3>Note</h3>
+
+<p>print.jobj since 1.4.0
+</p>
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.0 <a href="00Index.html">Index</a>]</div>
+</body></html>

http://git-wip-us.apache.org/repos/asf/spark-website/blob/52917ac4/site/docs/2.4.0/api/R/print.structField.html
----------------------------------------------------------------------
diff --git a/site/docs/2.4.0/api/R/print.structField.html b/site/docs/2.4.0/api/R/print.structField.html
new file mode 100644
index 0000000..4c2a474
--- /dev/null
+++ b/site/docs/2.4.0/api/R/print.structField.html
@@ -0,0 +1,45 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: Print a Spark StructField.</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+</head><body>
+
+<table width="100%" summary="page for print.structField {SparkR}"><tr><td>print.structField {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>Print a Spark StructField.</h2>
+
+<h3>Description</h3>
+
+<p>This function prints the contents of a StructField returned from the
+SparkR JVM backend.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+## S3 method for class 'structField'
+print(x, ...)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>x</code></td>
+<td>
+<p>A StructField object</p>
+</td></tr>
+<tr valign="top"><td><code>...</code></td>
+<td>
+<p>further arguments passed to or from other methods</p>
+</td></tr>
+</table>
+
+
+<h3>Note</h3>
+
+<p>print.structField since 1.4.0
+</p>
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.0 <a href="00Index.html">Index</a>]</div>
+</body></html>

http://git-wip-us.apache.org/repos/asf/spark-website/blob/52917ac4/site/docs/2.4.0/api/R/print.structType.html
----------------------------------------------------------------------
diff --git a/site/docs/2.4.0/api/R/print.structType.html b/site/docs/2.4.0/api/R/print.structType.html
new file mode 100644
index 0000000..0549a43
--- /dev/null
+++ b/site/docs/2.4.0/api/R/print.structType.html
@@ -0,0 +1,45 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: Print a Spark StructType.</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+</head><body>
+
+<table width="100%" summary="page for print.structType {SparkR}"><tr><td>print.structType {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>Print a Spark StructType.</h2>
+
+<h3>Description</h3>
+
+<p>This function prints the contents of a StructType returned from the
+SparkR JVM backend.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+## S3 method for class 'structType'
+print(x, ...)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>x</code></td>
+<td>
+<p>A StructType object</p>
+</td></tr>
+<tr valign="top"><td><code>...</code></td>
+<td>
+<p>further arguments passed to or from other methods</p>
+</td></tr>
+</table>
+
+
+<h3>Note</h3>
+
+<p>print.structType since 1.4.0
+</p>
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.0 <a href="00Index.html">Index</a>]</div>
+</body></html>

http://git-wip-us.apache.org/repos/asf/spark-website/blob/52917ac4/site/docs/2.4.0/api/R/printSchema.html
----------------------------------------------------------------------
diff --git a/site/docs/2.4.0/api/R/printSchema.html b/site/docs/2.4.0/api/R/printSchema.html
new file mode 100644
index 0000000..dea41c7
--- /dev/null
+++ b/site/docs/2.4.0/api/R/printSchema.html
@@ -0,0 +1,107 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: Print Schema of a SparkDataFrame</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for printSchema {SparkR}"><tr><td>printSchema {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>Print Schema of a SparkDataFrame</h2>
+
+<h3>Description</h3>
+
+<p>Prints out the schema in tree format
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+printSchema(x)
+
+## S4 method for signature 'SparkDataFrame'
+printSchema(x)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>x</code></td>
+<td>
+<p>A SparkDataFrame</p>
+</td></tr>
+</table>
+
+
+<h3>Note</h3>
+
+<p>printSchema since 1.4.0
+</p>
+
+
+<h3>See Also</h3>
+
+<p>Other SparkDataFrame functions: <code><a href="SparkDataFrame.html">SparkDataFrame-class</a></code>,
+<code><a href="summarize.html">agg</a></code>, <code><a href="alias.html">alias</a></code>,
+<code><a href="arrange.html">arrange</a></code>, <code><a href="as.data.frame.html">as.data.frame</a></code>,
+<code><a href="attach.html">attach,SparkDataFrame-method</a></code>,
+<code><a href="broadcast.html">broadcast</a></code>, <code><a href="cache.html">cache</a></code>,
+<code><a href="checkpoint.html">checkpoint</a></code>, <code><a href="coalesce.html">coalesce</a></code>,
+<code><a href="collect.html">collect</a></code>, <code><a href="columns.html">colnames</a></code>,
+<code><a href="coltypes.html">coltypes</a></code>,
+<code><a href="createOrReplaceTempView.html">createOrReplaceTempView</a></code>,
+<code><a href="crossJoin.html">crossJoin</a></code>, <code><a href="cube.html">cube</a></code>,
+<code><a href="dapplyCollect.html">dapplyCollect</a></code>, <code><a href="dapply.html">dapply</a></code>,
+<code><a href="describe.html">describe</a></code>, <code><a href="dim.html">dim</a></code>,
+<code><a href="distinct.html">distinct</a></code>, <code><a href="dropDuplicates.html">dropDuplicates</a></code>,
+<code><a href="nafunctions.html">dropna</a></code>, <code><a href="drop.html">drop</a></code>,
+<code><a href="dtypes.html">dtypes</a></code>, <code><a href="exceptAll.html">exceptAll</a></code>,
+<code><a href="except.html">except</a></code>, <code><a href="explain.html">explain</a></code>,
+<code><a href="filter.html">filter</a></code>, <code><a href="first.html">first</a></code>,
+<code><a href="gapplyCollect.html">gapplyCollect</a></code>, <code><a href="gapply.html">gapply</a></code>,
+<code><a href="getNumPartitions.html">getNumPartitions</a></code>, <code><a href="groupBy.html">group_by</a></code>,
+<code><a href="head.html">head</a></code>, <code><a href="hint.html">hint</a></code>,
+<code><a href="histogram.html">histogram</a></code>, <code><a href="insertInto.html">insertInto</a></code>,
+<code><a href="intersectAll.html">intersectAll</a></code>, <code><a href="intersect.html">intersect</a></code>,
+<code><a href="isLocal.html">isLocal</a></code>, <code><a href="isStreaming.html">isStreaming</a></code>,
+<code><a href="join.html">join</a></code>, <code><a href="limit.html">limit</a></code>,
+<code><a href="localCheckpoint.html">localCheckpoint</a></code>, <code><a href="merge.html">merge</a></code>,
+<code><a href="mutate.html">mutate</a></code>, <code><a href="ncol.html">ncol</a></code>,
+<code><a href="nrow.html">nrow</a></code>, <code><a href="persist.html">persist</a></code>,
+<code><a href="randomSplit.html">randomSplit</a></code>, <code><a href="rbind.html">rbind</a></code>,
+<code><a href="rename.html">rename</a></code>, <code><a href="repartitionByRange.html">repartitionByRange</a></code>,
+<code><a href="repartition.html">repartition</a></code>, <code><a href="rollup.html">rollup</a></code>,
+<code><a href="sample.html">sample</a></code>, <code><a href="saveAsTable.html">saveAsTable</a></code>,
+<code><a href="schema.html">schema</a></code>, <code><a href="selectExpr.html">selectExpr</a></code>,
+<code><a href="select.html">select</a></code>, <code><a href="showDF.html">showDF</a></code>,
+<code><a href="show.html">show</a></code>, <code><a href="storageLevel.html">storageLevel</a></code>,
+<code><a href="str.html">str</a></code>, <code><a href="subset.html">subset</a></code>,
+<code><a href="summary.html">summary</a></code>, <code><a href="take.html">take</a></code>,
+<code><a href="toJSON.html">toJSON</a></code>, <code><a href="unionByName.html">unionByName</a></code>,
+<code><a href="union.html">union</a></code>, <code><a href="unpersist.html">unpersist</a></code>,
+<code><a href="withColumn.html">withColumn</a></code>, <code><a href="withWatermark.html">withWatermark</a></code>,
+<code><a href="with.html">with</a></code>, <code><a href="write.df.html">write.df</a></code>,
+<code><a href="write.jdbc.html">write.jdbc</a></code>, <code><a href="write.json.html">write.json</a></code>,
+<code><a href="write.orc.html">write.orc</a></code>, <code><a href="write.parquet.html">write.parquet</a></code>,
+<code><a href="write.stream.html">write.stream</a></code>, <code><a href="write.text.html">write.text</a></code>
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D sparkR.session()
+##D path &lt;- &quot;path/to/file.json&quot;
+##D df &lt;- read.json(path)
+##D printSchema(df)
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.0 <a href="00Index.html">Index</a>]</div>
+</body></html>

http://git-wip-us.apache.org/repos/asf/spark-website/blob/52917ac4/site/docs/2.4.0/api/R/queryName.html
----------------------------------------------------------------------
diff --git a/site/docs/2.4.0/api/R/queryName.html b/site/docs/2.4.0/api/R/queryName.html
new file mode 100644
index 0000000..01c9331
--- /dev/null
+++ b/site/docs/2.4.0/api/R/queryName.html
@@ -0,0 +1,75 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: queryName</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for queryName {SparkR}"><tr><td>queryName {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>queryName</h2>
+
+<h3>Description</h3>
+
+<p>Returns the user-specified name of the query. This is specified in
+<code>write.stream(df, queryName = "query")</code>. This name, if set, must be unique across all active
+queries.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+queryName(x)
+
+## S4 method for signature 'StreamingQuery'
+queryName(x)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>x</code></td>
+<td>
+<p>a StreamingQuery.</p>
+</td></tr>
+</table>
+
+
+<h3>Value</h3>
+
+<p>The name of the query, or NULL if not specified.
+</p>
+
+
+<h3>Note</h3>
+
+<p>queryName(StreamingQuery) since 2.2.0
+</p>
+<p>experimental
+</p>
+
+
+<h3>See Also</h3>
+
+<p><a href="write.stream.html">write.stream</a>
+</p>
+<p>Other StreamingQuery methods: <code><a href="awaitTermination.html">awaitTermination</a></code>,
+<code><a href="explain.html">explain</a></code>, <code><a href="isActive.html">isActive</a></code>,
+<code><a href="lastProgress.html">lastProgress</a></code>, <code><a href="status.html">status</a></code>,
+<code><a href="stopQuery.html">stopQuery</a></code>
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run:  queryName(sq) 
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.0 <a href="00Index.html">Index</a>]</div>
+</body></html>

http://git-wip-us.apache.org/repos/asf/spark-website/blob/52917ac4/site/docs/2.4.0/api/R/randomSplit.html
----------------------------------------------------------------------
diff --git a/site/docs/2.4.0/api/R/randomSplit.html b/site/docs/2.4.0/api/R/randomSplit.html
new file mode 100644
index 0000000..14af526
--- /dev/null
+++ b/site/docs/2.4.0/api/R/randomSplit.html
@@ -0,0 +1,116 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: randomSplit</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for randomSplit {SparkR}"><tr><td>randomSplit {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>randomSplit</h2>
+
+<h3>Description</h3>
+
+<p>Return a list of randomly split dataframes with the provided weights.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+randomSplit(x, weights, seed)
+
+## S4 method for signature 'SparkDataFrame,numeric'
+randomSplit(x, weights, seed)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>x</code></td>
+<td>
+<p>A SparkDataFrame</p>
+</td></tr>
+<tr valign="top"><td><code>weights</code></td>
+<td>
+<p>A vector of weights for splits, will be normalized if they don't sum to 1</p>
+</td></tr>
+<tr valign="top"><td><code>seed</code></td>
+<td>
+<p>A seed to use for random split</p>
+</td></tr>
+</table>
+
+
+<h3>Note</h3>
+
+<p>randomSplit since 2.0.0
+</p>
+
+
+<h3>See Also</h3>
+
+<p>Other SparkDataFrame functions: <code><a href="SparkDataFrame.html">SparkDataFrame-class</a></code>,
+<code><a href="summarize.html">agg</a></code>, <code><a href="alias.html">alias</a></code>,
+<code><a href="arrange.html">arrange</a></code>, <code><a href="as.data.frame.html">as.data.frame</a></code>,
+<code><a href="attach.html">attach,SparkDataFrame-method</a></code>,
+<code><a href="broadcast.html">broadcast</a></code>, <code><a href="cache.html">cache</a></code>,
+<code><a href="checkpoint.html">checkpoint</a></code>, <code><a href="coalesce.html">coalesce</a></code>,
+<code><a href="collect.html">collect</a></code>, <code><a href="columns.html">colnames</a></code>,
+<code><a href="coltypes.html">coltypes</a></code>,
+<code><a href="createOrReplaceTempView.html">createOrReplaceTempView</a></code>,
+<code><a href="crossJoin.html">crossJoin</a></code>, <code><a href="cube.html">cube</a></code>,
+<code><a href="dapplyCollect.html">dapplyCollect</a></code>, <code><a href="dapply.html">dapply</a></code>,
+<code><a href="describe.html">describe</a></code>, <code><a href="dim.html">dim</a></code>,
+<code><a href="distinct.html">distinct</a></code>, <code><a href="dropDuplicates.html">dropDuplicates</a></code>,
+<code><a href="nafunctions.html">dropna</a></code>, <code><a href="drop.html">drop</a></code>,
+<code><a href="dtypes.html">dtypes</a></code>, <code><a href="exceptAll.html">exceptAll</a></code>,
+<code><a href="except.html">except</a></code>, <code><a href="explain.html">explain</a></code>,
+<code><a href="filter.html">filter</a></code>, <code><a href="first.html">first</a></code>,
+<code><a href="gapplyCollect.html">gapplyCollect</a></code>, <code><a href="gapply.html">gapply</a></code>,
+<code><a href="getNumPartitions.html">getNumPartitions</a></code>, <code><a href="groupBy.html">group_by</a></code>,
+<code><a href="head.html">head</a></code>, <code><a href="hint.html">hint</a></code>,
+<code><a href="histogram.html">histogram</a></code>, <code><a href="insertInto.html">insertInto</a></code>,
+<code><a href="intersectAll.html">intersectAll</a></code>, <code><a href="intersect.html">intersect</a></code>,
+<code><a href="isLocal.html">isLocal</a></code>, <code><a href="isStreaming.html">isStreaming</a></code>,
+<code><a href="join.html">join</a></code>, <code><a href="limit.html">limit</a></code>,
+<code><a href="localCheckpoint.html">localCheckpoint</a></code>, <code><a href="merge.html">merge</a></code>,
+<code><a href="mutate.html">mutate</a></code>, <code><a href="ncol.html">ncol</a></code>,
+<code><a href="nrow.html">nrow</a></code>, <code><a href="persist.html">persist</a></code>,
+<code><a href="printSchema.html">printSchema</a></code>, <code><a href="rbind.html">rbind</a></code>,
+<code><a href="rename.html">rename</a></code>, <code><a href="repartitionByRange.html">repartitionByRange</a></code>,
+<code><a href="repartition.html">repartition</a></code>, <code><a href="rollup.html">rollup</a></code>,
+<code><a href="sample.html">sample</a></code>, <code><a href="saveAsTable.html">saveAsTable</a></code>,
+<code><a href="schema.html">schema</a></code>, <code><a href="selectExpr.html">selectExpr</a></code>,
+<code><a href="select.html">select</a></code>, <code><a href="showDF.html">showDF</a></code>,
+<code><a href="show.html">show</a></code>, <code><a href="storageLevel.html">storageLevel</a></code>,
+<code><a href="str.html">str</a></code>, <code><a href="subset.html">subset</a></code>,
+<code><a href="summary.html">summary</a></code>, <code><a href="take.html">take</a></code>,
+<code><a href="toJSON.html">toJSON</a></code>, <code><a href="unionByName.html">unionByName</a></code>,
+<code><a href="union.html">union</a></code>, <code><a href="unpersist.html">unpersist</a></code>,
+<code><a href="withColumn.html">withColumn</a></code>, <code><a href="withWatermark.html">withWatermark</a></code>,
+<code><a href="with.html">with</a></code>, <code><a href="write.df.html">write.df</a></code>,
+<code><a href="write.jdbc.html">write.jdbc</a></code>, <code><a href="write.json.html">write.json</a></code>,
+<code><a href="write.orc.html">write.orc</a></code>, <code><a href="write.parquet.html">write.parquet</a></code>,
+<code><a href="write.stream.html">write.stream</a></code>, <code><a href="write.text.html">write.text</a></code>
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D sparkR.session()
+##D df &lt;- createDataFrame(data.frame(id = 1:1000))
+##D df_list &lt;- randomSplit(df, c(2, 3, 5), 0)
+##D # df_list contains 3 SparkDataFrames with each having about 200, 300 and 500 rows respectively
+##D sapply(df_list, count)
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.0 <a href="00Index.html">Index</a>]</div>
+</body></html>

http://git-wip-us.apache.org/repos/asf/spark-website/blob/52917ac4/site/docs/2.4.0/api/R/rangeBetween.html
----------------------------------------------------------------------
diff --git a/site/docs/2.4.0/api/R/rangeBetween.html b/site/docs/2.4.0/api/R/rangeBetween.html
new file mode 100644
index 0000000..1be8f32
--- /dev/null
+++ b/site/docs/2.4.0/api/R/rangeBetween.html
@@ -0,0 +1,87 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: rangeBetween</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for rangeBetween {SparkR}"><tr><td>rangeBetween {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>rangeBetween</h2>
+
+<h3>Description</h3>
+
+<p>Defines the frame boundaries, from <code>start</code> (inclusive) to <code>end</code> (inclusive).
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+rangeBetween(x, start, end)
+
+## S4 method for signature 'WindowSpec,numeric,numeric'
+rangeBetween(x, start, end)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>x</code></td>
+<td>
+<p>a WindowSpec</p>
+</td></tr>
+<tr valign="top"><td><code>start</code></td>
+<td>
+<p>boundary start, inclusive.
+The frame is unbounded if this is the minimum long value.</p>
+</td></tr>
+<tr valign="top"><td><code>end</code></td>
+<td>
+<p>boundary end, inclusive.
+The frame is unbounded if this is the maximum long value.</p>
+</td></tr>
+</table>
+
+
+<h3>Details</h3>
+
+<p>Both <code>start</code> and <code>end</code> are relative from the current row. For example, &quot;0&quot; means
+&quot;current row&quot;, while &quot;-1&quot; means one off before the current row, and &quot;5&quot; means the five off
+after the current row.
+</p>
+
+
+<h3>Value</h3>
+
+<p>a WindowSpec
+</p>
+
+
+<h3>Note</h3>
+
+<p>rangeBetween since 2.0.0
+</p>
+
+
+<h3>See Also</h3>
+
+<p>Other windowspec_method: <code><a href="orderBy.html">orderBy</a></code>,
+<code><a href="partitionBy.html">partitionBy</a></code>, <code><a href="rowsBetween.html">rowsBetween</a></code>
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D   rangeBetween(ws, 0, 3)
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.0 <a href="00Index.html">Index</a>]</div>
+</body></html>

http://git-wip-us.apache.org/repos/asf/spark-website/blob/52917ac4/site/docs/2.4.0/api/R/rbind.html
----------------------------------------------------------------------
diff --git a/site/docs/2.4.0/api/R/rbind.html b/site/docs/2.4.0/api/R/rbind.html
new file mode 100644
index 0000000..0db32a6
--- /dev/null
+++ b/site/docs/2.4.0/api/R/rbind.html
@@ -0,0 +1,129 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: Union two or more SparkDataFrames</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for rbind {SparkR}"><tr><td>rbind {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>Union two or more SparkDataFrames</h2>
+
+<h3>Description</h3>
+
+<p>Union two or more SparkDataFrames by row. As in R's <code>rbind</code>, this method
+requires that the input SparkDataFrames have the same column names.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+rbind(..., deparse.level = 1)
+
+## S4 method for signature 'SparkDataFrame'
+rbind(x, ..., deparse.level = 1)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>...</code></td>
+<td>
+<p>additional SparkDataFrame(s).</p>
+</td></tr>
+<tr valign="top"><td><code>deparse.level</code></td>
+<td>
+<p>currently not used (put here to match the signature of
+the base implementation).</p>
+</td></tr>
+<tr valign="top"><td><code>x</code></td>
+<td>
+<p>a SparkDataFrame.</p>
+</td></tr>
+</table>
+
+
+<h3>Details</h3>
+
+<p>Note: This does not remove duplicate rows across the two SparkDataFrames.
+</p>
+
+
+<h3>Value</h3>
+
+<p>A SparkDataFrame containing the result of the union.
+</p>
+
+
+<h3>Note</h3>
+
+<p>rbind since 1.5.0
+</p>
+
+
+<h3>See Also</h3>
+
+<p><a href="union.html">union</a> <a href="unionByName.html">unionByName</a>
+</p>
+<p>Other SparkDataFrame functions: <code><a href="SparkDataFrame.html">SparkDataFrame-class</a></code>,
+<code><a href="summarize.html">agg</a></code>, <code><a href="alias.html">alias</a></code>,
+<code><a href="arrange.html">arrange</a></code>, <code><a href="as.data.frame.html">as.data.frame</a></code>,
+<code><a href="attach.html">attach,SparkDataFrame-method</a></code>,
+<code><a href="broadcast.html">broadcast</a></code>, <code><a href="cache.html">cache</a></code>,
+<code><a href="checkpoint.html">checkpoint</a></code>, <code><a href="coalesce.html">coalesce</a></code>,
+<code><a href="collect.html">collect</a></code>, <code><a href="columns.html">colnames</a></code>,
+<code><a href="coltypes.html">coltypes</a></code>,
+<code><a href="createOrReplaceTempView.html">createOrReplaceTempView</a></code>,
+<code><a href="crossJoin.html">crossJoin</a></code>, <code><a href="cube.html">cube</a></code>,
+<code><a href="dapplyCollect.html">dapplyCollect</a></code>, <code><a href="dapply.html">dapply</a></code>,
+<code><a href="describe.html">describe</a></code>, <code><a href="dim.html">dim</a></code>,
+<code><a href="distinct.html">distinct</a></code>, <code><a href="dropDuplicates.html">dropDuplicates</a></code>,
+<code><a href="nafunctions.html">dropna</a></code>, <code><a href="drop.html">drop</a></code>,
+<code><a href="dtypes.html">dtypes</a></code>, <code><a href="exceptAll.html">exceptAll</a></code>,
+<code><a href="except.html">except</a></code>, <code><a href="explain.html">explain</a></code>,
+<code><a href="filter.html">filter</a></code>, <code><a href="first.html">first</a></code>,
+<code><a href="gapplyCollect.html">gapplyCollect</a></code>, <code><a href="gapply.html">gapply</a></code>,
+<code><a href="getNumPartitions.html">getNumPartitions</a></code>, <code><a href="groupBy.html">group_by</a></code>,
+<code><a href="head.html">head</a></code>, <code><a href="hint.html">hint</a></code>,
+<code><a href="histogram.html">histogram</a></code>, <code><a href="insertInto.html">insertInto</a></code>,
+<code><a href="intersectAll.html">intersectAll</a></code>, <code><a href="intersect.html">intersect</a></code>,
+<code><a href="isLocal.html">isLocal</a></code>, <code><a href="isStreaming.html">isStreaming</a></code>,
+<code><a href="join.html">join</a></code>, <code><a href="limit.html">limit</a></code>,
+<code><a href="localCheckpoint.html">localCheckpoint</a></code>, <code><a href="merge.html">merge</a></code>,
+<code><a href="mutate.html">mutate</a></code>, <code><a href="ncol.html">ncol</a></code>,
+<code><a href="nrow.html">nrow</a></code>, <code><a href="persist.html">persist</a></code>,
+<code><a href="printSchema.html">printSchema</a></code>, <code><a href="randomSplit.html">randomSplit</a></code>,
+<code><a href="rename.html">rename</a></code>, <code><a href="repartitionByRange.html">repartitionByRange</a></code>,
+<code><a href="repartition.html">repartition</a></code>, <code><a href="rollup.html">rollup</a></code>,
+<code><a href="sample.html">sample</a></code>, <code><a href="saveAsTable.html">saveAsTable</a></code>,
+<code><a href="schema.html">schema</a></code>, <code><a href="selectExpr.html">selectExpr</a></code>,
+<code><a href="select.html">select</a></code>, <code><a href="showDF.html">showDF</a></code>,
+<code><a href="show.html">show</a></code>, <code><a href="storageLevel.html">storageLevel</a></code>,
+<code><a href="str.html">str</a></code>, <code><a href="subset.html">subset</a></code>,
+<code><a href="summary.html">summary</a></code>, <code><a href="take.html">take</a></code>,
+<code><a href="toJSON.html">toJSON</a></code>, <code><a href="unionByName.html">unionByName</a></code>,
+<code><a href="union.html">union</a></code>, <code><a href="unpersist.html">unpersist</a></code>,
+<code><a href="withColumn.html">withColumn</a></code>, <code><a href="withWatermark.html">withWatermark</a></code>,
+<code><a href="with.html">with</a></code>, <code><a href="write.df.html">write.df</a></code>,
+<code><a href="write.jdbc.html">write.jdbc</a></code>, <code><a href="write.json.html">write.json</a></code>,
+<code><a href="write.orc.html">write.orc</a></code>, <code><a href="write.parquet.html">write.parquet</a></code>,
+<code><a href="write.stream.html">write.stream</a></code>, <code><a href="write.text.html">write.text</a></code>
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D sparkR.session()
+##D unions &lt;- rbind(df, df2, df3, df4)
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.0 <a href="00Index.html">Index</a>]</div>
+</body></html>

http://git-wip-us.apache.org/repos/asf/spark-website/blob/52917ac4/site/docs/2.4.0/api/R/read.df.html
----------------------------------------------------------------------
diff --git a/site/docs/2.4.0/api/R/read.df.html b/site/docs/2.4.0/api/R/read.df.html
new file mode 100644
index 0000000..06ffcf3
--- /dev/null
+++ b/site/docs/2.4.0/api/R/read.df.html
@@ -0,0 +1,106 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: Load a SparkDataFrame</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for read.df {SparkR}"><tr><td>read.df {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>Load a SparkDataFrame</h2>
+
+<h3>Description</h3>
+
+<p>Returns the dataset in a data source as a SparkDataFrame
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+## Default S3 method:
+read.df(path = NULL, source = NULL, schema = NULL,
+  na.strings = "NA", ...)
+
+## Default S3 method:
+loadDF(path = NULL, source = NULL, schema = NULL,
+  ...)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>path</code></td>
+<td>
+<p>The path of files to load</p>
+</td></tr>
+<tr valign="top"><td><code>source</code></td>
+<td>
+<p>The name of external data source</p>
+</td></tr>
+<tr valign="top"><td><code>schema</code></td>
+<td>
+<p>The data schema defined in structType or a DDL-formatted string.</p>
+</td></tr>
+<tr valign="top"><td><code>na.strings</code></td>
+<td>
+<p>Default string value for NA when source is &quot;csv&quot;</p>
+</td></tr>
+<tr valign="top"><td><code>...</code></td>
+<td>
+<p>additional external data source specific named properties.</p>
+</td></tr>
+</table>
+
+
+<h3>Details</h3>
+
+<p>The data source is specified by the <code>source</code> and a set of options(...).
+If <code>source</code> is not specified, the default data source configured by
+&quot;spark.sql.sources.default&quot; will be used. <br />
+Similar to R read.csv, when <code>source</code> is &quot;csv&quot;, by default, a value of &quot;NA&quot; will be
+interpreted as NA.
+</p>
+
+
+<h3>Value</h3>
+
+<p>SparkDataFrame
+</p>
+
+
+<h3>Note</h3>
+
+<p>read.df since 1.4.0
+</p>
+<p>loadDF since 1.6.0
+</p>
+
+
+<h3>See Also</h3>
+
+<p><a href="read.json.html">read.json</a>
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D sparkR.session()
+##D df1 &lt;- read.df(&quot;path/to/file.json&quot;, source = &quot;json&quot;)
+##D schema &lt;- structType(structField(&quot;name&quot;, &quot;string&quot;),
+##D                      structField(&quot;info&quot;, &quot;map&lt;string,double&gt;&quot;))
+##D df2 &lt;- read.df(mapTypeJsonPath, &quot;json&quot;, schema, multiLine = TRUE)
+##D df3 &lt;- loadDF(&quot;data/test_table&quot;, &quot;parquet&quot;, mergeSchema = &quot;true&quot;)
+##D stringSchema &lt;- &quot;name STRING, info MAP&lt;STRING, DOUBLE&gt;&quot;
+##D df4 &lt;- read.df(mapTypeJsonPath, &quot;json&quot;, stringSchema, multiLine = TRUE)
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.0 <a href="00Index.html">Index</a>]</div>
+</body></html>

http://git-wip-us.apache.org/repos/asf/spark-website/blob/52917ac4/site/docs/2.4.0/api/R/read.jdbc.html
----------------------------------------------------------------------
diff --git a/site/docs/2.4.0/api/R/read.jdbc.html b/site/docs/2.4.0/api/R/read.jdbc.html
new file mode 100644
index 0000000..a4c48c1
--- /dev/null
+++ b/site/docs/2.4.0/api/R/read.jdbc.html
@@ -0,0 +1,105 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: Create a SparkDataFrame representing the database table...</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for read.jdbc {SparkR}"><tr><td>read.jdbc {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>Create a SparkDataFrame representing the database table accessible via JDBC URL</h2>
+
+<h3>Description</h3>
+
+<p>Additional JDBC database connection properties can be set (...)
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+read.jdbc(url, tableName, partitionColumn = NULL, lowerBound = NULL,
+  upperBound = NULL, numPartitions = 0L, predicates = list(), ...)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>url</code></td>
+<td>
+<p>JDBC database url of the form <code>jdbc:subprotocol:subname</code></p>
+</td></tr>
+<tr valign="top"><td><code>tableName</code></td>
+<td>
+<p>the name of the table in the external database</p>
+</td></tr>
+<tr valign="top"><td><code>partitionColumn</code></td>
+<td>
+<p>the name of a column of integral type that will be used for partitioning</p>
+</td></tr>
+<tr valign="top"><td><code>lowerBound</code></td>
+<td>
+<p>the minimum value of <code>partitionColumn</code> used to decide partition stride</p>
+</td></tr>
+<tr valign="top"><td><code>upperBound</code></td>
+<td>
+<p>the maximum value of <code>partitionColumn</code> used to decide partition stride</p>
+</td></tr>
+<tr valign="top"><td><code>numPartitions</code></td>
+<td>
+<p>the number of partitions, This, along with <code>lowerBound</code> (inclusive),
+<code>upperBound</code> (exclusive), form partition strides for generated WHERE
+clause expressions used to split the column <code>partitionColumn</code> evenly.
+This defaults to SparkContext.defaultParallelism when unset.</p>
+</td></tr>
+<tr valign="top"><td><code>predicates</code></td>
+<td>
+<p>a list of conditions in the where clause; each one defines one partition</p>
+</td></tr>
+<tr valign="top"><td><code>...</code></td>
+<td>
+<p>additional JDBC database connection named properties.</p>
+</td></tr>
+</table>
+
+
+<h3>Details</h3>
+
+<p>Only one of partitionColumn or predicates should be set. Partitions of the table will be
+retrieved in parallel based on the <code>numPartitions</code> or by the predicates.
+</p>
+<p>Don't create too many partitions in parallel on a large cluster; otherwise Spark might crash
+your external database systems.
+</p>
+
+
+<h3>Value</h3>
+
+<p>SparkDataFrame
+</p>
+
+
+<h3>Note</h3>
+
+<p>read.jdbc since 2.0.0
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D sparkR.session()
+##D jdbcUrl &lt;- &quot;jdbc:mysql://localhost:3306/databasename&quot;
+##D df &lt;- read.jdbc(jdbcUrl, &quot;table&quot;, predicates = list(&quot;field&lt;=123&quot;), user = &quot;username&quot;)
+##D df2 &lt;- read.jdbc(jdbcUrl, &quot;table2&quot;, partitionColumn = &quot;index&quot;, lowerBound = 0,
+##D                  upperBound = 10000, user = &quot;username&quot;, password = &quot;password&quot;)
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.0 <a href="00Index.html">Index</a>]</div>
+</body></html>

http://git-wip-us.apache.org/repos/asf/spark-website/blob/52917ac4/site/docs/2.4.0/api/R/read.json.html
----------------------------------------------------------------------
diff --git a/site/docs/2.4.0/api/R/read.json.html b/site/docs/2.4.0/api/R/read.json.html
new file mode 100644
index 0000000..0cd1ce0
--- /dev/null
+++ b/site/docs/2.4.0/api/R/read.json.html
@@ -0,0 +1,77 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: Create a SparkDataFrame from a JSON file.</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for read.json {SparkR}"><tr><td>read.json {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>Create a SparkDataFrame from a JSON file.</h2>
+
+<h3>Description</h3>
+
+<p>Loads a JSON file, returning the result as a SparkDataFrame
+By default, (<a href="http://jsonlines.org/">JSON Lines text format or newline-delimited JSON</a>
+) is supported. For JSON (one record per file), set a named property <code>multiLine</code> to
+<code>TRUE</code>.
+It goes through the entire dataset once to determine the schema.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+## Default S3 method:
+read.json(path, ...)
+
+## Default S3 method:
+jsonFile(path)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>path</code></td>
+<td>
+<p>Path of file to read. A vector of multiple paths is allowed.</p>
+</td></tr>
+<tr valign="top"><td><code>...</code></td>
+<td>
+<p>additional external data source specific named properties.</p>
+</td></tr>
+</table>
+
+
+<h3>Value</h3>
+
+<p>SparkDataFrame
+</p>
+
+
+<h3>Note</h3>
+
+<p>read.json since 1.6.0
+</p>
+<p>jsonFile since 1.4.0
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D sparkR.session()
+##D path &lt;- &quot;path/to/file.json&quot;
+##D df &lt;- read.json(path)
+##D df &lt;- read.json(path, multiLine = TRUE)
+##D df &lt;- jsonFile(path)
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.0 <a href="00Index.html">Index</a>]</div>
+</body></html>

http://git-wip-us.apache.org/repos/asf/spark-website/blob/52917ac4/site/docs/2.4.0/api/R/read.ml.html
----------------------------------------------------------------------
diff --git a/site/docs/2.4.0/api/R/read.ml.html b/site/docs/2.4.0/api/R/read.ml.html
new file mode 100644
index 0000000..c023193
--- /dev/null
+++ b/site/docs/2.4.0/api/R/read.ml.html
@@ -0,0 +1,66 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: Load a fitted MLlib model from the input path.</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for read.ml {SparkR}"><tr><td>read.ml {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>Load a fitted MLlib model from the input path.</h2>
+
+<h3>Description</h3>
+
+<p>Load a fitted MLlib model from the input path.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+read.ml(path)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>path</code></td>
+<td>
+<p>path of the model to read.</p>
+</td></tr>
+</table>
+
+
+<h3>Value</h3>
+
+<p>A fitted MLlib model.
+</p>
+
+
+<h3>Note</h3>
+
+<p>read.ml since 2.0.0
+</p>
+
+
+<h3>See Also</h3>
+
+<p><a href="write.ml.html">write.ml</a>
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D path &lt;- &quot;path/to/model&quot;
+##D model &lt;- read.ml(path)
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.0 <a href="00Index.html">Index</a>]</div>
+</body></html>

http://git-wip-us.apache.org/repos/asf/spark-website/blob/52917ac4/site/docs/2.4.0/api/R/read.orc.html
----------------------------------------------------------------------
diff --git a/site/docs/2.4.0/api/R/read.orc.html b/site/docs/2.4.0/api/R/read.orc.html
new file mode 100644
index 0000000..5e916f6
--- /dev/null
+++ b/site/docs/2.4.0/api/R/read.orc.html
@@ -0,0 +1,49 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: Create a SparkDataFrame from an ORC file.</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+</head><body>
+
+<table width="100%" summary="page for read.orc {SparkR}"><tr><td>read.orc {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>Create a SparkDataFrame from an ORC file.</h2>
+
+<h3>Description</h3>
+
+<p>Loads an ORC file, returning the result as a SparkDataFrame.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+read.orc(path, ...)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>path</code></td>
+<td>
+<p>Path of file to read.</p>
+</td></tr>
+<tr valign="top"><td><code>...</code></td>
+<td>
+<p>additional external data source specific named properties.</p>
+</td></tr>
+</table>
+
+
+<h3>Value</h3>
+
+<p>SparkDataFrame
+</p>
+
+
+<h3>Note</h3>
+
+<p>read.orc since 2.0.0
+</p>
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.0 <a href="00Index.html">Index</a>]</div>
+</body></html>

http://git-wip-us.apache.org/repos/asf/spark-website/blob/52917ac4/site/docs/2.4.0/api/R/read.parquet.html
----------------------------------------------------------------------
diff --git a/site/docs/2.4.0/api/R/read.parquet.html b/site/docs/2.4.0/api/R/read.parquet.html
new file mode 100644
index 0000000..a504f54
--- /dev/null
+++ b/site/docs/2.4.0/api/R/read.parquet.html
@@ -0,0 +1,55 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: Create a SparkDataFrame from a Parquet file.</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+</head><body>
+
+<table width="100%" summary="page for read.parquet {SparkR}"><tr><td>read.parquet {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>Create a SparkDataFrame from a Parquet file.</h2>
+
+<h3>Description</h3>
+
+<p>Loads a Parquet file, returning the result as a SparkDataFrame.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+## Default S3 method:
+read.parquet(path, ...)
+
+## Default S3 method:
+parquetFile(...)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>path</code></td>
+<td>
+<p>path of file to read. A vector of multiple paths is allowed.</p>
+</td></tr>
+<tr valign="top"><td><code>...</code></td>
+<td>
+<p>argument(s) passed to the method.</p>
+</td></tr>
+</table>
+
+
+<h3>Value</h3>
+
+<p>SparkDataFrame
+</p>
+
+
+<h3>Note</h3>
+
+<p>read.parquet since 1.6.0
+</p>
+<p>parquetFile since 1.4.0
+</p>
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.0 <a href="00Index.html">Index</a>]</div>
+</body></html>

http://git-wip-us.apache.org/repos/asf/spark-website/blob/52917ac4/site/docs/2.4.0/api/R/read.stream.html
----------------------------------------------------------------------
diff --git a/site/docs/2.4.0/api/R/read.stream.html b/site/docs/2.4.0/api/R/read.stream.html
new file mode 100644
index 0000000..0662162
--- /dev/null
+++ b/site/docs/2.4.0/api/R/read.stream.html
@@ -0,0 +1,93 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: Load a streaming SparkDataFrame</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for read.stream {SparkR}"><tr><td>read.stream {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>Load a streaming SparkDataFrame</h2>
+
+<h3>Description</h3>
+
+<p>Returns the dataset in a data source as a SparkDataFrame
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+read.stream(source = NULL, schema = NULL, ...)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>source</code></td>
+<td>
+<p>The name of external data source</p>
+</td></tr>
+<tr valign="top"><td><code>schema</code></td>
+<td>
+<p>The data schema defined in structType or a DDL-formatted string, this is
+required for file-based streaming data source</p>
+</td></tr>
+<tr valign="top"><td><code>...</code></td>
+<td>
+<p>additional external data source specific named options, for instance <code>path</code> for
+file-based streaming data source. <code>timeZone</code> to indicate a timezone to be used to
+parse timestamps in the JSON/CSV data sources or partition values; If it isn't set, it
+uses the default value, session local timezone.</p>
+</td></tr>
+</table>
+
+
+<h3>Details</h3>
+
+<p>The data source is specified by the <code>source</code> and a set of options(...).
+If <code>source</code> is not specified, the default data source configured by
+&quot;spark.sql.sources.default&quot; will be used.
+</p>
+
+
+<h3>Value</h3>
+
+<p>SparkDataFrame
+</p>
+
+
+<h3>Note</h3>
+
+<p>read.stream since 2.2.0
+</p>
+<p>experimental
+</p>
+
+
+<h3>See Also</h3>
+
+<p><a href="write.stream.html">write.stream</a>
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D sparkR.session()
+##D df &lt;- read.stream(&quot;socket&quot;, host = &quot;localhost&quot;, port = 9999)
+##D q &lt;- write.stream(df, &quot;text&quot;, path = &quot;/home/user/out&quot;, checkpointLocation = &quot;/home/user/cp&quot;)
+##D 
+##D df &lt;- read.stream(&quot;json&quot;, path = jsonDir, schema = schema, maxFilesPerTrigger = 1)
+##D stringSchema &lt;- &quot;name STRING, info MAP&lt;STRING, DOUBLE&gt;&quot;
+##D df1 &lt;- read.stream(&quot;json&quot;, path = jsonDir, schema = stringSchema, maxFilesPerTrigger = 1)
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.0 <a href="00Index.html">Index</a>]</div>
+</body></html>

http://git-wip-us.apache.org/repos/asf/spark-website/blob/52917ac4/site/docs/2.4.0/api/R/read.text.html
----------------------------------------------------------------------
diff --git a/site/docs/2.4.0/api/R/read.text.html b/site/docs/2.4.0/api/R/read.text.html
new file mode 100644
index 0000000..46ab5fe
--- /dev/null
+++ b/site/docs/2.4.0/api/R/read.text.html
@@ -0,0 +1,74 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: Create a SparkDataFrame from a text file.</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for read.text {SparkR}"><tr><td>read.text {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>Create a SparkDataFrame from a text file.</h2>
+
+<h3>Description</h3>
+
+<p>Loads text files and returns a SparkDataFrame whose schema starts with
+a string column named &quot;value&quot;, and followed by partitioned columns if
+there are any.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+## Default S3 method:
+read.text(path, ...)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>path</code></td>
+<td>
+<p>Path of file to read. A vector of multiple paths is allowed.</p>
+</td></tr>
+<tr valign="top"><td><code>...</code></td>
+<td>
+<p>additional external data source specific named properties.</p>
+</td></tr>
+</table>
+
+
+<h3>Details</h3>
+
+<p>Each line in the text file is a new row in the resulting SparkDataFrame.
+</p>
+
+
+<h3>Value</h3>
+
+<p>SparkDataFrame
+</p>
+
+
+<h3>Note</h3>
+
+<p>read.text since 1.6.1
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D sparkR.session()
+##D path &lt;- &quot;path/to/file.txt&quot;
+##D df &lt;- read.text(path)
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.0 <a href="00Index.html">Index</a>]</div>
+</body></html>

http://git-wip-us.apache.org/repos/asf/spark-website/blob/52917ac4/site/docs/2.4.0/api/R/recoverPartitions.html
----------------------------------------------------------------------
diff --git a/site/docs/2.4.0/api/R/recoverPartitions.html b/site/docs/2.4.0/api/R/recoverPartitions.html
new file mode 100644
index 0000000..21ac068
--- /dev/null
+++ b/site/docs/2.4.0/api/R/recoverPartitions.html
@@ -0,0 +1,56 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: Recovers all the partitions in the directory of a table and...</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for recoverPartitions {SparkR}"><tr><td>recoverPartitions {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>Recovers all the partitions in the directory of a table and update the catalog</h2>
+
+<h3>Description</h3>
+
+<p>Recovers all the partitions in the directory of a table and update the catalog. The name should
+reference a partitioned table, and not a view.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+recoverPartitions(tableName)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>tableName</code></td>
+<td>
+<p>the qualified or unqualified name that designates a table. If no database
+identifier is provided, it refers to a table in the current database.</p>
+</td></tr>
+</table>
+
+
+<h3>Note</h3>
+
+<p>since 2.2.0
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D sparkR.session()
+##D recoverPartitions(&quot;myTable&quot;)
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.0 <a href="00Index.html">Index</a>]</div>
+</body></html>

http://git-wip-us.apache.org/repos/asf/spark-website/blob/52917ac4/site/docs/2.4.0/api/R/refreshByPath.html
----------------------------------------------------------------------
diff --git a/site/docs/2.4.0/api/R/refreshByPath.html b/site/docs/2.4.0/api/R/refreshByPath.html
new file mode 100644
index 0000000..32edab5
--- /dev/null
+++ b/site/docs/2.4.0/api/R/refreshByPath.html
@@ -0,0 +1,56 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: Invalidates and refreshes all the cached data and metadata...</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for refreshByPath {SparkR}"><tr><td>refreshByPath {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>Invalidates and refreshes all the cached data and metadata for SparkDataFrame containing path</h2>
+
+<h3>Description</h3>
+
+<p>Invalidates and refreshes all the cached data (and the associated metadata) for any
+SparkDataFrame that contains the given data source path. Path matching is by prefix, i.e. &quot;/&quot;
+would invalidate everything that is cached.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+refreshByPath(path)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>path</code></td>
+<td>
+<p>the path of the data source.</p>
+</td></tr>
+</table>
+
+
+<h3>Note</h3>
+
+<p>since 2.2.0
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D sparkR.session()
+##D refreshByPath(&quot;/path&quot;)
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.0 <a href="00Index.html">Index</a>]</div>
+</body></html>

http://git-wip-us.apache.org/repos/asf/spark-website/blob/52917ac4/site/docs/2.4.0/api/R/refreshTable.html
----------------------------------------------------------------------
diff --git a/site/docs/2.4.0/api/R/refreshTable.html b/site/docs/2.4.0/api/R/refreshTable.html
new file mode 100644
index 0000000..69c8a5c
--- /dev/null
+++ b/site/docs/2.4.0/api/R/refreshTable.html
@@ -0,0 +1,65 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: Invalidates and refreshes all the cached data and metadata of...</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for refreshTable {SparkR}"><tr><td>refreshTable {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>Invalidates and refreshes all the cached data and metadata of the given table</h2>
+
+<h3>Description</h3>
+
+<p>Invalidates and refreshes all the cached data and metadata of the given table. For performance
+reasons, Spark SQL or the external data source library it uses might cache certain metadata about
+a table, such as the location of blocks. When those change outside of Spark SQL, users should
+call this function to invalidate the cache.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+refreshTable(tableName)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>tableName</code></td>
+<td>
+<p>the qualified or unqualified name that designates a table. If no database
+identifier is provided, it refers to a table in the current database.</p>
+</td></tr>
+</table>
+
+
+<h3>Details</h3>
+
+<p>If this table is cached as an InMemoryRelation, drop the original cached version and make the
+new version cached lazily.
+</p>
+
+
+<h3>Note</h3>
+
+<p>since 2.2.0
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D sparkR.session()
+##D refreshTable(&quot;myTable&quot;)
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.0 <a href="00Index.html">Index</a>]</div>
+</body></html>

http://git-wip-us.apache.org/repos/asf/spark-website/blob/52917ac4/site/docs/2.4.0/api/R/registerTempTable-deprecated.html
----------------------------------------------------------------------
diff --git a/site/docs/2.4.0/api/R/registerTempTable-deprecated.html b/site/docs/2.4.0/api/R/registerTempTable-deprecated.html
new file mode 100644
index 0000000..ae96830
--- /dev/null
+++ b/site/docs/2.4.0/api/R/registerTempTable-deprecated.html
@@ -0,0 +1,70 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: (Deprecated) Register Temporary Table</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for registerTempTable {SparkR}"><tr><td>registerTempTable {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>(Deprecated) Register Temporary Table</h2>
+
+<h3>Description</h3>
+
+<p>Registers a SparkDataFrame as a Temporary Table in the SparkSession
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+registerTempTable(x, tableName)
+
+## S4 method for signature 'SparkDataFrame,character'
+registerTempTable(x, tableName)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>x</code></td>
+<td>
+<p>A SparkDataFrame</p>
+</td></tr>
+<tr valign="top"><td><code>tableName</code></td>
+<td>
+<p>A character vector containing the name of the table</p>
+</td></tr>
+</table>
+
+
+<h3>Note</h3>
+
+<p>registerTempTable since 1.4.0
+</p>
+
+
+<h3>See Also</h3>
+
+<p><a href="createOrReplaceTempView.html">createOrReplaceTempView</a>
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D sparkR.session()
+##D path &lt;- &quot;path/to/file.json&quot;
+##D df &lt;- read.json(path)
+##D registerTempTable(df, &quot;json_df&quot;)
+##D new_df &lt;- sql(&quot;SELECT * FROM json_df&quot;)
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.0 <a href="00Index.html">Index</a>]</div>
+</body></html>

http://git-wip-us.apache.org/repos/asf/spark-website/blob/52917ac4/site/docs/2.4.0/api/R/rename.html
----------------------------------------------------------------------
diff --git a/site/docs/2.4.0/api/R/rename.html b/site/docs/2.4.0/api/R/rename.html
new file mode 100644
index 0000000..db45034
--- /dev/null
+++ b/site/docs/2.4.0/api/R/rename.html
@@ -0,0 +1,141 @@
+<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd"><html xmlns="http://www.w3.org/1999/xhtml"><head><title>R: rename</title>
+<meta http-equiv="Content-Type" content="text/html; charset=utf-8" />
+<link rel="stylesheet" type="text/css" href="R.css" />
+
+<link rel="stylesheet" href="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/styles/github.min.css">
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/highlight.min.js"></script>
+<script src="https://cdnjs.cloudflare.com/ajax/libs/highlight.js/8.3/languages/r.min.js"></script>
+<script>hljs.initHighlightingOnLoad();</script>
+</head><body>
+
+<table width="100%" summary="page for rename {SparkR}"><tr><td>rename {SparkR}</td><td style="text-align: right;">R Documentation</td></tr></table>
+
+<h2>rename</h2>
+
+<h3>Description</h3>
+
+<p>Rename an existing column in a SparkDataFrame.
+</p>
+
+
+<h3>Usage</h3>
+
+<pre>
+rename(x, ...)
+
+withColumnRenamed(x, existingCol, newCol)
+
+## S4 method for signature 'SparkDataFrame,character,character'
+withColumnRenamed(x,
+  existingCol, newCol)
+
+## S4 method for signature 'SparkDataFrame'
+rename(x, ...)
+</pre>
+
+
+<h3>Arguments</h3>
+
+<table summary="R argblock">
+<tr valign="top"><td><code>x</code></td>
+<td>
+<p>A SparkDataFrame</p>
+</td></tr>
+<tr valign="top"><td><code>...</code></td>
+<td>
+<p>A named pair of the form new_column_name = existing_column</p>
+</td></tr>
+<tr valign="top"><td><code>existingCol</code></td>
+<td>
+<p>The name of the column you want to change.</p>
+</td></tr>
+<tr valign="top"><td><code>newCol</code></td>
+<td>
+<p>The new column name.</p>
+</td></tr>
+</table>
+
+
+<h3>Value</h3>
+
+<p>A SparkDataFrame with the column name changed.
+</p>
+
+
+<h3>Note</h3>
+
+<p>withColumnRenamed since 1.4.0
+</p>
+<p>rename since 1.4.0
+</p>
+
+
+<h3>See Also</h3>
+
+<p><a href="mutate.html">mutate</a>
+</p>
+<p>Other SparkDataFrame functions: <code><a href="SparkDataFrame.html">SparkDataFrame-class</a></code>,
+<code><a href="summarize.html">agg</a></code>, <code><a href="alias.html">alias</a></code>,
+<code><a href="arrange.html">arrange</a></code>, <code><a href="as.data.frame.html">as.data.frame</a></code>,
+<code><a href="attach.html">attach,SparkDataFrame-method</a></code>,
+<code><a href="broadcast.html">broadcast</a></code>, <code><a href="cache.html">cache</a></code>,
+<code><a href="checkpoint.html">checkpoint</a></code>, <code><a href="coalesce.html">coalesce</a></code>,
+<code><a href="collect.html">collect</a></code>, <code><a href="columns.html">colnames</a></code>,
+<code><a href="coltypes.html">coltypes</a></code>,
+<code><a href="createOrReplaceTempView.html">createOrReplaceTempView</a></code>,
+<code><a href="crossJoin.html">crossJoin</a></code>, <code><a href="cube.html">cube</a></code>,
+<code><a href="dapplyCollect.html">dapplyCollect</a></code>, <code><a href="dapply.html">dapply</a></code>,
+<code><a href="describe.html">describe</a></code>, <code><a href="dim.html">dim</a></code>,
+<code><a href="distinct.html">distinct</a></code>, <code><a href="dropDuplicates.html">dropDuplicates</a></code>,
+<code><a href="nafunctions.html">dropna</a></code>, <code><a href="drop.html">drop</a></code>,
+<code><a href="dtypes.html">dtypes</a></code>, <code><a href="exceptAll.html">exceptAll</a></code>,
+<code><a href="except.html">except</a></code>, <code><a href="explain.html">explain</a></code>,
+<code><a href="filter.html">filter</a></code>, <code><a href="first.html">first</a></code>,
+<code><a href="gapplyCollect.html">gapplyCollect</a></code>, <code><a href="gapply.html">gapply</a></code>,
+<code><a href="getNumPartitions.html">getNumPartitions</a></code>, <code><a href="groupBy.html">group_by</a></code>,
+<code><a href="head.html">head</a></code>, <code><a href="hint.html">hint</a></code>,
+<code><a href="histogram.html">histogram</a></code>, <code><a href="insertInto.html">insertInto</a></code>,
+<code><a href="intersectAll.html">intersectAll</a></code>, <code><a href="intersect.html">intersect</a></code>,
+<code><a href="isLocal.html">isLocal</a></code>, <code><a href="isStreaming.html">isStreaming</a></code>,
+<code><a href="join.html">join</a></code>, <code><a href="limit.html">limit</a></code>,
+<code><a href="localCheckpoint.html">localCheckpoint</a></code>, <code><a href="merge.html">merge</a></code>,
+<code><a href="mutate.html">mutate</a></code>, <code><a href="ncol.html">ncol</a></code>,
+<code><a href="nrow.html">nrow</a></code>, <code><a href="persist.html">persist</a></code>,
+<code><a href="printSchema.html">printSchema</a></code>, <code><a href="randomSplit.html">randomSplit</a></code>,
+<code><a href="rbind.html">rbind</a></code>, <code><a href="repartitionByRange.html">repartitionByRange</a></code>,
+<code><a href="repartition.html">repartition</a></code>, <code><a href="rollup.html">rollup</a></code>,
+<code><a href="sample.html">sample</a></code>, <code><a href="saveAsTable.html">saveAsTable</a></code>,
+<code><a href="schema.html">schema</a></code>, <code><a href="selectExpr.html">selectExpr</a></code>,
+<code><a href="select.html">select</a></code>, <code><a href="showDF.html">showDF</a></code>,
+<code><a href="show.html">show</a></code>, <code><a href="storageLevel.html">storageLevel</a></code>,
+<code><a href="str.html">str</a></code>, <code><a href="subset.html">subset</a></code>,
+<code><a href="summary.html">summary</a></code>, <code><a href="take.html">take</a></code>,
+<code><a href="toJSON.html">toJSON</a></code>, <code><a href="unionByName.html">unionByName</a></code>,
+<code><a href="union.html">union</a></code>, <code><a href="unpersist.html">unpersist</a></code>,
+<code><a href="withColumn.html">withColumn</a></code>, <code><a href="withWatermark.html">withWatermark</a></code>,
+<code><a href="with.html">with</a></code>, <code><a href="write.df.html">write.df</a></code>,
+<code><a href="write.jdbc.html">write.jdbc</a></code>, <code><a href="write.json.html">write.json</a></code>,
+<code><a href="write.orc.html">write.orc</a></code>, <code><a href="write.parquet.html">write.parquet</a></code>,
+<code><a href="write.stream.html">write.stream</a></code>, <code><a href="write.text.html">write.text</a></code>
+</p>
+
+
+<h3>Examples</h3>
+
+<pre><code class="r">## Not run: 
+##D sparkR.session()
+##D path &lt;- &quot;path/to/file.json&quot;
+##D df &lt;- read.json(path)
+##D newDF &lt;- withColumnRenamed(df, &quot;col1&quot;, &quot;newCol1&quot;)
+## End(Not run)
+## Not run: 
+##D sparkR.session()
+##D path &lt;- &quot;path/to/file.json&quot;
+##D df &lt;- read.json(path)
+##D newDF &lt;- rename(df, col1 = df$newCol1)
+## End(Not run)
+</code></pre>
+
+
+<hr /><div style="text-align: center;">[Package <em>SparkR</em> version 2.4.0 <a href="00Index.html">Index</a>]</div>
+</body></html>


---------------------------------------------------------------------
To unsubscribe, e-mail: commits-unsubscribe@spark.apache.org
For additional commands, e-mail: commits-help@spark.apache.org


Mime
View raw message