spark-commits mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From hol...@apache.org
Subject [01/51] [partial] spark-website git commit: Add 2.1.2 docs
Date Tue, 17 Oct 2017 20:17:57 GMT
Repository: spark-website
Updated Branches:
  refs/heads/asf-site 0490125a8 -> a6155a89d


http://git-wip-us.apache.org/repos/asf/spark-website/blob/a6155a89/site/docs/2.1.2/api/java/org/apache/spark/api/r/RRDD.html
----------------------------------------------------------------------
diff --git a/site/docs/2.1.2/api/java/org/apache/spark/api/r/RRDD.html b/site/docs/2.1.2/api/java/org/apache/spark/api/r/RRDD.html
new file mode 100644
index 0000000..19fea40
--- /dev/null
+++ b/site/docs/2.1.2/api/java/org/apache/spark/api/r/RRDD.html
@@ -0,0 +1,2158 @@
+<!DOCTYPE HTML PUBLIC "-//W3C//DTD HTML 4.01 Transitional//EN" "http://www.w3.org/TR/html4/loose.dtd">
+<!-- NewPage -->
+<html lang="en">
+<head>
+<!-- Generated by javadoc (1.7.0_151) on Mon Oct 02 14:47:17 PDT 2017 -->
+<title>RRDD (Spark 2.1.2 JavaDoc)</title>
+<meta name="date" content="2017-10-02">
+<link rel="stylesheet" type="text/css" href="../../../../../stylesheet.css" title="Style">
+</head>
+<body>
+<script type="text/javascript"><!--
+    try {
+        if (location.href.indexOf('is-external=true') == -1) {
+            parent.document.title="RRDD (Spark 2.1.2 JavaDoc)";
+        }
+    }
+    catch(err) {
+    }
+//-->
+</script>
+<noscript>
+<div>JavaScript is disabled on your browser.</div>
+</noscript>
+<!-- ========= START OF TOP NAVBAR ======= -->
+<div class="topNav"><a name="navbar_top">
+<!--   -->
+</a><a href="#skip-navbar_top" title="Skip navigation links"></a><a name="navbar_top_firstrow">
+<!--   -->
+</a>
+<ul class="navList" title="Navigation">
+<li><a href="../../../../../overview-summary.html">Overview</a></li>
+<li><a href="package-summary.html">Package</a></li>
+<li class="navBarCell1Rev">Class</li>
+<li><a href="package-tree.html">Tree</a></li>
+<li><a href="../../../../../deprecated-list.html">Deprecated</a></li>
+<li><a href="../../../../../index-all.html">Index</a></li>
+<li><a href="../../../../../help-doc.html">Help</a></li>
+</ul>
+</div>
+<div class="subNav">
+<ul class="navList">
+<li><a href="../../../../../org/apache/spark/api/r/PairwiseRRDD.html" title="class in org.apache.spark.api.r"><span class="strong">Prev Class</span></a></li>
+<li><a href="../../../../../org/apache/spark/api/r/RRunnerModes.html" title="class in org.apache.spark.api.r"><span class="strong">Next Class</span></a></li>
+</ul>
+<ul class="navList">
+<li><a href="../../../../../index.html?org/apache/spark/api/r/RRDD.html" target="_top">Frames</a></li>
+<li><a href="RRDD.html" target="_top">No Frames</a></li>
+</ul>
+<ul class="navList" id="allclasses_navbar_top">
+<li><a href="../../../../../allclasses-noframe.html">All Classes</a></li>
+</ul>
+<div>
+<script type="text/javascript"><!--
+  allClassesLink = document.getElementById("allclasses_navbar_top");
+  if(window==top) {
+    allClassesLink.style.display = "block";
+  }
+  else {
+    allClassesLink.style.display = "none";
+  }
+  //-->
+</script>
+</div>
+<div>
+<ul class="subNavList">
+<li>Summary:&nbsp;</li>
+<li>Nested&nbsp;|&nbsp;</li>
+<li>Field&nbsp;|&nbsp;</li>
+<li><a href="#constructor_summary">Constr</a>&nbsp;|&nbsp;</li>
+<li><a href="#method_summary">Method</a></li>
+</ul>
+<ul class="subNavList">
+<li>Detail:&nbsp;</li>
+<li>Field&nbsp;|&nbsp;</li>
+<li><a href="#constructor_detail">Constr</a>&nbsp;|&nbsp;</li>
+<li><a href="#method_detail">Method</a></li>
+</ul>
+</div>
+<a name="skip-navbar_top">
+<!--   -->
+</a></div>
+<!-- ========= END OF TOP NAVBAR ========= -->
+<!-- ======== START OF CLASS DATA ======== -->
+<div class="header">
+<div class="subTitle">org.apache.spark.api.r</div>
+<h2 title="Class RRDD" class="title">Class RRDD&lt;T&gt;</h2>
+</div>
+<div class="contentContainer">
+<ul class="inheritance">
+<li>Object</li>
+<li>
+<ul class="inheritance">
+<li><a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">org.apache.spark.rdd.RDD</a>&lt;U&gt;</li>
+<li>
+<ul class="inheritance">
+<li><a href="../../../../../org/apache/spark/api/r/BaseRRDD.html" title="class in org.apache.spark.api.r">org.apache.spark.api.r.BaseRRDD</a>&lt;T,byte[]&gt;</li>
+<li>
+<ul class="inheritance">
+<li>org.apache.spark.api.r.RRDD&lt;T&gt;</li>
+</ul>
+</li>
+</ul>
+</li>
+</ul>
+</li>
+</ul>
+<div class="description">
+<ul class="blockList">
+<li class="blockList">
+<dl>
+<dt>All Implemented Interfaces:</dt>
+<dd>java.io.Serializable</dd>
+</dl>
+<hr>
+<br>
+<pre>public class <span class="strong">RRDD&lt;T&gt;</span>
+extends <a href="../../../../../org/apache/spark/api/r/BaseRRDD.html" title="class in org.apache.spark.api.r">BaseRRDD</a>&lt;T,byte[]&gt;</pre>
+<div class="block">An RDD that stores serialized R objects as Array[Byte].</div>
+<dl><dt><span class="strong">See Also:</span></dt><dd><a href="../../../../../serialized-form.html#org.apache.spark.api.r.RRDD">Serialized Form</a></dd></dl>
+</li>
+</ul>
+</div>
+<div class="summary">
+<ul class="blockList">
+<li class="blockList">
+<!-- ======== CONSTRUCTOR SUMMARY ======== -->
+<ul class="blockList">
+<li class="blockList"><a name="constructor_summary">
+<!--   -->
+</a>
+<h3>Constructor Summary</h3>
+<table class="overviewSummary" border="0" cellpadding="3" cellspacing="0" summary="Constructor Summary table, listing constructors, and an explanation">
+<caption><span>Constructors</span><span class="tabEnd">&nbsp;</span></caption>
+<tr>
+<th class="colOne" scope="col">Constructor and Description</th>
+</tr>
+<tr class="altColor">
+<td class="colOne"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#RRDD(org.apache.spark.rdd.RDD,%20byte[],%20java.lang.String,%20java.lang.String,%20byte[],%20java.lang.Object[],%20scala.reflect.ClassTag)">RRDD</a></strong>(<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;<a href="../../../../../org/apache/spark/api/r/RRDD.html" title="type parameter in RRDD">T</a>&gt;&nbsp;parent,
+    byte[]&nbsp;func,
+    String&nbsp;deserializer,
+    String&nbsp;serializer,
+    byte[]&nbsp;packageNames,
+    Object[]&nbsp;broadcastVars,
+    scala.reflect.ClassTag&lt;<a href="../../../../../org/apache/spark/api/r/RRDD.html" title="type parameter in RRDD">T</a>&gt;&nbsp;evidence$4)</code>&nbsp;</td>
+</tr>
+</table>
+</li>
+</ul>
+<!-- ========== METHOD SUMMARY =========== -->
+<ul class="blockList">
+<li class="blockList"><a name="method_summary">
+<!--   -->
+</a>
+<h3>Method Summary</h3>
+<table class="overviewSummary" border="0" cellpadding="3" cellspacing="0" summary="Method Summary table, listing methods, and an explanation">
+<caption><span>Methods</span><span class="tabEnd">&nbsp;</span></caption>
+<tr>
+<th class="colFirst" scope="col">Modifier and Type</th>
+<th class="colLast" scope="col">Method and Description</th>
+</tr>
+<tr class="altColor">
+<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#$plus$plus(org.apache.spark.rdd.RDD)">$plus$plus</a></strong>(<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;&nbsp;other)</code>&nbsp;</td>
+</tr>
+<tr class="rowColor">
+<td class="colFirst"><code>static &lt;U&gt;&nbsp;U</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#aggregate(U,%20scala.Function2,%20scala.Function2,%20scala.reflect.ClassTag)">aggregate</a></strong>(U&nbsp;zeroValue,
+         scala.Function2&lt;U,T,U&gt;&nbsp;seqOp,
+         scala.Function2&lt;U,U,U&gt;&nbsp;combOp,
+         scala.reflect.ClassTag&lt;U&gt;&nbsp;evidence$30)</code>&nbsp;</td>
+</tr>
+<tr class="altColor">
+<td class="colFirst"><code><a href="../../../../../org/apache/spark/api/java/JavaRDD.html" title="class in org.apache.spark.api.java">JavaRDD</a>&lt;byte[]&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#asJavaRDD()">asJavaRDD</a></strong>()</code>&nbsp;</td>
+</tr>
+<tr class="rowColor">
+<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#cache()">cache</a></strong>()</code>&nbsp;</td>
+</tr>
+<tr class="altColor">
+<td class="colFirst"><code>static &lt;U&gt;&nbsp;<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;scala.Tuple2&lt;T,U&gt;&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#cartesian(org.apache.spark.rdd.RDD,%20scala.reflect.ClassTag)">cartesian</a></strong>(<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;U&gt;&nbsp;other,
+         scala.reflect.ClassTag&lt;U&gt;&nbsp;evidence$5)</code>&nbsp;</td>
+</tr>
+<tr class="rowColor">
+<td class="colFirst"><code>static void</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#checkpoint()">checkpoint</a></strong>()</code>&nbsp;</td>
+</tr>
+<tr class="altColor">
+<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#coalesce(int,%20boolean,%20scala.Option,%20scala.math.Ordering)">coalesce</a></strong>(int&nbsp;numPartitions,
+        boolean&nbsp;shuffle,
+        scala.Option&lt;<a href="../../../../../org/apache/spark/rdd/PartitionCoalescer.html" title="interface in org.apache.spark.rdd">PartitionCoalescer</a>&gt;&nbsp;partitionCoalescer,
+        scala.math.Ordering&lt;T&gt;&nbsp;ord)</code>&nbsp;</td>
+</tr>
+<tr class="rowColor">
+<td class="colFirst"><code>static boolean</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#coalesce$default$2()">coalesce$default$2</a></strong>()</code>&nbsp;</td>
+</tr>
+<tr class="altColor">
+<td class="colFirst"><code>static scala.Option&lt;<a href="../../../../../org/apache/spark/rdd/PartitionCoalescer.html" title="interface in org.apache.spark.rdd">PartitionCoalescer</a>&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#coalesce$default$3()">coalesce$default$3</a></strong>()</code>&nbsp;</td>
+</tr>
+<tr class="rowColor">
+<td class="colFirst"><code>static scala.math.Ordering&lt;T&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#coalesce$default$4(int,%20boolean,%20scala.Option)">coalesce$default$4</a></strong>(int&nbsp;numPartitions,
+                  boolean&nbsp;shuffle,
+                  scala.Option&lt;<a href="../../../../../org/apache/spark/rdd/PartitionCoalescer.html" title="interface in org.apache.spark.rdd">PartitionCoalescer</a>&gt;&nbsp;partitionCoalescer)</code>&nbsp;</td>
+</tr>
+<tr class="altColor">
+<td class="colFirst"><code>static Object</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#collect()">collect</a></strong>()</code>&nbsp;</td>
+</tr>
+<tr class="rowColor">
+<td class="colFirst"><code>static &lt;U&gt;&nbsp;<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;U&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#collect(scala.PartialFunction,%20scala.reflect.ClassTag)">collect</a></strong>(scala.PartialFunction&lt;T,U&gt;&nbsp;f,
+       scala.reflect.ClassTag&lt;U&gt;&nbsp;evidence$29)</code>&nbsp;</td>
+</tr>
+<tr class="altColor">
+<td class="colFirst"><code>static scala.collection.Iterator&lt;U&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#compute(org.apache.spark.Partition,%20org.apache.spark.TaskContext)">compute</a></strong>(<a href="../../../../../org/apache/spark/Partition.html" title="interface in org.apache.spark">Partition</a>&nbsp;partition,
+       <a href="../../../../../org/apache/spark/TaskContext.html" title="class in org.apache.spark">TaskContext</a>&nbsp;context)</code>&nbsp;</td>
+</tr>
+<tr class="rowColor">
+<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/SparkContext.html" title="class in org.apache.spark">SparkContext</a></code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#context()">context</a></strong>()</code>&nbsp;</td>
+</tr>
+<tr class="altColor">
+<td class="colFirst"><code>static long</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#count()">count</a></strong>()</code>&nbsp;</td>
+</tr>
+<tr class="rowColor">
+<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/partial/PartialResult.html" title="class in org.apache.spark.partial">PartialResult</a>&lt;<a href="../../../../../org/apache/spark/partial/BoundedDouble.html" title="class in org.apache.spark.partial">BoundedDouble</a>&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#countApprox(long,%20double)">countApprox</a></strong>(long&nbsp;timeout,
+           double&nbsp;confidence)</code>&nbsp;</td>
+</tr>
+<tr class="altColor">
+<td class="colFirst"><code>static double</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#countApprox$default$2()">countApprox$default$2</a></strong>()</code>&nbsp;</td>
+</tr>
+<tr class="rowColor">
+<td class="colFirst"><code>static long</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#countApproxDistinct(double)">countApproxDistinct</a></strong>(double&nbsp;relativeSD)</code>&nbsp;</td>
+</tr>
+<tr class="altColor">
+<td class="colFirst"><code>static long</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#countApproxDistinct(int,%20int)">countApproxDistinct</a></strong>(int&nbsp;p,
+                   int&nbsp;sp)</code>&nbsp;</td>
+</tr>
+<tr class="rowColor">
+<td class="colFirst"><code>static double</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#countApproxDistinct$default$1()">countApproxDistinct$default$1</a></strong>()</code>&nbsp;</td>
+</tr>
+<tr class="altColor">
+<td class="colFirst"><code>static scala.collection.Map&lt;T,Object&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#countByValue(scala.math.Ordering)">countByValue</a></strong>(scala.math.Ordering&lt;T&gt;&nbsp;ord)</code>&nbsp;</td>
+</tr>
+<tr class="rowColor">
+<td class="colFirst"><code>static scala.math.Ordering&lt;T&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#countByValue$default$1()">countByValue$default$1</a></strong>()</code>&nbsp;</td>
+</tr>
+<tr class="altColor">
+<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/partial/PartialResult.html" title="class in org.apache.spark.partial">PartialResult</a>&lt;scala.collection.Map&lt;T,<a href="../../../../../org/apache/spark/partial/BoundedDouble.html" title="class in org.apache.spark.partial">BoundedDouble</a>&gt;&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#countByValueApprox(long,%20double,%20scala.math.Ordering)">countByValueApprox</a></strong>(long&nbsp;timeout,
+                  double&nbsp;confidence,
+                  scala.math.Ordering&lt;T&gt;&nbsp;ord)</code>&nbsp;</td>
+</tr>
+<tr class="rowColor">
+<td class="colFirst"><code>static double</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#countByValueApprox$default$2()">countByValueApprox$default$2</a></strong>()</code>&nbsp;</td>
+</tr>
+<tr class="altColor">
+<td class="colFirst"><code>static scala.math.Ordering&lt;T&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#countByValueApprox$default$3(long,%20double)">countByValueApprox$default$3</a></strong>(long&nbsp;timeout,
+                            double&nbsp;confidence)</code>&nbsp;</td>
+</tr>
+<tr class="rowColor">
+<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/api/java/JavaRDD.html" title="class in org.apache.spark.api.java">JavaRDD</a>&lt;byte[]&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#createRDDFromArray(org.apache.spark.api.java.JavaSparkContext,%20byte[][])">createRDDFromArray</a></strong>(<a href="../../../../../org/apache/spark/api/java/JavaSparkContext.html" title="class in org.apache.spark.api.java">JavaSparkContext</a>&nbsp;jsc,
+                  byte[][]&nbsp;arr)</code>
+<div class="block">Create an RRDD given a sequence of byte arrays.</div>
+</td>
+</tr>
+<tr class="altColor">
+<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/api/java/JavaRDD.html" title="class in org.apache.spark.api.java">JavaRDD</a>&lt;byte[]&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#createRDDFromFile(org.apache.spark.api.java.JavaSparkContext,%20java.lang.String,%20int)">createRDDFromFile</a></strong>(<a href="../../../../../org/apache/spark/api/java/JavaSparkContext.html" title="class in org.apache.spark.api.java">JavaSparkContext</a>&nbsp;jsc,
+                 String&nbsp;fileName,
+                 int&nbsp;parallelism)</code>
+<div class="block">Create an RRDD given a temporary file name.</div>
+</td>
+</tr>
+<tr class="rowColor">
+<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/api/java/JavaSparkContext.html" title="class in org.apache.spark.api.java">JavaSparkContext</a></code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#createSparkContext(java.lang.String,%20java.lang.String,%20java.lang.String,%20java.lang.String[],%20java.util.Map,%20java.util.Map)">createSparkContext</a></strong>(String&nbsp;master,
+                  String&nbsp;appName,
+                  String&nbsp;sparkHome,
+                  String[]&nbsp;jars,
+                  java.util.Map&lt;Object,Object&gt;&nbsp;sparkEnvirMap,
+                  java.util.Map&lt;Object,Object&gt;&nbsp;sparkExecutorEnvMap)</code>&nbsp;</td>
+</tr>
+<tr class="altColor">
+<td class="colFirst"><code>static scala.collection.Seq&lt;<a href="../../../../../org/apache/spark/Dependency.html" title="class in org.apache.spark">Dependency</a>&lt;?&gt;&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#dependencies()">dependencies</a></strong>()</code>&nbsp;</td>
+</tr>
+<tr class="rowColor">
+<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#distinct()">distinct</a></strong>()</code>&nbsp;</td>
+</tr>
+<tr class="altColor">
+<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#distinct(int,%20scala.math.Ordering)">distinct</a></strong>(int&nbsp;numPartitions,
+        scala.math.Ordering&lt;T&gt;&nbsp;ord)</code>&nbsp;</td>
+</tr>
+<tr class="rowColor">
+<td class="colFirst"><code>static scala.math.Ordering&lt;T&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#distinct$default$2(int)">distinct$default$2</a></strong>(int&nbsp;numPartitions)</code>&nbsp;</td>
+</tr>
+<tr class="altColor">
+<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#filter(scala.Function1)">filter</a></strong>(scala.Function1&lt;T,Object&gt;&nbsp;f)</code>&nbsp;</td>
+</tr>
+<tr class="rowColor">
+<td class="colFirst"><code>static T</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#first()">first</a></strong>()</code>&nbsp;</td>
+</tr>
+<tr class="altColor">
+<td class="colFirst"><code>static &lt;U&gt;&nbsp;<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;U&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#flatMap(scala.Function1,%20scala.reflect.ClassTag)">flatMap</a></strong>(scala.Function1&lt;T,scala.collection.TraversableOnce&lt;U&gt;&gt;&nbsp;f,
+       scala.reflect.ClassTag&lt;U&gt;&nbsp;evidence$4)</code>&nbsp;</td>
+</tr>
+<tr class="rowColor">
+<td class="colFirst"><code>static T</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#fold(T,%20scala.Function2)">fold</a></strong>(T&nbsp;zeroValue,
+    scala.Function2&lt;T,T,T&gt;&nbsp;op)</code>&nbsp;</td>
+</tr>
+<tr class="altColor">
+<td class="colFirst"><code>static void</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#foreach(scala.Function1)">foreach</a></strong>(scala.Function1&lt;T,scala.runtime.BoxedUnit&gt;&nbsp;f)</code>&nbsp;</td>
+</tr>
+<tr class="rowColor">
+<td class="colFirst"><code>static void</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#foreachPartition(scala.Function1)">foreachPartition</a></strong>(scala.Function1&lt;scala.collection.Iterator&lt;T&gt;,scala.runtime.BoxedUnit&gt;&nbsp;f)</code>&nbsp;</td>
+</tr>
+<tr class="altColor">
+<td class="colFirst"><code>static scala.Option&lt;String&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#getCheckpointFile()">getCheckpointFile</a></strong>()</code>&nbsp;</td>
+</tr>
+<tr class="rowColor">
+<td class="colFirst"><code>static int</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#getNumPartitions()">getNumPartitions</a></strong>()</code>&nbsp;</td>
+</tr>
+<tr class="altColor">
+<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/Partition.html" title="interface in org.apache.spark">Partition</a>[]</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#getPartitions()">getPartitions</a></strong>()</code>&nbsp;</td>
+</tr>
+<tr class="rowColor">
+<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/storage/StorageLevel.html" title="class in org.apache.spark.storage">StorageLevel</a></code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#getStorageLevel()">getStorageLevel</a></strong>()</code>&nbsp;</td>
+</tr>
+<tr class="altColor">
+<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;Object&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#glom()">glom</a></strong>()</code>&nbsp;</td>
+</tr>
+<tr class="rowColor">
+<td class="colFirst"><code>static &lt;K&gt;&nbsp;<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;scala.Tuple2&lt;K,scala.collection.Iterable&lt;T&gt;&gt;&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#groupBy(scala.Function1,%20scala.reflect.ClassTag)">groupBy</a></strong>(scala.Function1&lt;T,K&gt;&nbsp;f,
+       scala.reflect.ClassTag&lt;K&gt;&nbsp;kt)</code>&nbsp;</td>
+</tr>
+<tr class="altColor">
+<td class="colFirst"><code>static &lt;K&gt;&nbsp;<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;scala.Tuple2&lt;K,scala.collection.Iterable&lt;T&gt;&gt;&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#groupBy(scala.Function1,%20int,%20scala.reflect.ClassTag)">groupBy</a></strong>(scala.Function1&lt;T,K&gt;&nbsp;f,
+       int&nbsp;numPartitions,
+       scala.reflect.ClassTag&lt;K&gt;&nbsp;kt)</code>&nbsp;</td>
+</tr>
+<tr class="rowColor">
+<td class="colFirst"><code>static &lt;K&gt;&nbsp;<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;scala.Tuple2&lt;K,scala.collection.Iterable&lt;T&gt;&gt;&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#groupBy(scala.Function1,%20org.apache.spark.Partitioner,%20scala.reflect.ClassTag,%20scala.math.Ordering)">groupBy</a></strong>(scala.Function1&lt;T,K&gt;&nbsp;f,
+       <a href="../../../../../org/apache/spark/Partitioner.html" title="class in org.apache.spark">Partitioner</a>&nbsp;p,
+       scala.reflect.ClassTag&lt;K&gt;&nbsp;kt,
+       scala.math.Ordering&lt;K&gt;&nbsp;ord)</code>&nbsp;</td>
+</tr>
+<tr class="altColor">
+<td class="colFirst"><code>static &lt;K&gt;&nbsp;scala.runtime.Null$</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#groupBy$default$4(scala.Function1,%20org.apache.spark.Partitioner)">groupBy$default$4</a></strong>(scala.Function1&lt;T,K&gt;&nbsp;f,
+                 <a href="../../../../../org/apache/spark/Partitioner.html" title="class in org.apache.spark">Partitioner</a>&nbsp;p)</code>&nbsp;</td>
+</tr>
+<tr class="rowColor">
+<td class="colFirst"><code>static int</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#id()">id</a></strong>()</code>&nbsp;</td>
+</tr>
+<tr class="altColor">
+<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#intersection(org.apache.spark.rdd.RDD)">intersection</a></strong>(<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;&nbsp;other)</code>&nbsp;</td>
+</tr>
+<tr class="rowColor">
+<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#intersection(org.apache.spark.rdd.RDD,%20int)">intersection</a></strong>(<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;&nbsp;other,
+            int&nbsp;numPartitions)</code>&nbsp;</td>
+</tr>
+<tr class="altColor">
+<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#intersection(org.apache.spark.rdd.RDD,%20org.apache.spark.Partitioner,%20scala.math.Ordering)">intersection</a></strong>(<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;&nbsp;other,
+            <a href="../../../../../org/apache/spark/Partitioner.html" title="class in org.apache.spark">Partitioner</a>&nbsp;partitioner,
+            scala.math.Ordering&lt;T&gt;&nbsp;ord)</code>&nbsp;</td>
+</tr>
+<tr class="rowColor">
+<td class="colFirst"><code>static scala.math.Ordering&lt;T&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#intersection$default$3(org.apache.spark.rdd.RDD,%20org.apache.spark.Partitioner)">intersection$default$3</a></strong>(<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;&nbsp;other,
+                      <a href="../../../../../org/apache/spark/Partitioner.html" title="class in org.apache.spark">Partitioner</a>&nbsp;partitioner)</code>&nbsp;</td>
+</tr>
+<tr class="altColor">
+<td class="colFirst"><code>static boolean</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#isCheckpointed()">isCheckpointed</a></strong>()</code>&nbsp;</td>
+</tr>
+<tr class="rowColor">
+<td class="colFirst"><code>static boolean</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#isEmpty()">isEmpty</a></strong>()</code>&nbsp;</td>
+</tr>
+<tr class="altColor">
+<td class="colFirst"><code>static scala.collection.Iterator&lt;T&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#iterator(org.apache.spark.Partition,%20org.apache.spark.TaskContext)">iterator</a></strong>(<a href="../../../../../org/apache/spark/Partition.html" title="interface in org.apache.spark">Partition</a>&nbsp;split,
+        <a href="../../../../../org/apache/spark/TaskContext.html" title="class in org.apache.spark">TaskContext</a>&nbsp;context)</code>&nbsp;</td>
+</tr>
+<tr class="rowColor">
+<td class="colFirst"><code>static &lt;K&gt;&nbsp;<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;scala.Tuple2&lt;K,T&gt;&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#keyBy(scala.Function1)">keyBy</a></strong>(scala.Function1&lt;T,K&gt;&nbsp;f)</code>&nbsp;</td>
+</tr>
+<tr class="altColor">
+<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#localCheckpoint()">localCheckpoint</a></strong>()</code>&nbsp;</td>
+</tr>
+<tr class="rowColor">
+<td class="colFirst"><code>static &lt;U&gt;&nbsp;<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;U&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#map(scala.Function1,%20scala.reflect.ClassTag)">map</a></strong>(scala.Function1&lt;T,U&gt;&nbsp;f,
+   scala.reflect.ClassTag&lt;U&gt;&nbsp;evidence$3)</code>&nbsp;</td>
+</tr>
+<tr class="altColor">
+<td class="colFirst"><code>static &lt;U&gt;&nbsp;<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;U&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#mapPartitions(scala.Function1,%20boolean,%20scala.reflect.ClassTag)">mapPartitions</a></strong>(scala.Function1&lt;scala.collection.Iterator&lt;T&gt;,scala.collection.Iterator&lt;U&gt;&gt;&nbsp;f,
+             boolean&nbsp;preservesPartitioning,
+             scala.reflect.ClassTag&lt;U&gt;&nbsp;evidence$6)</code>&nbsp;</td>
+</tr>
+<tr class="rowColor">
+<td class="colFirst"><code>static &lt;U&gt;&nbsp;boolean</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#mapPartitions$default$2()">mapPartitions$default$2</a></strong>()</code>&nbsp;</td>
+</tr>
+<tr class="altColor">
+<td class="colFirst"><code>static &lt;U&gt;&nbsp;boolean</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#mapPartitionsInternal$default$2()">mapPartitionsInternal$default$2</a></strong>()</code>&nbsp;</td>
+</tr>
+<tr class="rowColor">
+<td class="colFirst"><code>static &lt;U&gt;&nbsp;<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;U&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#mapPartitionsWithIndex(scala.Function2,%20boolean,%20scala.reflect.ClassTag)">mapPartitionsWithIndex</a></strong>(scala.Function2&lt;Object,scala.collection.Iterator&lt;T&gt;,scala.collection.Iterator&lt;U&gt;&gt;&nbsp;f,
+                      boolean&nbsp;preservesPartitioning,
+                      scala.reflect.ClassTag&lt;U&gt;&nbsp;evidence$9)</code>&nbsp;</td>
+</tr>
+<tr class="altColor">
+<td class="colFirst"><code>static &lt;U&gt;&nbsp;boolean</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#mapPartitionsWithIndex$default$2()">mapPartitionsWithIndex$default$2</a></strong>()</code>&nbsp;</td>
+</tr>
+<tr class="rowColor">
+<td class="colFirst"><code>static &lt;U&gt;&nbsp;boolean</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#mapPartitionsWithIndexInternal$default$2()">mapPartitionsWithIndexInternal$default$2</a></strong>()</code>&nbsp;</td>
+</tr>
+<tr class="altColor">
+<td class="colFirst"><code>static T</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#max(scala.math.Ordering)">max</a></strong>(scala.math.Ordering&lt;T&gt;&nbsp;ord)</code>&nbsp;</td>
+</tr>
+<tr class="rowColor">
+<td class="colFirst"><code>static T</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#min(scala.math.Ordering)">min</a></strong>(scala.math.Ordering&lt;T&gt;&nbsp;ord)</code>&nbsp;</td>
+</tr>
+<tr class="altColor">
+<td class="colFirst"><code>static void</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#name_$eq(java.lang.String)">name_$eq</a></strong>(String&nbsp;x$1)</code>&nbsp;</td>
+</tr>
+<tr class="rowColor">
+<td class="colFirst"><code>static String</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#name()">name</a></strong>()</code>&nbsp;</td>
+</tr>
+<tr class="altColor">
+<td class="colFirst"><code>static scala.Option&lt;<a href="../../../../../org/apache/spark/Partitioner.html" title="class in org.apache.spark">Partitioner</a>&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#partitioner()">partitioner</a></strong>()</code>&nbsp;</td>
+</tr>
+<tr class="rowColor">
+<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/Partition.html" title="interface in org.apache.spark">Partition</a>[]</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#partitions()">partitions</a></strong>()</code>&nbsp;</td>
+</tr>
+<tr class="altColor">
+<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#persist()">persist</a></strong>()</code>&nbsp;</td>
+</tr>
+<tr class="rowColor">
+<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#persist(org.apache.spark.storage.StorageLevel)">persist</a></strong>(<a href="../../../../../org/apache/spark/storage/StorageLevel.html" title="class in org.apache.spark.storage">StorageLevel</a>&nbsp;newLevel)</code>&nbsp;</td>
+</tr>
+<tr class="altColor">
+<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;String&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#pipe(scala.collection.Seq,%20scala.collection.Map,%20scala.Function1,%20scala.Function2,%20boolean,%20int,%20java.lang.String)">pipe</a></strong>(scala.collection.Seq&lt;String&gt;&nbsp;command,
+    scala.collection.Map&lt;String,String&gt;&nbsp;env,
+    scala.Function1&lt;scala.Function1&lt;String,scala.runtime.BoxedUnit&gt;,scala.runtime.BoxedUnit&gt;&nbsp;printPipeContext,
+    scala.Function2&lt;T,scala.Function1&lt;String,scala.runtime.BoxedUnit&gt;,scala.runtime.BoxedUnit&gt;&nbsp;printRDDElement,
+    boolean&nbsp;separateWorkingDir,
+    int&nbsp;bufferSize,
+    String&nbsp;encoding)</code>&nbsp;</td>
+</tr>
+<tr class="rowColor">
+<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;String&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#pipe(java.lang.String)">pipe</a></strong>(String&nbsp;command)</code>&nbsp;</td>
+</tr>
+<tr class="altColor">
+<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;String&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#pipe(java.lang.String,%20scala.collection.Map)">pipe</a></strong>(String&nbsp;command,
+    scala.collection.Map&lt;String,String&gt;&nbsp;env)</code>&nbsp;</td>
+</tr>
+<tr class="rowColor">
+<td class="colFirst"><code>static scala.collection.Map&lt;String,String&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#pipe$default$2()">pipe$default$2</a></strong>()</code>&nbsp;</td>
+</tr>
+<tr class="altColor">
+<td class="colFirst"><code>static scala.Function1&lt;scala.Function1&lt;String,scala.runtime.BoxedUnit&gt;,scala.runtime.BoxedUnit&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#pipe$default$3()">pipe$default$3</a></strong>()</code>&nbsp;</td>
+</tr>
+<tr class="rowColor">
+<td class="colFirst"><code>static scala.Function2&lt;T,scala.Function1&lt;String,scala.runtime.BoxedUnit&gt;,scala.runtime.BoxedUnit&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#pipe$default$4()">pipe$default$4</a></strong>()</code>&nbsp;</td>
+</tr>
+<tr class="altColor">
+<td class="colFirst"><code>static boolean</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#pipe$default$5()">pipe$default$5</a></strong>()</code>&nbsp;</td>
+</tr>
+<tr class="rowColor">
+<td class="colFirst"><code>static int</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#pipe$default$6()">pipe$default$6</a></strong>()</code>&nbsp;</td>
+</tr>
+<tr class="altColor">
+<td class="colFirst"><code>static String</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#pipe$default$7()">pipe$default$7</a></strong>()</code>&nbsp;</td>
+</tr>
+<tr class="rowColor">
+<td class="colFirst"><code>static scala.collection.Seq&lt;String&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#preferredLocations(org.apache.spark.Partition)">preferredLocations</a></strong>(<a href="../../../../../org/apache/spark/Partition.html" title="interface in org.apache.spark">Partition</a>&nbsp;split)</code>&nbsp;</td>
+</tr>
+<tr class="altColor">
+<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;[]</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#randomSplit(double[],%20long)">randomSplit</a></strong>(double[]&nbsp;weights,
+           long&nbsp;seed)</code>&nbsp;</td>
+</tr>
+<tr class="rowColor">
+<td class="colFirst"><code>static long</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#randomSplit$default$2()">randomSplit$default$2</a></strong>()</code>&nbsp;</td>
+</tr>
+<tr class="altColor">
+<td class="colFirst"><code>static T</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#reduce(scala.Function2)">reduce</a></strong>(scala.Function2&lt;T,T,T&gt;&nbsp;f)</code>&nbsp;</td>
+</tr>
+<tr class="rowColor">
+<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#repartition(int,%20scala.math.Ordering)">repartition</a></strong>(int&nbsp;numPartitions,
+           scala.math.Ordering&lt;T&gt;&nbsp;ord)</code>&nbsp;</td>
+</tr>
+<tr class="altColor">
+<td class="colFirst"><code>static scala.math.Ordering&lt;T&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#repartition$default$2(int)">repartition$default$2</a></strong>(int&nbsp;numPartitions)</code>&nbsp;</td>
+</tr>
+<tr class="rowColor">
+<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#sample(boolean,%20double,%20long)">sample</a></strong>(boolean&nbsp;withReplacement,
+      double&nbsp;fraction,
+      long&nbsp;seed)</code>&nbsp;</td>
+</tr>
+<tr class="altColor">
+<td class="colFirst"><code>static long</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#sample$default$3()">sample$default$3</a></strong>()</code>&nbsp;</td>
+</tr>
+<tr class="rowColor">
+<td class="colFirst"><code>static void</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#saveAsObjectFile(java.lang.String)">saveAsObjectFile</a></strong>(String&nbsp;path)</code>&nbsp;</td>
+</tr>
+<tr class="altColor">
+<td class="colFirst"><code>static void</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#saveAsTextFile(java.lang.String)">saveAsTextFile</a></strong>(String&nbsp;path)</code>&nbsp;</td>
+</tr>
+<tr class="rowColor">
+<td class="colFirst"><code>static void</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#saveAsTextFile(java.lang.String,%20java.lang.Class)">saveAsTextFile</a></strong>(String&nbsp;path,
+              Class&lt;? extends org.apache.hadoop.io.compress.CompressionCodec&gt;&nbsp;codec)</code>&nbsp;</td>
+</tr>
+<tr class="altColor">
+<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#setName(java.lang.String)">setName</a></strong>(String&nbsp;_name)</code>&nbsp;</td>
+</tr>
+<tr class="rowColor">
+<td class="colFirst"><code>static &lt;K&gt;&nbsp;<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#sortBy(scala.Function1,%20boolean,%20int,%20scala.math.Ordering,%20scala.reflect.ClassTag)">sortBy</a></strong>(scala.Function1&lt;T,K&gt;&nbsp;f,
+      boolean&nbsp;ascending,
+      int&nbsp;numPartitions,
+      scala.math.Ordering&lt;K&gt;&nbsp;ord,
+      scala.reflect.ClassTag&lt;K&gt;&nbsp;ctag)</code>&nbsp;</td>
+</tr>
+<tr class="altColor">
+<td class="colFirst"><code>static &lt;K&gt;&nbsp;boolean</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#sortBy$default$2()">sortBy$default$2</a></strong>()</code>&nbsp;</td>
+</tr>
+<tr class="rowColor">
+<td class="colFirst"><code>static &lt;K&gt;&nbsp;int</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#sortBy$default$3()">sortBy$default$3</a></strong>()</code>&nbsp;</td>
+</tr>
+<tr class="altColor">
+<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/SparkContext.html" title="class in org.apache.spark">SparkContext</a></code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#sparkContext()">sparkContext</a></strong>()</code>&nbsp;</td>
+</tr>
+<tr class="rowColor">
+<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#subtract(org.apache.spark.rdd.RDD)">subtract</a></strong>(<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;&nbsp;other)</code>&nbsp;</td>
+</tr>
+<tr class="altColor">
+<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#subtract(org.apache.spark.rdd.RDD,%20int)">subtract</a></strong>(<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;&nbsp;other,
+        int&nbsp;numPartitions)</code>&nbsp;</td>
+</tr>
+<tr class="rowColor">
+<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#subtract(org.apache.spark.rdd.RDD,%20org.apache.spark.Partitioner,%20scala.math.Ordering)">subtract</a></strong>(<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;&nbsp;other,
+        <a href="../../../../../org/apache/spark/Partitioner.html" title="class in org.apache.spark">Partitioner</a>&nbsp;p,
+        scala.math.Ordering&lt;T&gt;&nbsp;ord)</code>&nbsp;</td>
+</tr>
+<tr class="altColor">
+<td class="colFirst"><code>static scala.math.Ordering&lt;T&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#subtract$default$3(org.apache.spark.rdd.RDD,%20org.apache.spark.Partitioner)">subtract$default$3</a></strong>(<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;&nbsp;other,
+                  <a href="../../../../../org/apache/spark/Partitioner.html" title="class in org.apache.spark">Partitioner</a>&nbsp;p)</code>&nbsp;</td>
+</tr>
+<tr class="rowColor">
+<td class="colFirst"><code>static Object</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#take(int)">take</a></strong>(int&nbsp;num)</code>&nbsp;</td>
+</tr>
+<tr class="altColor">
+<td class="colFirst"><code>static Object</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#takeOrdered(int,%20scala.math.Ordering)">takeOrdered</a></strong>(int&nbsp;num,
+           scala.math.Ordering&lt;T&gt;&nbsp;ord)</code>&nbsp;</td>
+</tr>
+<tr class="rowColor">
+<td class="colFirst"><code>static Object</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#takeSample(boolean,%20int,%20long)">takeSample</a></strong>(boolean&nbsp;withReplacement,
+          int&nbsp;num,
+          long&nbsp;seed)</code>&nbsp;</td>
+</tr>
+<tr class="altColor">
+<td class="colFirst"><code>static long</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#takeSample$default$3()">takeSample$default$3</a></strong>()</code>&nbsp;</td>
+</tr>
+<tr class="rowColor">
+<td class="colFirst"><code>static String</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#toDebugString()">toDebugString</a></strong>()</code>&nbsp;</td>
+</tr>
+<tr class="altColor">
+<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/api/java/JavaRDD.html" title="class in org.apache.spark.api.java">JavaRDD</a>&lt;T&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#toJavaRDD()">toJavaRDD</a></strong>()</code>&nbsp;</td>
+</tr>
+<tr class="rowColor">
+<td class="colFirst"><code>static scala.collection.Iterator&lt;T&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#toLocalIterator()">toLocalIterator</a></strong>()</code>&nbsp;</td>
+</tr>
+<tr class="altColor">
+<td class="colFirst"><code>static Object</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#top(int,%20scala.math.Ordering)">top</a></strong>(int&nbsp;num,
+   scala.math.Ordering&lt;T&gt;&nbsp;ord)</code>&nbsp;</td>
+</tr>
+<tr class="rowColor">
+<td class="colFirst"><code>static String</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#toString()">toString</a></strong>()</code>&nbsp;</td>
+</tr>
+<tr class="altColor">
+<td class="colFirst"><code>static &lt;U&gt;&nbsp;U</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#treeAggregate(U,%20scala.Function2,%20scala.Function2,%20int,%20scala.reflect.ClassTag)">treeAggregate</a></strong>(U&nbsp;zeroValue,
+             scala.Function2&lt;U,T,U&gt;&nbsp;seqOp,
+             scala.Function2&lt;U,U,U&gt;&nbsp;combOp,
+             int&nbsp;depth,
+             scala.reflect.ClassTag&lt;U&gt;&nbsp;evidence$31)</code>&nbsp;</td>
+</tr>
+<tr class="rowColor">
+<td class="colFirst"><code>static &lt;U&gt;&nbsp;int</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#treeAggregate$default$4(U)">treeAggregate$default$4</a></strong>(U&nbsp;zeroValue)</code>&nbsp;</td>
+</tr>
+<tr class="altColor">
+<td class="colFirst"><code>static T</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#treeReduce(scala.Function2,%20int)">treeReduce</a></strong>(scala.Function2&lt;T,T,T&gt;&nbsp;f,
+          int&nbsp;depth)</code>&nbsp;</td>
+</tr>
+<tr class="rowColor">
+<td class="colFirst"><code>static int</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#treeReduce$default$2()">treeReduce$default$2</a></strong>()</code>&nbsp;</td>
+</tr>
+<tr class="altColor">
+<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#union(org.apache.spark.rdd.RDD)">union</a></strong>(<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;&nbsp;other)</code>&nbsp;</td>
+</tr>
+<tr class="rowColor">
+<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#unpersist(boolean)">unpersist</a></strong>(boolean&nbsp;blocking)</code>&nbsp;</td>
+</tr>
+<tr class="altColor">
+<td class="colFirst"><code>static boolean</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#unpersist$default$1()">unpersist$default$1</a></strong>()</code>&nbsp;</td>
+</tr>
+<tr class="rowColor">
+<td class="colFirst"><code>static &lt;U&gt;&nbsp;<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;scala.Tuple2&lt;T,U&gt;&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#zip(org.apache.spark.rdd.RDD,%20scala.reflect.ClassTag)">zip</a></strong>(<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;U&gt;&nbsp;other,
+   scala.reflect.ClassTag&lt;U&gt;&nbsp;evidence$10)</code>&nbsp;</td>
+</tr>
+<tr class="altColor">
+<td class="colFirst"><code>static &lt;B,V&gt;&nbsp;<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;V&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#zipPartitions(org.apache.spark.rdd.RDD,%20boolean,%20scala.Function2,%20scala.reflect.ClassTag,%20scala.reflect.ClassTag)">zipPartitions</a></strong>(<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;B&gt;&nbsp;rdd2,
+             boolean&nbsp;preservesPartitioning,
+             scala.Function2&lt;scala.collection.Iterator&lt;T&gt;,scala.collection.Iterator&lt;B&gt;,scala.collection.Iterator&lt;V&gt;&gt;&nbsp;f,
+             scala.reflect.ClassTag&lt;B&gt;&nbsp;evidence$11,
+             scala.reflect.ClassTag&lt;V&gt;&nbsp;evidence$12)</code>&nbsp;</td>
+</tr>
+<tr class="rowColor">
+<td class="colFirst"><code>static &lt;B,V&gt;&nbsp;<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;V&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#zipPartitions(org.apache.spark.rdd.RDD,%20scala.Function2,%20scala.reflect.ClassTag,%20scala.reflect.ClassTag)">zipPartitions</a></strong>(<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;B&gt;&nbsp;rdd2,
+             scala.Function2&lt;scala.collection.Iterator&lt;T&gt;,scala.collection.Iterator&lt;B&gt;,scala.collection.Iterator&lt;V&gt;&gt;&nbsp;f,
+             scala.reflect.ClassTag&lt;B&gt;&nbsp;evidence$13,
+             scala.reflect.ClassTag&lt;V&gt;&nbsp;evidence$14)</code>&nbsp;</td>
+</tr>
+<tr class="altColor">
+<td class="colFirst"><code>static &lt;B,C,V&gt;&nbsp;<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;V&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#zipPartitions(org.apache.spark.rdd.RDD,%20org.apache.spark.rdd.RDD,%20boolean,%20scala.Function3,%20scala.reflect.ClassTag,%20scala.reflect.ClassTag,%20scala.reflect.ClassTag)">zipPartitions</a></strong>(<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;B&gt;&nbsp;rdd2,
+             <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;C&gt;&nbsp;rdd3,
+             boolean&nbsp;preservesPartitioning,
+             scala.Function3&lt;scala.collection.Iterator&lt;T&gt;,scala.collection.Iterator&lt;B&gt;,scala.collection.Iterator&lt;C&gt;,scala.collection.Iterator&lt;V&gt;&gt;&nbsp;f,
+             scala.reflect.ClassTag&lt;B&gt;&nbsp;evidence$15,
+             scala.reflect.ClassTag&lt;C&gt;&nbsp;evidence$16,
+             scala.reflect.ClassTag&lt;V&gt;&nbsp;evidence$17)</code>&nbsp;</td>
+</tr>
+<tr class="rowColor">
+<td class="colFirst"><code>static &lt;B,C,V&gt;&nbsp;<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;V&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#zipPartitions(org.apache.spark.rdd.RDD,%20org.apache.spark.rdd.RDD,%20scala.Function3,%20scala.reflect.ClassTag,%20scala.reflect.ClassTag,%20scala.reflect.ClassTag)">zipPartitions</a></strong>(<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;B&gt;&nbsp;rdd2,
+             <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;C&gt;&nbsp;rdd3,
+             scala.Function3&lt;scala.collection.Iterator&lt;T&gt;,scala.collection.Iterator&lt;B&gt;,scala.collection.Iterator&lt;C&gt;,scala.collection.Iterator&lt;V&gt;&gt;&nbsp;f,
+             scala.reflect.ClassTag&lt;B&gt;&nbsp;evidence$18,
+             scala.reflect.ClassTag&lt;C&gt;&nbsp;evidence$19,
+             scala.reflect.ClassTag&lt;V&gt;&nbsp;evidence$20)</code>&nbsp;</td>
+</tr>
+<tr class="altColor">
+<td class="colFirst"><code>static &lt;B,C,D,V&gt;&nbsp;<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;V&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#zipPartitions(org.apache.spark.rdd.RDD,%20org.apache.spark.rdd.RDD,%20org.apache.spark.rdd.RDD,%20boolean,%20scala.Function4,%20scala.reflect.ClassTag,%20scala.reflect.ClassTag,%20scala.reflect.ClassTag,%20scala.reflect.ClassTag)">zipPartitions</a></strong>(<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;B&gt;&nbsp;rdd2,
+             <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;C&gt;&nbsp;rdd3,
+             <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;D&gt;&nbsp;rdd4,
+             boolean&nbsp;preservesPartitioning,
+             scala.Function4&lt;scala.collection.Iterator&lt;T&gt;,scala.collection.Iterator&lt;B&gt;,scala.collection.Iterator&lt;C&gt;,scala.collection.Iterator&lt;D&gt;,scala.collection.Iterator&lt;V&gt;&gt;&nbsp;f,
+             scala.reflect.ClassTag&lt;B&gt;&nbsp;evidence$21,
+             scala.reflect.ClassTag&lt;C&gt;&nbsp;evidence$22,
+             scala.reflect.ClassTag&lt;D&gt;&nbsp;evidence$23,
+             scala.reflect.ClassTag&lt;V&gt;&nbsp;evidence$24)</code>&nbsp;</td>
+</tr>
+<tr class="rowColor">
+<td class="colFirst"><code>static &lt;B,C,D,V&gt;&nbsp;<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;V&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#zipPartitions(org.apache.spark.rdd.RDD,%20org.apache.spark.rdd.RDD,%20org.apache.spark.rdd.RDD,%20scala.Function4,%20scala.reflect.ClassTag,%20scala.reflect.ClassTag,%20scala.reflect.ClassTag,%20scala.reflect.ClassTag)">zipPartitions</a></strong>(<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;B&gt;&nbsp;rdd2,
+             <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;C&gt;&nbsp;rdd3,
+             <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;D&gt;&nbsp;rdd4,
+             scala.Function4&lt;scala.collection.Iterator&lt;T&gt;,scala.collection.Iterator&lt;B&gt;,scala.collection.Iterator&lt;C&gt;,scala.collection.Iterator&lt;D&gt;,scala.collection.Iterator&lt;V&gt;&gt;&nbsp;f,
+             scala.reflect.ClassTag&lt;B&gt;&nbsp;evidence$25,
+             scala.reflect.ClassTag&lt;C&gt;&nbsp;evidence$26,
+             scala.reflect.ClassTag&lt;D&gt;&nbsp;evidence$27,
+             scala.reflect.ClassTag&lt;V&gt;&nbsp;evidence$28)</code>&nbsp;</td>
+</tr>
+<tr class="altColor">
+<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;scala.Tuple2&lt;T,Object&gt;&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#zipWithIndex()">zipWithIndex</a></strong>()</code>&nbsp;</td>
+</tr>
+<tr class="rowColor">
+<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;scala.Tuple2&lt;T,Object&gt;&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#zipWithUniqueId()">zipWithUniqueId</a></strong>()</code>&nbsp;</td>
+</tr>
+</table>
+<ul class="blockList">
+<li class="blockList"><a name="methods_inherited_from_class_org.apache.spark.api.r.BaseRRDD">
+<!--   -->
+</a>
+<h3>Methods inherited from class&nbsp;org.apache.spark.api.r.<a href="../../../../../org/apache/spark/api/r/BaseRRDD.html" title="class in org.apache.spark.api.r">BaseRRDD</a></h3>
+<code><a href="../../../../../org/apache/spark/api/r/BaseRRDD.html#compute(org.apache.spark.Partition,%20org.apache.spark.TaskContext)">compute</a>, <a href="../../../../../org/apache/spark/api/r/BaseRRDD.html#getPartitions()">getPartitions</a></code></li>
+</ul>
+<ul class="blockList">
+<li class="blockList"><a name="methods_inherited_from_class_org.apache.spark.rdd.RDD">
+<!--   -->
+</a>
+<h3>Methods inherited from class&nbsp;org.apache.spark.rdd.<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a></h3>
+<code><a href="../../../../../org/apache/spark/rdd/RDD.html#aggregate(U,%20scala.Function2,%20scala.Function2,%20scala.reflect.ClassTag)">aggregate</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#cache()">cache</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#cartesian(org.apache.spark.rdd.RDD,%20scala.reflect.ClassTag)">cartesian</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#checkpoint()">checkpoint</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#coalesce(int,%20boolean,%20scala.Option,%20scala.math.Ordering)">coalesce</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#collect()">collect</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#collect(scala.PartialFunction,%20scala.reflect.ClassTag)">collect</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#context()">context</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#count()">count</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#countApprox(
 long,%20double)">countApprox</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#countApproxDistinct(double)">countApproxDistinct</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#countApproxDistinct(int,%20int)">countApproxDistinct</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#countByValue(scala.math.Ordering)">countByValue</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#countByValueApprox(long,%20double,%20scala.math.Ordering)">countByValueApprox</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#dependencies()">dependencies</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#distinct()">distinct</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#distinct(int,%20scala.math.Ordering)">distinct</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#doubleRDDToDoubleRDDFunctions(org.apache.spark.rdd.RDD)">doubleRDDToDoubleRDDFunctions</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#filter(scala.Function1)">
 filter</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#first()">first</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#flatMap(scala.Function1,%20scala.reflect.ClassTag)">flatMap</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#fold(T,%20scala.Function2)">fold</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#foreach(scala.Function1)">foreach</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#foreachPartition(scala.Function1)">foreachPartition</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#getCheckpointFile()">getCheckpointFile</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#getNumPartitions()">getNumPartitions</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#getStorageLevel()">getStorageLevel</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#glom()">glom</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#groupBy(scala.Function1,%20scala.reflect.ClassTag)">groupBy</a>, <a href="../../..
 /../../org/apache/spark/rdd/RDD.html#groupBy(scala.Function1,%20int,%20scala.reflect.ClassTag)">groupBy</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#groupBy(scala.Function1,%20org.apache.spark.Partitioner,%20scala.reflect.ClassTag,%20scala.math.Ordering)">groupBy</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#id()">id</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#intersection(org.apache.spark.rdd.RDD)">intersection</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#intersection(org.apache.spark.rdd.RDD,%20int)">intersection</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#intersection(org.apache.spark.rdd.RDD,%20org.apache.spark.Partitioner,%20scala.math.Ordering)">intersection</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#isCheckpointed()">isCheckpointed</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#isEmpty()">isEmpty</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#iterator(org.apache.spar
 k.Partition,%20org.apache.spark.TaskContext)">iterator</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#keyBy(scala.Function1)">keyBy</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#localCheckpoint()">localCheckpoint</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#map(scala.Function1,%20scala.reflect.ClassTag)">map</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#mapPartitions(scala.Function1,%20boolean,%20scala.reflect.ClassTag)">mapPartitions</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#mapPartitionsWithIndex(scala.Function2,%20boolean,%20scala.reflect.ClassTag)">mapPartitionsWithIndex</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#max(scala.math.Ordering)">max</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#min(scala.math.Ordering)">min</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#name()">name</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#numericRDDToDoubleRDDFunctions(org.ap
 ache.spark.rdd.RDD,%20scala.math.Numeric)">numericRDDToDoubleRDDFunctions</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#partitioner()">partitioner</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#partitions()">partitions</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#persist()">persist</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#persist(org.apache.spark.storage.StorageLevel)">persist</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#pipe(scala.collection.Seq,%20scala.collection.Map,%20scala.Function1,%20scala.Function2,%20boolean,%20int,%20java.lang.String)">pipe</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#pipe(java.lang.String)">pipe</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#pipe(java.lang.String,%20scala.collection.Map)">pipe</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#preferredLocations(org.apache.spark.Partition)">preferredLocations</a>, <a href="../../../../../org/apache/sp
 ark/rdd/RDD.html#randomSplit(double[],%20long)">randomSplit</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#rddToAsyncRDDActions(org.apache.spark.rdd.RDD,%20scala.reflect.ClassTag)">rddToAsyncRDDActions</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#rddToOrderedRDDFunctions(org.apache.spark.rdd.RDD,%20scala.math.Ordering,%20scala.reflect.ClassTag,%20scala.reflect.ClassTag)">rddToOrderedRDDFunctions</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#rddToPairRDDFunctions(org.apache.spark.rdd.RDD,%20scala.reflect.ClassTag,%20scala.reflect.ClassTag,%20scala.math.Ordering)">rddToPairRDDFunctions</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#rddToSequenceFileRDDFunctions(org.apache.spark.rdd.RDD,%20scala.reflect.ClassTag,%20scala.reflect.ClassTag,%20,%20)">rddToSequenceFileRDDFunctions</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#reduce(scala.Function2)">reduce</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#repartition
 (int,%20scala.math.Ordering)">repartition</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#sample(boolean,%20double,%20long)">sample</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#saveAsObjectFile(java.lang.String)">saveAsObjectFile</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#saveAsTextFile(java.lang.String)">saveAsTextFile</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#saveAsTextFile(java.lang.String,%20java.lang.Class)">saveAsTextFile</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#setName(java.lang.String)">setName</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#sortBy(scala.Function1,%20boolean,%20int,%20scala.math.Ordering,%20scala.reflect.ClassTag)">sortBy</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#sparkContext()">sparkContext</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#subtract(org.apache.spark.rdd.RDD)">subtract</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#subt
 ract(org.apache.spark.rdd.RDD,%20int)">subtract</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#subtract(org.apache.spark.rdd.RDD,%20org.apache.spark.Partitioner,%20scala.math.Ordering)">subtract</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#take(int)">take</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#takeOrdered(int,%20scala.math.Ordering)">takeOrdered</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#takeSample(boolean,%20int,%20long)">takeSample</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#toDebugString()">toDebugString</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#toJavaRDD()">toJavaRDD</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#toLocalIterator()">toLocalIterator</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#top(int,%20scala.math.Ordering)">top</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#toString()">toString</a>, <a href="../../../../../org/apache/spark/rdd/RDD.h
 tml#treeAggregate(U,%20scala.Function2,%20scala.Function2,%20int,%20scala.reflect.ClassTag)">treeAggregate</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#treeReduce(scala.Function2,%20int)">treeReduce</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#union(org.apache.spark.rdd.RDD)">union</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#unpersist(boolean)">unpersist</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#zip(org.apache.spark.rdd.RDD,%20scala.reflect.ClassTag)">zip</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#zipPartitions(org.apache.spark.rdd.RDD,%20boolean,%20scala.Function2,%20scala.reflect.ClassTag,%20scala.reflect.ClassTag)">zipPartitions</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#zipPartitions(org.apache.spark.rdd.RDD,%20scala.Function2,%20scala.reflect.ClassTag,%20scala.reflect.ClassTag)">zipPartitions</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#zipPartitions(org.apache.spark.rdd.RD
 D,%20org.apache.spark.rdd.RDD,%20boolean,%20scala.Function3,%20scala.reflect.ClassTag,%20scala.reflect.ClassTag,%20scala.reflect.ClassTag)">zipPartitions</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#zipPartitions(org.apache.spark.rdd.RDD,%20org.apache.spark.rdd.RDD,%20scala.Function3,%20scala.reflect.ClassTag,%20scala.reflect.ClassTag,%20scala.reflect.ClassTag)">zipPartitions</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#zipPartitions(org.apache.spark.rdd.RDD,%20org.apache.spark.rdd.RDD,%20org.apache.spark.rdd.RDD,%20boolean,%20scala.Function4,%20scala.reflect.ClassTag,%20scala.reflect.ClassTag,%20scala.reflect.ClassTag,%20scala.reflect.ClassTag)">zipPartitions</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#zipPartitions(org.apache.spark.rdd.RDD,%20org.apache.spark.rdd.RDD,%20org.apache.spark.rdd.RDD,%20scala.Function4,%20scala.reflect.ClassTag,%20scala.reflect.ClassTag,%20scala.reflect.ClassTag,%20scala.reflect.ClassTag)">zipPartitions</a>, <a
  href="../../../../../org/apache/spark/rdd/RDD.html#zipWithIndex()">zipWithIndex</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#zipWithUniqueId()">zipWithUniqueId</a></code></li>
+</ul>
+<ul class="blockList">
+<li class="blockList"><a name="methods_inherited_from_class_Object">
+<!--   -->
+</a>
+<h3>Methods inherited from class&nbsp;Object</h3>
+<code>equals, getClass, hashCode, notify, notifyAll, wait, wait, wait</code></li>
+</ul>
+</li>
+</ul>
+</li>
+</ul>
+</div>
+<div class="details">
+<ul class="blockList">
+<li class="blockList">
+<!-- ========= CONSTRUCTOR DETAIL ======== -->
+<ul class="blockList">
+<li class="blockList"><a name="constructor_detail">
+<!--   -->
+</a>
+<h3>Constructor Detail</h3>
+<a name="RRDD(org.apache.spark.rdd.RDD, byte[], java.lang.String, java.lang.String, byte[], java.lang.Object[], scala.reflect.ClassTag)">
+<!--   -->
+</a>
+<ul class="blockListLast">
+<li class="blockList">
+<h4>RRDD</h4>
+<pre>public&nbsp;RRDD(<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;<a href="../../../../../org/apache/spark/api/r/RRDD.html" title="type parameter in RRDD">T</a>&gt;&nbsp;parent,
+    byte[]&nbsp;func,
+    String&nbsp;deserializer,
+    String&nbsp;serializer,
+    byte[]&nbsp;packageNames,
+    Object[]&nbsp;broadcastVars,
+    scala.reflect.ClassTag&lt;<a href="../../../../../org/apache/spark/api/r/RRDD.html" title="type parameter in RRDD">T</a>&gt;&nbsp;evidence$4)</pre>
+</li>
+</ul>
+</li>
+</ul>
+<!-- ============ METHOD DETAIL ========== -->
+<ul class="blockList">
+<li class="blockList"><a name="method_detail">
+<!--   -->
+</a>
+<h3>Method Detail</h3>
+<a name="createSparkContext(java.lang.String, java.lang.String, java.lang.String, java.lang.String[], java.util.Map, java.util.Map)">
+<!--   -->
+</a>
+<ul class="blockList">
+<li class="blockList">
+<h4>createSparkContext</h4>
+<pre>public static&nbsp;<a href="../../../../../org/apache/spark/api/java/JavaSparkContext.html" title="class in org.apache.spark.api.java">JavaSparkContext</a>&nbsp;createSparkContext(String&nbsp;master,
+                                  String&nbsp;appName,
+                                  String&nbsp;sparkHome,
+                                  String[]&nbsp;jars,
+                                  java.util.Map&lt;Object,Object&gt;&nbsp;sparkEnvirMap,
+                                  java.util.Map&lt;Object,Object&gt;&nbsp;sparkExecutorEnvMap)</pre>
+</li>
+</ul>
+<a name="createRDDFromArray(org.apache.spark.api.java.JavaSparkContext, byte[][])">
+<!--   -->
+</a>
+<ul class="blockList">
+<li class="blockList">
+<h4>createRDDFromArray</h4>
+<pre>public static&nbsp;<a href="../../../../../org/apache/spark/api/java/JavaRDD.html" title="class in org.apache.spark.api.java">JavaRDD</a>&lt;byte[]&gt;&nbsp;createRDDFromArray(<a href="../../../../../org/apache/spark/api/java/JavaSparkContext.html" title="class in org.apache.spark.api.java">JavaSparkContext</a>&nbsp;jsc,
+                                 byte[][]&nbsp;arr)</pre>
+<div class="block">Create an RRDD given a sequence of byte arrays. Used to create RRDD when <code>parallelize</code> is
+ called from R.</div>
+<dl><dt><span class="strong">Parameters:</span></dt><dd><code>jsc</code> - (undocumented)</dd><dd><code>arr</code> - (undocumented)</dd>
+<dt><span class="strong">Returns:</span></dt><dd>(undocumented)</dd></dl>
+</li>
+</ul>
+<a name="createRDDFromFile(org.apache.spark.api.java.JavaSparkContext, java.lang.String, int)">
+<!--   -->
+</a>
+<ul class="blockList">
+<li class="blockList">
+<h4>createRDDFromFile</h4>
+<pre>public static&nbsp;<a href="../../../../../org/apache/spark/api/java/JavaRDD.html" title="class in org.apache.spark.api.java">JavaRDD</a>&lt;byte[]&gt;&nbsp;createRDDFromFile(<a href="../../../../../org/apache/spark/api/java/JavaSparkContext.html" title="class in org.apache.spark.api.java">JavaSparkContext</a>&nbsp;jsc,
+                                String&nbsp;fileName,
+                                int&nbsp;parallelism)</pre>
+<div class="block">Create an RRDD given a temporary file name. This is used to create RRDD when parallelize is
+ called on large R objects.
+ <p></div>
+<dl><dt><span class="strong">Parameters:</span></dt><dd><code>fileName</code> - name of temporary file on driver machine</dd><dd><code>parallelism</code> - number of slices defaults to 4</dd><dd><code>jsc</code> - (undocumented)</dd>
+<dt><span class="strong">Returns:</span></dt><dd>(undocumented)</dd></dl>
+</li>
+</ul>
+<a name="partitioner()">
+<!--   -->
+</a>
+<ul class="blockList">
+<li class="blockList">
+<h4>partitioner</h4>
+<pre>public static&nbsp;scala.Option&lt;<a href="../../../../../org/apache/spark/Partitioner.html" title="class in org.apache.spark">Partitioner</a>&gt;&nbsp;partitioner()</pre>
+</li>
+</ul>
+<a name="sparkContext()">
+<!--   -->
+</a>
+<ul class="blockList">
+<li class="blockList">
+<h4>sparkContext</h4>
+<pre>public static&nbsp;<a href="../../../../../org/apache/spark/SparkContext.html" title="class in org.apache.spark">SparkContext</a>&nbsp;sparkContext()</pre>
+</li>
+</ul>
+<a name="id()">
+<!--   -->
+</a>
+<ul class="blockList">
+<li class="blockList">
+<h4>id</h4>
+<pre>public static&nbsp;int&nbsp;id()</pre>
+</li>
+</ul>
+<a name="name()">
+<!--   -->
+</a>
+<ul class="blockList">
+<li class="blockList">
+<h4>name</h4>
+<pre>public static&nbsp;String&nbsp;name()</pre>
+</li>
+</ul>
+<a name="name_$eq(java.lang.String)">
+<!--   -->
+</a>
+<ul class="blockList">
+<li class="blockList">
+<h4>name_$eq</h4>
+<pre>public static&nbsp;void&nbsp;name_$eq(String&nbsp;x$1)</pre>
+</li>
+</ul>
+<a name="setName(java.lang.String)">
+<!--   -->
+</a>
+<ul class="blockList">
+<li class="blockList">
+<h4>setName</h4>
+<pre>public static&nbsp;<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;&nbsp;setName(String&nbsp;_name)</pre>
+</li>
+</ul>
+<a name="persist(org.apache.spark.storage.StorageLevel)">
+<!--   -->
+</a>
+<ul class="blockList">
+<li class="blockList">
+<h4>persist</h4>
+<pre>public static&nbsp;<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;&nbsp;persist(<a href="../../../../../org/apache/spark/storage/StorageLevel.html" title="class in org.apache.spark.storage">StorageLevel</a>&nbsp;newLevel)</pre>
+</li>
+</ul>
+<a name="persist()">
+<!--   -->
+</a>
+<ul class="blockList">
+<li class="blockList">
+<h4>persist</h4>
+<pre>public static&nbsp;<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;&nbsp;persist()</pre>
+</li>
+</ul>
+<a name="cache()">
+<!--   -->
+</a>
+<ul class="blockList">
+<li class="blockList">
+<h4>cache</h4>
+<pre>public static&nbsp;<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;&nbsp;cache()</pre>
+</li>
+</ul>
+<a name="unpersist(boolean)">
+<!--   -->
+</a>
+<ul class="blockList">
+<li class="blockList">
+<h4>unpersist</h4>
+<pre>public static&nbsp;<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;&nbsp;unpersist(boolean&nbsp;blocking)</pre>
+</li>
+</ul>
+<a name="getStorageLevel()">
+<!--   -->
+</a>
+<ul class="blockList">
+<li class="blockList">
+<h4>getStorageLevel</h4>
+<pre>public static&nbsp;<a href="../../../../../org/apache/spark/storage/StorageLevel.html" title="class in org.apache.spark.storage">StorageLevel</a>&nbsp;getStorageLevel()</pre>
+</li>
+</ul>
+<a name="dependencies()">
+<!--   -->
+</a>
+<ul class="blockList">
+<li class="blockList">
+<h4>dependencies</h4>
+<pre>public static final&nbsp;scala.collection.Seq&lt;<a href="../../../../../org/apache/spark/Dependency.html" title="class in org.apache.spark">Dependency</a>&lt;?&gt;&gt;&nbsp;dependencies()</pre>
+</li>
+</ul>
+<a name="partitions()">
+<!--   -->
+</a>
+<ul class="blockList">
+<li class="blockList">
+<h4>partitions</h4>
+<pre>public static final&nbsp;<a href="../../../../../org/apache/spark/Partition.html" title="interface in org.apache.spark">Partition</a>[]&nbsp;partitions()</pre>
+</li>
+</ul>
+<a name="getNumPartitions()">
+<!--   -->
+</a>
+<ul class="blockList">
+<li class="blockList">
+<h4>getNumPartitions</h4>
+<pre>public static final&nbsp;int&nbsp;getNumPartitions()</pre>
+</li>
+</ul>
+<a name="preferredLocations(org.apache.spark.Partition)">
+<!--   -->
+</a>
+<ul class="blockList">
+<li class="blockList">
+<h4>preferredLocations</h4>
+<pre>public static final&nbsp;scala.collection.Seq&lt;String&gt;&nbsp;preferredLocations(<a href="../../../../../org/apache/spark/Partition.html" title="interface in org.apache.spark">Partition</a>&nbsp;split)</pre>
+</li>
+</ul>
+<a name="iterator(org.apache.spark.Partition, org.apache.spark.TaskContext)">
+<!--   -->
+</a>
+<ul class="blockList">
+<li class="blockList">
+<h4>iterator</h4>
+<pre>public static final&nbsp;scala.collection.Iterator&lt;T&gt;&nbsp;iterator(<a href="../../../../../org/apache/spark/Partition.html" title="interface in org.apache.spark">Partition</a>&nbsp;split,
+                                    <a href="../../../../../org/apache/spark/TaskContext.html" title="class in org.apache.spark">TaskContext</a>&nbsp;context)</pre>
+</li>
+</ul>
+<a name="map(scala.Function1, scala.reflect.ClassTag)">
+<!--   -->
+</a>
+<ul class="blockList">
+<li class="blockList">
+<h4>map</h4>
+<pre>public static&nbsp;&lt;U&gt;&nbsp;<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;U&gt;&nbsp;map(scala.Function1&lt;T,U&gt;&nbsp;f,
+             scala.reflect.ClassTag&lt;U&gt;&nbsp;evidence$3)</pre>
+</li>
+</ul>
+<a name="flatMap(scala.Function1, scala.reflect.ClassTag)">
+<!--   -->
+</a>
+<ul class="blockList">
+<li class="blockList">
+<h4>flatMap</h4>
+<pre>public static&nbsp;&lt;U&gt;&nbsp;<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;U&gt;&nbsp;flatMap(scala.Function1&lt;T,scala.collection.TraversableOnce&lt;U&gt;&gt;&nbsp;f,
+                 scala.reflect.ClassTag&lt;U&gt;&nbsp;evidence$4)</pre>
+</li>
+</ul>
+<a name="filter(scala.Function1)">
+<!--   -->
+</a>
+<ul class="blockList">
+<li class="blockList">
+<h4>filter</h4>
+<pre>public static&nbsp;<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;&nbsp;filter(scala.Function1&lt;T,Object&gt;&nbsp;f)</pre>
+</li>
+</ul>
+<a name="distinct(int, scala.math.Ordering)">
+<!--   -->
+</a>
+<ul class="blockList">
+<li class="blockList">
+<h4>distinct</h4>
+<pre>public static&nbsp;<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;&nbsp;distinct(int&nbsp;numPartitions,
+              scala.math.Ordering&lt;T&gt;&nbsp;ord)</pre>
+</li>
+</ul>
+<a name="distinct()">
+<!--   -->
+</a>
+<ul class="blockList">
+<li class="blockList">
+<h4>distinct</h4>
+<pre>public static&nbsp;<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;&nbsp;distinct()</pre>
+</li>
+</ul>
+<a name="repartition(int, scala.math.Ordering)">
+<!--   -->
+</a>
+<ul class="blockList">
+<li class="blockList">
+<h4>repartition</h4>
+<pre>public static&nbsp;<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;&nbsp;repartition(int&nbsp;numPartitions,
+                 scala.math.Ordering&lt;T&gt;&nbsp;ord)</pre>
+</li>
+</ul>
+<a name="coalesce(int, boolean, scala.Option, scala.math.Ordering)">
+<!--   -->
+</a>
+<ul class="blockList">
+<li class="blockList">
+<h4>coalesce</h4>
+<pre>public static&nbsp;<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;&nbsp;coalesce(int&nbsp;numPartitions,
+              boolean&nbsp;shuffle,
+              scala.Option&lt;<a href="../../../../../org/apache/spark/rdd/PartitionCoalescer.html" title="interface in org.apache.spark.rdd">PartitionCoalescer</a>&gt;&nbsp;partitionCoalescer,
+              scala.math.Ordering&lt;T&gt;&nbsp;ord)</pre>
+</li>
+</ul>
+<a name="sample(boolean, double, long)">
+<!--   -->
+</a>
+<ul class="blockList">
+<li class="blockList">
+<h4>sample</h4>
+<pre>public static&nbsp;<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;&nbsp;sample(boolean&nbsp;withReplacement,
+            double&nbsp;fraction,
+            long&nbsp;seed)</pre>
+</li>
+</ul>
+<a name="randomSplit(double[], long)">
+<!--   -->
+</a>
+<ul class="blockList">
+<li class="blockList">
+<h4>randomSplit</h4>
+<pre>public static&nbsp;<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;[]&nbsp;randomSplit(double[]&nbsp;weights,
+                   long&nbsp;seed)</pre>
+</li>
+</ul>
+<a name="takeSample(boolean, int, long)">
+<!--   -->
+</a>
+<ul class="blockList">
+<li class="blockList">
+<h4>takeSample</h4>
+<pre>public static&nbsp;Object&nbsp;takeSample(boolean&nbsp;withReplacement,
+                int&nbsp;num,
+                long&nbsp;seed)</pre>
+</li>
+</ul>
+<a name="union(org.apache.spark.rdd.RDD)">
+<!--   -->
+</a>
+<ul class="blockList">
+<li class="blockList">
+<h4>union</h4>
+<pre>public static&nbsp;<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;&nbsp;union(<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;&nbsp;other)</pre>
+</li>
+</ul>
+<a name="$plus$plus(org.apache.spark.rdd.RDD)">
+<!--   -->
+</a>
+<ul class="blockList">
+<li class="blockList">
+<h4>$plus$plus</h4>
+<pre>public static&nbsp;<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;&nbsp;$plus$plus(<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;&nbsp;other)</pre>
+</li>
+</ul>
+<a name="sortBy(scala.Function1, boolean, int, scala.math.Ordering, scala.reflect.ClassTag)">
+<!--   -->
+</a>
+<ul class="blockList">
+<li class="blockList">
+<h4>sortBy</h4>
+<pre>public static&nbsp;&lt;K&gt;&nbsp;<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;&nbsp;sortBy(scala.Function1&lt;T,K&gt;&nbsp;f,
+                boolean&nbsp;ascending,
+                int&nbsp;numPartitions,
+                scala.math.Ordering&lt;K&gt;&nbsp;ord,
+                scala.reflect.ClassTag&lt;K&gt;&nbsp;ctag)</pre>
+</li>
+</ul>
+<a name="intersection(org.apache.spark.rdd.RDD)">
+<!--   -->
+</a>
+<ul class="blockList">
+<li class="blockList">
+<h4>intersection</h4>
+<pre>public static&nbsp;<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;&nbsp;intersection(<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;&nbsp;other)</pre>
+</li>
+</ul>
+<a name="intersection(org.apache.spark.rdd.RDD, org.apache.spark.Partitioner, scala.math.Ordering)">
+<!--   -->
+</a>
+<ul class="blockList">
+<li class="blockList">
+<h4>intersection</h4>
+<pre>public static&nbsp;<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;&nbsp;intersection(<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;&nbsp;other,
+                  <a href="../../../../../org/apache/spark/Partitioner.html" title="class in org.apache.spark">Partitioner</a>&nbsp;partitioner,
+                  scala.math.Ordering&lt;T&gt;&nbsp;ord)</pre>
+</li>
+</ul>
+<a name="intersection(org.apache.spark.rdd.RDD, int)">
+<!--   -->
+</a>
+<ul class="blockList">
+<li class="blockList">
+<h4>intersection</h4>
+<pre>public static&nbsp;<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;&nbsp;intersection(<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;&nbsp;other,
+                  int&nbsp;numPartitions)</pre>
+</li>
+</ul>
+<a name="glom()">
+<!--   -->
+</a>
+<ul class="blockList">
+<li class="blockList">
+<h4>glom</h4>
+<pre>public static&nbsp;<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;Object&gt;&nbsp;glom()</pre>
+</li>
+</ul>
+<a name="cartesian(org.apache.spark.rdd.RDD, scala.reflect.ClassTag)">
+<!--   -->
+</a>
+<ul class="blockList">
+<li class="blockList">
+<h4>cartesian</h4>
+<pre>public static&nbsp;&lt;U&gt;&nbsp;<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;scala.Tuple2&lt;T,U&gt;&gt;&nbsp;cartesian(<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;U&gt;&nbsp;other,
+                     

<TRUNCATED>

---------------------------------------------------------------------
To unsubscribe, e-mail: commits-unsubscribe@spark.apache.org
For additional commands, e-mail: commits-help@spark.apache.org


Mime
View raw message