Repository: spark-website Updated Branches: refs/heads/asf-site ac6ba355d -> 3a6349595
http://git-wip-us.apache.org/repos/asf/spark-website/blob/3a634959/site/docs/2.1.3/api/java/org/apache/spark/api/r/RRDD.html ---------------------------------------------------------------------- diff --git a/site/docs/2.1.3/api/java/org/apache/spark/api/r/RRDD.html b/site/docs/2.1.3/api/java/org/apache/spark/api/r/RRDD.html new file mode 100644 index 0000000..3cc1f29 --- /dev/null +++ b/site/docs/2.1.3/api/java/org/apache/spark/api/r/RRDD.html @@ -0,0 +1,2154 @@ +<!DOCTYPE HTML PUBLIC "-//W3C//DTD HTML 4.01 Transitional//EN" "http://www.w3.org/TR/html4/loose.dtd"> +<!-- NewPage --> +<html lang="en"> +<head> +<!-- Generated by javadoc (version 1.7.0_80) on Tue Jun 26 17:20:36 UTC 2018 --> +<title>RRDD (Spark 2.1.3 JavaDoc)</title> +<meta name="date" content="2018-06-26"> +<link rel="stylesheet" type="text/css" href="../../../../../stylesheet.css" title="Style"> +</head> +<body> +<script type="text/javascript"><!-- + if (location.href.indexOf('is-external=true') == -1) { + parent.document.title="RRDD (Spark 2.1.3 JavaDoc)"; + } +//--> +</script> +<noscript> +<div>JavaScript is disabled on your browser.</div> +</noscript> +<!-- ========= START OF TOP NAVBAR ======= --> +<div class="topNav"><a name="navbar_top"> +<!-- --> +</a><a href="#skip-navbar_top" title="Skip navigation links"></a><a name="navbar_top_firstrow"> +<!-- --> +</a> +<ul class="navList" title="Navigation"> +<li><a href="../../../../../overview-summary.html">Overview</a></li> +<li><a href="package-summary.html">Package</a></li> +<li class="navBarCell1Rev">Class</li> +<li><a href="package-tree.html">Tree</a></li> +<li><a href="../../../../../deprecated-list.html">Deprecated</a></li> +<li><a href="../../../../../index-all.html">Index</a></li> +<li><a href="../../../../../help-doc.html">Help</a></li> +</ul> +</div> +<div class="subNav"> +<ul class="navList"> +<li><a href="../../../../../org/apache/spark/api/r/RBackendAuthHandler.html" title="class in org.apache.spark.api.r"><span class="strong">Prev Class</span></a></li> +<li><a href="../../../../../org/apache/spark/api/r/RRunnerModes.html" title="class in org.apache.spark.api.r"><span class="strong">Next Class</span></a></li> +</ul> +<ul class="navList"> +<li><a href="../../../../../index.html?org/apache/spark/api/r/RRDD.html" target="_top">Frames</a></li> +<li><a href="RRDD.html" target="_top">No Frames</a></li> +</ul> +<ul class="navList" id="allclasses_navbar_top"> +<li><a href="../../../../../allclasses-noframe.html">All Classes</a></li> +</ul> +<div> +<script type="text/javascript"><!-- + allClassesLink = document.getElementById("allclasses_navbar_top"); + if(window==top) { + allClassesLink.style.display = "block"; + } + else { + allClassesLink.style.display = "none"; + } + //--> +</script> +</div> +<div> +<ul class="subNavList"> +<li>Summary: </li> +<li>Nested | </li> +<li>Field | </li> +<li><a href="#constructor_summary">Constr</a> | </li> +<li><a href="#method_summary">Method</a></li> +</ul> +<ul class="subNavList"> +<li>Detail: </li> +<li>Field | </li> +<li><a href="#constructor_detail">Constr</a> | </li> +<li><a href="#method_detail">Method</a></li> +</ul> +</div> +<a name="skip-navbar_top"> +<!-- --> +</a></div> +<!-- ========= END OF TOP NAVBAR ========= --> +<!-- ======== START OF CLASS DATA ======== --> +<div class="header"> +<div class="subTitle">org.apache.spark.api.r</div> +<h2 title="Class RRDD" class="title">Class RRDD<T></h2> +</div> +<div class="contentContainer"> +<ul class="inheritance"> +<li>Object</li> +<li> +<ul class="inheritance"> +<li><a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">org.apache.spark.rdd.RDD</a><U></li> +<li> +<ul class="inheritance"> +<li><a href="../../../../../org/apache/spark/api/r/BaseRRDD.html" title="class in org.apache.spark.api.r">org.apache.spark.api.r.BaseRRDD</a><T,byte[]></li> +<li> +<ul class="inheritance"> +<li>org.apache.spark.api.r.RRDD<T></li> +</ul> +</li> +</ul> +</li> +</ul> +</li> +</ul> +<div class="description"> +<ul class="blockList"> +<li class="blockList"> +<dl> +<dt>All Implemented Interfaces:</dt> +<dd>java.io.Serializable</dd> +</dl> +<hr> +<br> +<pre>public class <span class="strong">RRDD<T></span> +extends <a href="../../../../../org/apache/spark/api/r/BaseRRDD.html" title="class in org.apache.spark.api.r">BaseRRDD</a><T,byte[]></pre> +<div class="block">An RDD that stores serialized R objects as Array[Byte].</div> +<dl><dt><span class="strong">See Also:</span></dt><dd><a href="../../../../../serialized-form.html#org.apache.spark.api.r.RRDD">Serialized Form</a></dd></dl> +</li> +</ul> +</div> +<div class="summary"> +<ul class="blockList"> +<li class="blockList"> +<!-- ======== CONSTRUCTOR SUMMARY ======== --> +<ul class="blockList"> +<li class="blockList"><a name="constructor_summary"> +<!-- --> +</a> +<h3>Constructor Summary</h3> +<table class="overviewSummary" border="0" cellpadding="3" cellspacing="0" summary="Constructor Summary table, listing constructors, and an explanation"> +<caption><span>Constructors</span><span class="tabEnd"> </span></caption> +<tr> +<th class="colOne" scope="col">Constructor and Description</th> +</tr> +<tr class="altColor"> +<td class="colOne"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#RRDD(org.apache.spark.rdd.RDD,%20byte[],%20java.lang.String,%20java.lang.String,%20byte[],%20java.lang.Object[],%20scala.reflect.ClassTag)">RRDD</a></strong>(<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><<a href="../../../../../org/apache/spark/api/r/RRDD.html" title="type parameter in RRDD">T</a>> parent, + byte[] func, + String deserializer, + String serializer, + byte[] packageNames, + Object[] broadcastVars, + scala.reflect.ClassTag<<a href="../../../../../org/apache/spark/api/r/RRDD.html" title="type parameter in RRDD">T</a>> evidence$4)</code> </td> +</tr> +</table> +</li> +</ul> +<!-- ========== METHOD SUMMARY =========== --> +<ul class="blockList"> +<li class="blockList"><a name="method_summary"> +<!-- --> +</a> +<h3>Method Summary</h3> +<table class="overviewSummary" border="0" cellpadding="3" cellspacing="0" summary="Method Summary table, listing methods, and an explanation"> +<caption><span>Methods</span><span class="tabEnd"> </span></caption> +<tr> +<th class="colFirst" scope="col">Modifier and Type</th> +<th class="colLast" scope="col">Method and Description</th> +</tr> +<tr class="altColor"> +<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><T></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#$plus$plus(org.apache.spark.rdd.RDD)">$plus$plus</a></strong>(<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><T> other)</code> </td> +</tr> +<tr class="rowColor"> +<td class="colFirst"><code>static <U> U</code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#aggregate(U,%20scala.Function2,%20scala.Function2,%20scala.reflect.ClassTag)">aggregate</a></strong>(U zeroValue, + scala.Function2<U,T,U> seqOp, + scala.Function2<U,U,U> combOp, + scala.reflect.ClassTag<U> evidence$30)</code> </td> +</tr> +<tr class="altColor"> +<td class="colFirst"><code><a href="../../../../../org/apache/spark/api/java/JavaRDD.html" title="class in org.apache.spark.api.java">JavaRDD</a><byte[]></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#asJavaRDD()">asJavaRDD</a></strong>()</code> </td> +</tr> +<tr class="rowColor"> +<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><T></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#cache()">cache</a></strong>()</code> </td> +</tr> +<tr class="altColor"> +<td class="colFirst"><code>static <U> <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><scala.Tuple2<T,U>></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#cartesian(org.apache.spark.rdd.RDD,%20scala.reflect.ClassTag)">cartesian</a></strong>(<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><U> other, + scala.reflect.ClassTag<U> evidence$5)</code> </td> +</tr> +<tr class="rowColor"> +<td class="colFirst"><code>static void</code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#checkpoint()">checkpoint</a></strong>()</code> </td> +</tr> +<tr class="altColor"> +<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><T></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#coalesce(int,%20boolean,%20scala.Option,%20scala.math.Ordering)">coalesce</a></strong>(int numPartitions, + boolean shuffle, + scala.Option<<a href="../../../../../org/apache/spark/rdd/PartitionCoalescer.html" title="interface in org.apache.spark.rdd">PartitionCoalescer</a>> partitionCoalescer, + scala.math.Ordering<T> ord)</code> </td> +</tr> +<tr class="rowColor"> +<td class="colFirst"><code>static boolean</code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#coalesce$default$2()">coalesce$default$2</a></strong>()</code> </td> +</tr> +<tr class="altColor"> +<td class="colFirst"><code>static scala.Option<<a href="../../../../../org/apache/spark/rdd/PartitionCoalescer.html" title="interface in org.apache.spark.rdd">PartitionCoalescer</a>></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#coalesce$default$3()">coalesce$default$3</a></strong>()</code> </td> +</tr> +<tr class="rowColor"> +<td class="colFirst"><code>static scala.math.Ordering<T></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#coalesce$default$4(int,%20boolean,%20scala.Option)">coalesce$default$4</a></strong>(int numPartitions, + boolean shuffle, + scala.Option<<a href="../../../../../org/apache/spark/rdd/PartitionCoalescer.html" title="interface in org.apache.spark.rdd">PartitionCoalescer</a>> partitionCoalescer)</code> </td> +</tr> +<tr class="altColor"> +<td class="colFirst"><code>static Object</code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#collect()">collect</a></strong>()</code> </td> +</tr> +<tr class="rowColor"> +<td class="colFirst"><code>static <U> <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><U></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#collect(scala.PartialFunction,%20scala.reflect.ClassTag)">collect</a></strong>(scala.PartialFunction<T,U> f, + scala.reflect.ClassTag<U> evidence$29)</code> </td> +</tr> +<tr class="altColor"> +<td class="colFirst"><code>static scala.collection.Iterator<U></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#compute(org.apache.spark.Partition,%20org.apache.spark.TaskContext)">compute</a></strong>(<a href="../../../../../org/apache/spark/Partition.html" title="interface in org.apache.spark">Partition</a> partition, + <a href="../../../../../org/apache/spark/TaskContext.html" title="class in org.apache.spark">TaskContext</a> context)</code> </td> +</tr> +<tr class="rowColor"> +<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/SparkContext.html" title="class in org.apache.spark">SparkContext</a></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#context()">context</a></strong>()</code> </td> +</tr> +<tr class="altColor"> +<td class="colFirst"><code>static long</code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#count()">count</a></strong>()</code> </td> +</tr> +<tr class="rowColor"> +<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/partial/PartialResult.html" title="class in org.apache.spark.partial">PartialResult</a><<a href="../../../../../org/apache/spark/partial/BoundedDouble.html" title="class in org.apache.spark.partial">BoundedDouble</a>></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#countApprox(long,%20double)">countApprox</a></strong>(long timeout, + double confidence)</code> </td> +</tr> +<tr class="altColor"> +<td class="colFirst"><code>static double</code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#countApprox$default$2()">countApprox$default$2</a></strong>()</code> </td> +</tr> +<tr class="rowColor"> +<td class="colFirst"><code>static long</code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#countApproxDistinct(double)">countApproxDistinct</a></strong>(double relativeSD)</code> </td> +</tr> +<tr class="altColor"> +<td class="colFirst"><code>static long</code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#countApproxDistinct(int,%20int)">countApproxDistinct</a></strong>(int p, + int sp)</code> </td> +</tr> +<tr class="rowColor"> +<td class="colFirst"><code>static double</code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#countApproxDistinct$default$1()">countApproxDistinct$default$1</a></strong>()</code> </td> +</tr> +<tr class="altColor"> +<td class="colFirst"><code>static scala.collection.Map<T,Object></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#countByValue(scala.math.Ordering)">countByValue</a></strong>(scala.math.Ordering<T> ord)</code> </td> +</tr> +<tr class="rowColor"> +<td class="colFirst"><code>static scala.math.Ordering<T></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#countByValue$default$1()">countByValue$default$1</a></strong>()</code> </td> +</tr> +<tr class="altColor"> +<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/partial/PartialResult.html" title="class in org.apache.spark.partial">PartialResult</a><scala.collection.Map<T,<a href="../../../../../org/apache/spark/partial/BoundedDouble.html" title="class in org.apache.spark.partial">BoundedDouble</a>>></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#countByValueApprox(long,%20double,%20scala.math.Ordering)">countByValueApprox</a></strong>(long timeout, + double confidence, + scala.math.Ordering<T> ord)</code> </td> +</tr> +<tr class="rowColor"> +<td class="colFirst"><code>static double</code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#countByValueApprox$default$2()">countByValueApprox$default$2</a></strong>()</code> </td> +</tr> +<tr class="altColor"> +<td class="colFirst"><code>static scala.math.Ordering<T></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#countByValueApprox$default$3(long,%20double)">countByValueApprox$default$3</a></strong>(long timeout, + double confidence)</code> </td> +</tr> +<tr class="rowColor"> +<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/api/java/JavaRDD.html" title="class in org.apache.spark.api.java">JavaRDD</a><byte[]></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#createRDDFromArray(org.apache.spark.api.java.JavaSparkContext,%20byte[][])">createRDDFromArray</a></strong>(<a href="../../../../../org/apache/spark/api/java/JavaSparkContext.html" title="class in org.apache.spark.api.java">JavaSparkContext</a> jsc, + byte[][] arr)</code> +<div class="block">Create an RRDD given a sequence of byte arrays.</div> +</td> +</tr> +<tr class="altColor"> +<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/api/java/JavaRDD.html" title="class in org.apache.spark.api.java">JavaRDD</a><byte[]></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#createRDDFromFile(org.apache.spark.api.java.JavaSparkContext,%20java.lang.String,%20int)">createRDDFromFile</a></strong>(<a href="../../../../../org/apache/spark/api/java/JavaSparkContext.html" title="class in org.apache.spark.api.java">JavaSparkContext</a> jsc, + String fileName, + int parallelism)</code> +<div class="block">Create an RRDD given a temporary file name.</div> +</td> +</tr> +<tr class="rowColor"> +<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/api/java/JavaSparkContext.html" title="class in org.apache.spark.api.java">JavaSparkContext</a></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#createSparkContext(java.lang.String,%20java.lang.String,%20java.lang.String,%20java.lang.String[],%20java.util.Map,%20java.util.Map)">createSparkContext</a></strong>(String master, + String appName, + String sparkHome, + String[] jars, + java.util.Map<Object,Object> sparkEnvirMap, + java.util.Map<Object,Object> sparkExecutorEnvMap)</code> </td> +</tr> +<tr class="altColor"> +<td class="colFirst"><code>static scala.collection.Seq<<a href="../../../../../org/apache/spark/Dependency.html" title="class in org.apache.spark">Dependency</a><?>></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#dependencies()">dependencies</a></strong>()</code> </td> +</tr> +<tr class="rowColor"> +<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><T></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#distinct()">distinct</a></strong>()</code> </td> +</tr> +<tr class="altColor"> +<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><T></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#distinct(int,%20scala.math.Ordering)">distinct</a></strong>(int numPartitions, + scala.math.Ordering<T> ord)</code> </td> +</tr> +<tr class="rowColor"> +<td class="colFirst"><code>static scala.math.Ordering<T></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#distinct$default$2(int)">distinct$default$2</a></strong>(int numPartitions)</code> </td> +</tr> +<tr class="altColor"> +<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><T></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#filter(scala.Function1)">filter</a></strong>(scala.Function1<T,Object> f)</code> </td> +</tr> +<tr class="rowColor"> +<td class="colFirst"><code>static T</code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#first()">first</a></strong>()</code> </td> +</tr> +<tr class="altColor"> +<td class="colFirst"><code>static <U> <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><U></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#flatMap(scala.Function1,%20scala.reflect.ClassTag)">flatMap</a></strong>(scala.Function1<T,scala.collection.TraversableOnce<U>> f, + scala.reflect.ClassTag<U> evidence$4)</code> </td> +</tr> +<tr class="rowColor"> +<td class="colFirst"><code>static T</code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#fold(T,%20scala.Function2)">fold</a></strong>(T zeroValue, + scala.Function2<T,T,T> op)</code> </td> +</tr> +<tr class="altColor"> +<td class="colFirst"><code>static void</code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#foreach(scala.Function1)">foreach</a></strong>(scala.Function1<T,scala.runtime.BoxedUnit> f)</code> </td> +</tr> +<tr class="rowColor"> +<td class="colFirst"><code>static void</code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#foreachPartition(scala.Function1)">foreachPartition</a></strong>(scala.Function1<scala.collection.Iterator<T>,scala.runtime.BoxedUnit> f)</code> </td> +</tr> +<tr class="altColor"> +<td class="colFirst"><code>static scala.Option<String></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#getCheckpointFile()">getCheckpointFile</a></strong>()</code> </td> +</tr> +<tr class="rowColor"> +<td class="colFirst"><code>static int</code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#getNumPartitions()">getNumPartitions</a></strong>()</code> </td> +</tr> +<tr class="altColor"> +<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/Partition.html" title="interface in org.apache.spark">Partition</a>[]</code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#getPartitions()">getPartitions</a></strong>()</code> </td> +</tr> +<tr class="rowColor"> +<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/storage/StorageLevel.html" title="class in org.apache.spark.storage">StorageLevel</a></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#getStorageLevel()">getStorageLevel</a></strong>()</code> </td> +</tr> +<tr class="altColor"> +<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><Object></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#glom()">glom</a></strong>()</code> </td> +</tr> +<tr class="rowColor"> +<td class="colFirst"><code>static <K> <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><scala.Tuple2<K,scala.collection.Iterable<T>>></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#groupBy(scala.Function1,%20scala.reflect.ClassTag)">groupBy</a></strong>(scala.Function1<T,K> f, + scala.reflect.ClassTag<K> kt)</code> </td> +</tr> +<tr class="altColor"> +<td class="colFirst"><code>static <K> <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><scala.Tuple2<K,scala.collection.Iterable<T>>></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#groupBy(scala.Function1,%20int,%20scala.reflect.ClassTag)">groupBy</a></strong>(scala.Function1<T,K> f, + int numPartitions, + scala.reflect.ClassTag<K> kt)</code> </td> +</tr> +<tr class="rowColor"> +<td class="colFirst"><code>static <K> <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><scala.Tuple2<K,scala.collection.Iterable<T>>></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#groupBy(scala.Function1,%20org.apache.spark.Partitioner,%20scala.reflect.ClassTag,%20scala.math.Ordering)">groupBy</a></strong>(scala.Function1<T,K> f, + <a href="../../../../../org/apache/spark/Partitioner.html" title="class in org.apache.spark">Partitioner</a> p, + scala.reflect.ClassTag<K> kt, + scala.math.Ordering<K> ord)</code> </td> +</tr> +<tr class="altColor"> +<td class="colFirst"><code>static <K> scala.runtime.Null$</code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#groupBy$default$4(scala.Function1,%20org.apache.spark.Partitioner)">groupBy$default$4</a></strong>(scala.Function1<T,K> f, + <a href="../../../../../org/apache/spark/Partitioner.html" title="class in org.apache.spark">Partitioner</a> p)</code> </td> +</tr> +<tr class="rowColor"> +<td class="colFirst"><code>static int</code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#id()">id</a></strong>()</code> </td> +</tr> +<tr class="altColor"> +<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><T></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#intersection(org.apache.spark.rdd.RDD)">intersection</a></strong>(<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><T> other)</code> </td> +</tr> +<tr class="rowColor"> +<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><T></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#intersection(org.apache.spark.rdd.RDD,%20int)">intersection</a></strong>(<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><T> other, + int numPartitions)</code> </td> +</tr> +<tr class="altColor"> +<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><T></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#intersection(org.apache.spark.rdd.RDD,%20org.apache.spark.Partitioner,%20scala.math.Ordering)">intersection</a></strong>(<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><T> other, + <a href="../../../../../org/apache/spark/Partitioner.html" title="class in org.apache.spark">Partitioner</a> partitioner, + scala.math.Ordering<T> ord)</code> </td> +</tr> +<tr class="rowColor"> +<td class="colFirst"><code>static scala.math.Ordering<T></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#intersection$default$3(org.apache.spark.rdd.RDD,%20org.apache.spark.Partitioner)">intersection$default$3</a></strong>(<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><T> other, + <a href="../../../../../org/apache/spark/Partitioner.html" title="class in org.apache.spark">Partitioner</a> partitioner)</code> </td> +</tr> +<tr class="altColor"> +<td class="colFirst"><code>static boolean</code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#isCheckpointed()">isCheckpointed</a></strong>()</code> </td> +</tr> +<tr class="rowColor"> +<td class="colFirst"><code>static boolean</code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#isEmpty()">isEmpty</a></strong>()</code> </td> +</tr> +<tr class="altColor"> +<td class="colFirst"><code>static scala.collection.Iterator<T></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#iterator(org.apache.spark.Partition,%20org.apache.spark.TaskContext)">iterator</a></strong>(<a href="../../../../../org/apache/spark/Partition.html" title="interface in org.apache.spark">Partition</a> split, + <a href="../../../../../org/apache/spark/TaskContext.html" title="class in org.apache.spark">TaskContext</a> context)</code> </td> +</tr> +<tr class="rowColor"> +<td class="colFirst"><code>static <K> <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><scala.Tuple2<K,T>></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#keyBy(scala.Function1)">keyBy</a></strong>(scala.Function1<T,K> f)</code> </td> +</tr> +<tr class="altColor"> +<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><T></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#localCheckpoint()">localCheckpoint</a></strong>()</code> </td> +</tr> +<tr class="rowColor"> +<td class="colFirst"><code>static <U> <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><U></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#map(scala.Function1,%20scala.reflect.ClassTag)">map</a></strong>(scala.Function1<T,U> f, + scala.reflect.ClassTag<U> evidence$3)</code> </td> +</tr> +<tr class="altColor"> +<td class="colFirst"><code>static <U> <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><U></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#mapPartitions(scala.Function1,%20boolean,%20scala.reflect.ClassTag)">mapPartitions</a></strong>(scala.Function1<scala.collection.Iterator<T>,scala.collection.Iterator<U>> f, + boolean preservesPartitioning, + scala.reflect.ClassTag<U> evidence$6)</code> </td> +</tr> +<tr class="rowColor"> +<td class="colFirst"><code>static <U> boolean</code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#mapPartitions$default$2()">mapPartitions$default$2</a></strong>()</code> </td> +</tr> +<tr class="altColor"> +<td class="colFirst"><code>static <U> boolean</code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#mapPartitionsInternal$default$2()">mapPartitionsInternal$default$2</a></strong>()</code> </td> +</tr> +<tr class="rowColor"> +<td class="colFirst"><code>static <U> <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><U></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#mapPartitionsWithIndex(scala.Function2,%20boolean,%20scala.reflect.ClassTag)">mapPartitionsWithIndex</a></strong>(scala.Function2<Object,scala.collection.Iterator<T>,scala.collection.Iterator<U>> f, + boolean preservesPartitioning, + scala.reflect.ClassTag<U> evidence$9)</code> </td> +</tr> +<tr class="altColor"> +<td class="colFirst"><code>static <U> boolean</code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#mapPartitionsWithIndex$default$2()">mapPartitionsWithIndex$default$2</a></strong>()</code> </td> +</tr> +<tr class="rowColor"> +<td class="colFirst"><code>static <U> boolean</code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#mapPartitionsWithIndexInternal$default$2()">mapPartitionsWithIndexInternal$default$2</a></strong>()</code> </td> +</tr> +<tr class="altColor"> +<td class="colFirst"><code>static T</code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#max(scala.math.Ordering)">max</a></strong>(scala.math.Ordering<T> ord)</code> </td> +</tr> +<tr class="rowColor"> +<td class="colFirst"><code>static T</code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#min(scala.math.Ordering)">min</a></strong>(scala.math.Ordering<T> ord)</code> </td> +</tr> +<tr class="altColor"> +<td class="colFirst"><code>static void</code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#name_$eq(java.lang.String)">name_$eq</a></strong>(String x$1)</code> </td> +</tr> +<tr class="rowColor"> +<td class="colFirst"><code>static String</code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#name()">name</a></strong>()</code> </td> +</tr> +<tr class="altColor"> +<td class="colFirst"><code>static scala.Option<<a href="../../../../../org/apache/spark/Partitioner.html" title="class in org.apache.spark">Partitioner</a>></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#partitioner()">partitioner</a></strong>()</code> </td> +</tr> +<tr class="rowColor"> +<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/Partition.html" title="interface in org.apache.spark">Partition</a>[]</code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#partitions()">partitions</a></strong>()</code> </td> +</tr> +<tr class="altColor"> +<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><T></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#persist()">persist</a></strong>()</code> </td> +</tr> +<tr class="rowColor"> +<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><T></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#persist(org.apache.spark.storage.StorageLevel)">persist</a></strong>(<a href="../../../../../org/apache/spark/storage/StorageLevel.html" title="class in org.apache.spark.storage">StorageLevel</a> newLevel)</code> </td> +</tr> +<tr class="altColor"> +<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><String></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#pipe(scala.collection.Seq,%20scala.collection.Map,%20scala.Function1,%20scala.Function2,%20boolean,%20int,%20java.lang.String)">pipe</a></strong>(scala.collection.Seq<String> command, + scala.collection.Map<String,String> env, + scala.Function1<scala.Function1<String,scala.runtime.BoxedUnit>,scala.runtime.BoxedUnit> printPipeContext, + scala.Function2<T,scala.Function1<String,scala.runtime.BoxedUnit>,scala.runtime.BoxedUnit> printRDDElement, + boolean separateWorkingDir, + int bufferSize, + String encoding)</code> </td> +</tr> +<tr class="rowColor"> +<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><String></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#pipe(java.lang.String)">pipe</a></strong>(String command)</code> </td> +</tr> +<tr class="altColor"> +<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><String></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#pipe(java.lang.String,%20scala.collection.Map)">pipe</a></strong>(String command, + scala.collection.Map<String,String> env)</code> </td> +</tr> +<tr class="rowColor"> +<td class="colFirst"><code>static scala.collection.Map<String,String></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#pipe$default$2()">pipe$default$2</a></strong>()</code> </td> +</tr> +<tr class="altColor"> +<td class="colFirst"><code>static scala.Function1<scala.Function1<String,scala.runtime.BoxedUnit>,scala.runtime.BoxedUnit></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#pipe$default$3()">pipe$default$3</a></strong>()</code> </td> +</tr> +<tr class="rowColor"> +<td class="colFirst"><code>static scala.Function2<T,scala.Function1<String,scala.runtime.BoxedUnit>,scala.runtime.BoxedUnit></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#pipe$default$4()">pipe$default$4</a></strong>()</code> </td> +</tr> +<tr class="altColor"> +<td class="colFirst"><code>static boolean</code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#pipe$default$5()">pipe$default$5</a></strong>()</code> </td> +</tr> +<tr class="rowColor"> +<td class="colFirst"><code>static int</code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#pipe$default$6()">pipe$default$6</a></strong>()</code> </td> +</tr> +<tr class="altColor"> +<td class="colFirst"><code>static String</code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#pipe$default$7()">pipe$default$7</a></strong>()</code> </td> +</tr> +<tr class="rowColor"> +<td class="colFirst"><code>static scala.collection.Seq<String></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#preferredLocations(org.apache.spark.Partition)">preferredLocations</a></strong>(<a href="../../../../../org/apache/spark/Partition.html" title="interface in org.apache.spark">Partition</a> split)</code> </td> +</tr> +<tr class="altColor"> +<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><T>[]</code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#randomSplit(double[],%20long)">randomSplit</a></strong>(double[] weights, + long seed)</code> </td> +</tr> +<tr class="rowColor"> +<td class="colFirst"><code>static long</code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#randomSplit$default$2()">randomSplit$default$2</a></strong>()</code> </td> +</tr> +<tr class="altColor"> +<td class="colFirst"><code>static T</code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#reduce(scala.Function2)">reduce</a></strong>(scala.Function2<T,T,T> f)</code> </td> +</tr> +<tr class="rowColor"> +<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><T></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#repartition(int,%20scala.math.Ordering)">repartition</a></strong>(int numPartitions, + scala.math.Ordering<T> ord)</code> </td> +</tr> +<tr class="altColor"> +<td class="colFirst"><code>static scala.math.Ordering<T></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#repartition$default$2(int)">repartition$default$2</a></strong>(int numPartitions)</code> </td> +</tr> +<tr class="rowColor"> +<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><T></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#sample(boolean,%20double,%20long)">sample</a></strong>(boolean withReplacement, + double fraction, + long seed)</code> </td> +</tr> +<tr class="altColor"> +<td class="colFirst"><code>static long</code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#sample$default$3()">sample$default$3</a></strong>()</code> </td> +</tr> +<tr class="rowColor"> +<td class="colFirst"><code>static void</code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#saveAsObjectFile(java.lang.String)">saveAsObjectFile</a></strong>(String path)</code> </td> +</tr> +<tr class="altColor"> +<td class="colFirst"><code>static void</code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#saveAsTextFile(java.lang.String)">saveAsTextFile</a></strong>(String path)</code> </td> +</tr> +<tr class="rowColor"> +<td class="colFirst"><code>static void</code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#saveAsTextFile(java.lang.String,%20java.lang.Class)">saveAsTextFile</a></strong>(String path, + Class<? extends org.apache.hadoop.io.compress.CompressionCodec> codec)</code> </td> +</tr> +<tr class="altColor"> +<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><T></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#setName(java.lang.String)">setName</a></strong>(String _name)</code> </td> +</tr> +<tr class="rowColor"> +<td class="colFirst"><code>static <K> <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><T></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#sortBy(scala.Function1,%20boolean,%20int,%20scala.math.Ordering,%20scala.reflect.ClassTag)">sortBy</a></strong>(scala.Function1<T,K> f, + boolean ascending, + int numPartitions, + scala.math.Ordering<K> ord, + scala.reflect.ClassTag<K> ctag)</code> </td> +</tr> +<tr class="altColor"> +<td class="colFirst"><code>static <K> boolean</code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#sortBy$default$2()">sortBy$default$2</a></strong>()</code> </td> +</tr> +<tr class="rowColor"> +<td class="colFirst"><code>static <K> int</code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#sortBy$default$3()">sortBy$default$3</a></strong>()</code> </td> +</tr> +<tr class="altColor"> +<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/SparkContext.html" title="class in org.apache.spark">SparkContext</a></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#sparkContext()">sparkContext</a></strong>()</code> </td> +</tr> +<tr class="rowColor"> +<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><T></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#subtract(org.apache.spark.rdd.RDD)">subtract</a></strong>(<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><T> other)</code> </td> +</tr> +<tr class="altColor"> +<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><T></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#subtract(org.apache.spark.rdd.RDD,%20int)">subtract</a></strong>(<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><T> other, + int numPartitions)</code> </td> +</tr> +<tr class="rowColor"> +<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><T></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#subtract(org.apache.spark.rdd.RDD,%20org.apache.spark.Partitioner,%20scala.math.Ordering)">subtract</a></strong>(<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><T> other, + <a href="../../../../../org/apache/spark/Partitioner.html" title="class in org.apache.spark">Partitioner</a> p, + scala.math.Ordering<T> ord)</code> </td> +</tr> +<tr class="altColor"> +<td class="colFirst"><code>static scala.math.Ordering<T></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#subtract$default$3(org.apache.spark.rdd.RDD,%20org.apache.spark.Partitioner)">subtract$default$3</a></strong>(<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><T> other, + <a href="../../../../../org/apache/spark/Partitioner.html" title="class in org.apache.spark">Partitioner</a> p)</code> </td> +</tr> +<tr class="rowColor"> +<td class="colFirst"><code>static Object</code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#take(int)">take</a></strong>(int num)</code> </td> +</tr> +<tr class="altColor"> +<td class="colFirst"><code>static Object</code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#takeOrdered(int,%20scala.math.Ordering)">takeOrdered</a></strong>(int num, + scala.math.Ordering<T> ord)</code> </td> +</tr> +<tr class="rowColor"> +<td class="colFirst"><code>static Object</code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#takeSample(boolean,%20int,%20long)">takeSample</a></strong>(boolean withReplacement, + int num, + long seed)</code> </td> +</tr> +<tr class="altColor"> +<td class="colFirst"><code>static long</code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#takeSample$default$3()">takeSample$default$3</a></strong>()</code> </td> +</tr> +<tr class="rowColor"> +<td class="colFirst"><code>static String</code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#toDebugString()">toDebugString</a></strong>()</code> </td> +</tr> +<tr class="altColor"> +<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/api/java/JavaRDD.html" title="class in org.apache.spark.api.java">JavaRDD</a><T></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#toJavaRDD()">toJavaRDD</a></strong>()</code> </td> +</tr> +<tr class="rowColor"> +<td class="colFirst"><code>static scala.collection.Iterator<T></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#toLocalIterator()">toLocalIterator</a></strong>()</code> </td> +</tr> +<tr class="altColor"> +<td class="colFirst"><code>static Object</code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#top(int,%20scala.math.Ordering)">top</a></strong>(int num, + scala.math.Ordering<T> ord)</code> </td> +</tr> +<tr class="rowColor"> +<td class="colFirst"><code>static String</code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#toString()">toString</a></strong>()</code> </td> +</tr> +<tr class="altColor"> +<td class="colFirst"><code>static <U> U</code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#treeAggregate(U,%20scala.Function2,%20scala.Function2,%20int,%20scala.reflect.ClassTag)">treeAggregate</a></strong>(U zeroValue, + scala.Function2<U,T,U> seqOp, + scala.Function2<U,U,U> combOp, + int depth, + scala.reflect.ClassTag<U> evidence$31)</code> </td> +</tr> +<tr class="rowColor"> +<td class="colFirst"><code>static <U> int</code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#treeAggregate$default$4(U)">treeAggregate$default$4</a></strong>(U zeroValue)</code> </td> +</tr> +<tr class="altColor"> +<td class="colFirst"><code>static T</code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#treeReduce(scala.Function2,%20int)">treeReduce</a></strong>(scala.Function2<T,T,T> f, + int depth)</code> </td> +</tr> +<tr class="rowColor"> +<td class="colFirst"><code>static int</code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#treeReduce$default$2()">treeReduce$default$2</a></strong>()</code> </td> +</tr> +<tr class="altColor"> +<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><T></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#union(org.apache.spark.rdd.RDD)">union</a></strong>(<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><T> other)</code> </td> +</tr> +<tr class="rowColor"> +<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><T></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#unpersist(boolean)">unpersist</a></strong>(boolean blocking)</code> </td> +</tr> +<tr class="altColor"> +<td class="colFirst"><code>static boolean</code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#unpersist$default$1()">unpersist$default$1</a></strong>()</code> </td> +</tr> +<tr class="rowColor"> +<td class="colFirst"><code>static <U> <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><scala.Tuple2<T,U>></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#zip(org.apache.spark.rdd.RDD,%20scala.reflect.ClassTag)">zip</a></strong>(<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><U> other, + scala.reflect.ClassTag<U> evidence$10)</code> </td> +</tr> +<tr class="altColor"> +<td class="colFirst"><code>static <B,V> <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><V></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#zipPartitions(org.apache.spark.rdd.RDD,%20boolean,%20scala.Function2,%20scala.reflect.ClassTag,%20scala.reflect.ClassTag)">zipPartitions</a></strong>(<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><B> rdd2, + boolean preservesPartitioning, + scala.Function2<scala.collection.Iterator<T>,scala.collection.Iterator<B>,scala.collection.Iterator<V>> f, + scala.reflect.ClassTag<B> evidence$11, + scala.reflect.ClassTag<V> evidence$12)</code> </td> +</tr> +<tr class="rowColor"> +<td class="colFirst"><code>static <B,V> <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><V></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#zipPartitions(org.apache.spark.rdd.RDD,%20scala.Function2,%20scala.reflect.ClassTag,%20scala.reflect.ClassTag)">zipPartitions</a></strong>(<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><B> rdd2, + scala.Function2<scala.collection.Iterator<T>,scala.collection.Iterator<B>,scala.collection.Iterator<V>> f, + scala.reflect.ClassTag<B> evidence$13, + scala.reflect.ClassTag<V> evidence$14)</code> </td> +</tr> +<tr class="altColor"> +<td class="colFirst"><code>static <B,C,V> <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><V></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#zipPartitions(org.apache.spark.rdd.RDD,%20org.apache.spark.rdd.RDD,%20boolean,%20scala.Function3,%20scala.reflect.ClassTag,%20scala.reflect.ClassTag,%20scala.reflect.ClassTag)">zipPartitions</a></strong>(<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><B> rdd2, + <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><C> rdd3, + boolean preservesPartitioning, + scala.Function3<scala.collection.Iterator<T>,scala.collection.Iterator<B>,scala.collection.Iterator<C>,scala.collection.Iterator<V>> f, + scala.reflect.ClassTag<B> evidence$15, + scala.reflect.ClassTag<C> evidence$16, + scala.reflect.ClassTag<V> evidence$17)</code> </td> +</tr> +<tr class="rowColor"> +<td class="colFirst"><code>static <B,C,V> <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><V></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#zipPartitions(org.apache.spark.rdd.RDD,%20org.apache.spark.rdd.RDD,%20scala.Function3,%20scala.reflect.ClassTag,%20scala.reflect.ClassTag,%20scala.reflect.ClassTag)">zipPartitions</a></strong>(<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><B> rdd2, + <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><C> rdd3, + scala.Function3<scala.collection.Iterator<T>,scala.collection.Iterator<B>,scala.collection.Iterator<C>,scala.collection.Iterator<V>> f, + scala.reflect.ClassTag<B> evidence$18, + scala.reflect.ClassTag<C> evidence$19, + scala.reflect.ClassTag<V> evidence$20)</code> </td> +</tr> +<tr class="altColor"> +<td class="colFirst"><code>static <B,C,D,V> <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><V></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#zipPartitions(org.apache.spark.rdd.RDD,%20org.apache.spark.rdd.RDD,%20org.apache.spark.rdd.RDD,%20boolean,%20scala.Function4,%20scala.reflect.ClassTag,%20scala.reflect.ClassTag,%20scala.reflect.ClassTag,%20scala.reflect.ClassTag)">zipPartitions</a></strong>(<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><B> rdd2, + <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><C> rdd3, + <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><D> rdd4, + boolean preservesPartitioning, + scala.Function4<scala.collection.Iterator<T>,scala.collection.Iterator<B>,scala.collection.Iterator<C>,scala.collection.Iterator<D>,scala.collection.Iterator<V>> f, + scala.reflect.ClassTag<B> evidence$21, + scala.reflect.ClassTag<C> evidence$22, + scala.reflect.ClassTag<D> evidence$23, + scala.reflect.ClassTag<V> evidence$24)</code> </td> +</tr> +<tr class="rowColor"> +<td class="colFirst"><code>static <B,C,D,V> <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><V></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#zipPartitions(org.apache.spark.rdd.RDD,%20org.apache.spark.rdd.RDD,%20org.apache.spark.rdd.RDD,%20scala.Function4,%20scala.reflect.ClassTag,%20scala.reflect.ClassTag,%20scala.reflect.ClassTag,%20scala.reflect.ClassTag)">zipPartitions</a></strong>(<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><B> rdd2, + <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><C> rdd3, + <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><D> rdd4, + scala.Function4<scala.collection.Iterator<T>,scala.collection.Iterator<B>,scala.collection.Iterator<C>,scala.collection.Iterator<D>,scala.collection.Iterator<V>> f, + scala.reflect.ClassTag<B> evidence$25, + scala.reflect.ClassTag<C> evidence$26, + scala.reflect.ClassTag<D> evidence$27, + scala.reflect.ClassTag<V> evidence$28)</code> </td> +</tr> +<tr class="altColor"> +<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><scala.Tuple2<T,Object>></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#zipWithIndex()">zipWithIndex</a></strong>()</code> </td> +</tr> +<tr class="rowColor"> +<td class="colFirst"><code>static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><scala.Tuple2<T,Object>></code></td> +<td class="colLast"><code><strong><a href="../../../../../org/apache/spark/api/r/RRDD.html#zipWithUniqueId()">zipWithUniqueId</a></strong>()</code> </td> +</tr> +</table> +<ul class="blockList"> +<li class="blockList"><a name="methods_inherited_from_class_org.apache.spark.api.r.BaseRRDD"> +<!-- --> +</a> +<h3>Methods inherited from class org.apache.spark.api.r.<a href="../../../../../org/apache/spark/api/r/BaseRRDD.html" title="class in org.apache.spark.api.r">BaseRRDD</a></h3> +<code><a href="../../../../../org/apache/spark/api/r/BaseRRDD.html#compute(org.apache.spark.Partition,%20org.apache.spark.TaskContext)">compute</a>, <a href="../../../../../org/apache/spark/api/r/BaseRRDD.html#getPartitions()">getPartitions</a></code></li> +</ul> +<ul class="blockList"> +<li class="blockList"><a name="methods_inherited_from_class_org.apache.spark.rdd.RDD"> +<!-- --> +</a> +<h3>Methods inherited from class org.apache.spark.rdd.<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a></h3> +<code><a href="../../../../../org/apache/spark/rdd/RDD.html#aggregate(U,%20scala.Function2,%20scala.Function2,%20scala.reflect.ClassTag)">aggregate</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#cache()">cache</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#cartesian(org.apache.spark.rdd.RDD,%20scala.reflect.ClassTag)">cartesian</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#checkpoint()">checkpoint</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#coalesce(int,%20boolean,%20scala.Option,%20scala.math.Ordering)">coalesce</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#collect()">collect</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#collect(scala.PartialFunction,%20scala.reflect.ClassTag)">collect</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#context()">context</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#count()">count</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#countApprox( long,%20double)">countApprox</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#countApproxDistinct(double)">countApproxDistinct</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#countApproxDistinct(int,%20int)">countApproxDistinct</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#countByValue(scala.math.Ordering)">countByValue</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#countByValueApprox(long,%20double,%20scala.math.Ordering)">countByValueApprox</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#dependencies()">dependencies</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#distinct()">distinct</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#distinct(int,%20scala.math.Ordering)">distinct</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#doubleRDDToDoubleRDDFunctions(org.apache.spark.rdd.RDD)">doubleRDDToDoubleRDDFunctions</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#filter(scala.Function1)"> filter</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#first()">first</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#flatMap(scala.Function1,%20scala.reflect.ClassTag)">flatMap</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#fold(T,%20scala.Function2)">fold</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#foreach(scala.Function1)">foreach</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#foreachPartition(scala.Function1)">foreachPartition</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#getCheckpointFile()">getCheckpointFile</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#getNumPartitions()">getNumPartitions</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#getStorageLevel()">getStorageLevel</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#glom()">glom</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#groupBy(scala.Function1,%20scala.reflect.ClassTag)">groupBy</a>, <a href="../../.. /../../org/apache/spark/rdd/RDD.html#groupBy(scala.Function1,%20int,%20scala.reflect.ClassTag)">groupBy</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#groupBy(scala.Function1,%20org.apache.spark.Partitioner,%20scala.reflect.ClassTag,%20scala.math.Ordering)">groupBy</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#id()">id</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#intersection(org.apache.spark.rdd.RDD)">intersection</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#intersection(org.apache.spark.rdd.RDD,%20int)">intersection</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#intersection(org.apache.spark.rdd.RDD,%20org.apache.spark.Partitioner,%20scala.math.Ordering)">intersection</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#isCheckpointed()">isCheckpointed</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#isEmpty()">isEmpty</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#iterator(org.apache.spar k.Partition,%20org.apache.spark.TaskContext)">iterator</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#keyBy(scala.Function1)">keyBy</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#localCheckpoint()">localCheckpoint</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#map(scala.Function1,%20scala.reflect.ClassTag)">map</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#mapPartitions(scala.Function1,%20boolean,%20scala.reflect.ClassTag)">mapPartitions</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#mapPartitionsWithIndex(scala.Function2,%20boolean,%20scala.reflect.ClassTag)">mapPartitionsWithIndex</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#max(scala.math.Ordering)">max</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#min(scala.math.Ordering)">min</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#name()">name</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#numericRDDToDoubleRDDFunctions(org.ap ache.spark.rdd.RDD,%20scala.math.Numeric)">numericRDDToDoubleRDDFunctions</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#partitioner()">partitioner</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#partitions()">partitions</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#persist()">persist</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#persist(org.apache.spark.storage.StorageLevel)">persist</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#pipe(scala.collection.Seq,%20scala.collection.Map,%20scala.Function1,%20scala.Function2,%20boolean,%20int,%20java.lang.String)">pipe</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#pipe(java.lang.String)">pipe</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#pipe(java.lang.String,%20scala.collection.Map)">pipe</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#preferredLocations(org.apache.spark.Partition)">preferredLocations</a>, <a href="../../../../../org/apache/sp ark/rdd/RDD.html#randomSplit(double[],%20long)">randomSplit</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#rddToAsyncRDDActions(org.apache.spark.rdd.RDD,%20scala.reflect.ClassTag)">rddToAsyncRDDActions</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#rddToOrderedRDDFunctions(org.apache.spark.rdd.RDD,%20scala.math.Ordering,%20scala.reflect.ClassTag,%20scala.reflect.ClassTag)">rddToOrderedRDDFunctions</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#rddToPairRDDFunctions(org.apache.spark.rdd.RDD,%20scala.reflect.ClassTag,%20scala.reflect.ClassTag,%20scala.math.Ordering)">rddToPairRDDFunctions</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#rddToSequenceFileRDDFunctions(org.apache.spark.rdd.RDD,%20scala.reflect.ClassTag,%20scala.reflect.ClassTag,%20,%20)">rddToSequenceFileRDDFunctions</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#reduce(scala.Function2)">reduce</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#repartition (int,%20scala.math.Ordering)">repartition</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#sample(boolean,%20double,%20long)">sample</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#saveAsObjectFile(java.lang.String)">saveAsObjectFile</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#saveAsTextFile(java.lang.String)">saveAsTextFile</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#saveAsTextFile(java.lang.String,%20java.lang.Class)">saveAsTextFile</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#setName(java.lang.String)">setName</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#sortBy(scala.Function1,%20boolean,%20int,%20scala.math.Ordering,%20scala.reflect.ClassTag)">sortBy</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#sparkContext()">sparkContext</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#subtract(org.apache.spark.rdd.RDD)">subtract</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#subt ract(org.apache.spark.rdd.RDD,%20int)">subtract</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#subtract(org.apache.spark.rdd.RDD,%20org.apache.spark.Partitioner,%20scala.math.Ordering)">subtract</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#take(int)">take</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#takeOrdered(int,%20scala.math.Ordering)">takeOrdered</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#takeSample(boolean,%20int,%20long)">takeSample</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#toDebugString()">toDebugString</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#toJavaRDD()">toJavaRDD</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#toLocalIterator()">toLocalIterator</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#top(int,%20scala.math.Ordering)">top</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#toString()">toString</a>, <a href="../../../../../org/apache/spark/rdd/RDD.h tml#treeAggregate(U,%20scala.Function2,%20scala.Function2,%20int,%20scala.reflect.ClassTag)">treeAggregate</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#treeReduce(scala.Function2,%20int)">treeReduce</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#union(org.apache.spark.rdd.RDD)">union</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#unpersist(boolean)">unpersist</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#zip(org.apache.spark.rdd.RDD,%20scala.reflect.ClassTag)">zip</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#zipPartitions(org.apache.spark.rdd.RDD,%20boolean,%20scala.Function2,%20scala.reflect.ClassTag,%20scala.reflect.ClassTag)">zipPartitions</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#zipPartitions(org.apache.spark.rdd.RDD,%20scala.Function2,%20scala.reflect.ClassTag,%20scala.reflect.ClassTag)">zipPartitions</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#zipPartitions(org.apache.spark.rdd.RD D,%20org.apache.spark.rdd.RDD,%20boolean,%20scala.Function3,%20scala.reflect.ClassTag,%20scala.reflect.ClassTag,%20scala.reflect.ClassTag)">zipPartitions</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#zipPartitions(org.apache.spark.rdd.RDD,%20org.apache.spark.rdd.RDD,%20scala.Function3,%20scala.reflect.ClassTag,%20scala.reflect.ClassTag,%20scala.reflect.ClassTag)">zipPartitions</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#zipPartitions(org.apache.spark.rdd.RDD,%20org.apache.spark.rdd.RDD,%20org.apache.spark.rdd.RDD,%20boolean,%20scala.Function4,%20scala.reflect.ClassTag,%20scala.reflect.ClassTag,%20scala.reflect.ClassTag,%20scala.reflect.ClassTag)">zipPartitions</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#zipPartitions(org.apache.spark.rdd.RDD,%20org.apache.spark.rdd.RDD,%20org.apache.spark.rdd.RDD,%20scala.Function4,%20scala.reflect.ClassTag,%20scala.reflect.ClassTag,%20scala.reflect.ClassTag,%20scala.reflect.ClassTag)">zipPartitions</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#zipWithIndex()">zipWithIndex</a>, <a href="../../../../../org/apache/spark/rdd/RDD.html#zipWithUniqueId()">zipWithUniqueId</a></code></li> +</ul> +<ul class="blockList"> +<li class="blockList"><a name="methods_inherited_from_class_Object"> +<!-- --> +</a> +<h3>Methods inherited from class Object</h3> +<code>equals, getClass, hashCode, notify, notifyAll, wait, wait, wait</code></li> +</ul> +</li> +</ul> +</li> +</ul> +</div> +<div class="details"> +<ul class="blockList"> +<li class="blockList"> +<!-- ========= CONSTRUCTOR DETAIL ======== --> +<ul class="blockList"> +<li class="blockList"><a name="constructor_detail"> +<!-- --> +</a> +<h3>Constructor Detail</h3> +<a name="RRDD(org.apache.spark.rdd.RDD, byte[], java.lang.String, java.lang.String, byte[], java.lang.Object[], scala.reflect.ClassTag)"> +<!-- --> +</a> +<ul class="blockListLast"> +<li class="blockList"> +<h4>RRDD</h4> +<pre>public RRDD(<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><<a href="../../../../../org/apache/spark/api/r/RRDD.html" title="type parameter in RRDD">T</a>> parent, + byte[] func, + String deserializer, + String serializer, + byte[] packageNames, + Object[] broadcastVars, + scala.reflect.ClassTag<<a href="../../../../../org/apache/spark/api/r/RRDD.html" title="type parameter in RRDD">T</a>> evidence$4)</pre> +</li> +</ul> +</li> +</ul> +<!-- ============ METHOD DETAIL ========== --> +<ul class="blockList"> +<li class="blockList"><a name="method_detail"> +<!-- --> +</a> +<h3>Method Detail</h3> +<a name="createSparkContext(java.lang.String, java.lang.String, java.lang.String, java.lang.String[], java.util.Map, java.util.Map)"> +<!-- --> +</a> +<ul class="blockList"> +<li class="blockList"> +<h4>createSparkContext</h4> +<pre>public static <a href="../../../../../org/apache/spark/api/java/JavaSparkContext.html" title="class in org.apache.spark.api.java">JavaSparkContext</a> createSparkContext(String master, + String appName, + String sparkHome, + String[] jars, + java.util.Map<Object,Object> sparkEnvirMap, + java.util.Map<Object,Object> sparkExecutorEnvMap)</pre> +</li> +</ul> +<a name="createRDDFromArray(org.apache.spark.api.java.JavaSparkContext, byte[][])"> +<!-- --> +</a> +<ul class="blockList"> +<li class="blockList"> +<h4>createRDDFromArray</h4> +<pre>public static <a href="../../../../../org/apache/spark/api/java/JavaRDD.html" title="class in org.apache.spark.api.java">JavaRDD</a><byte[]> createRDDFromArray(<a href="../../../../../org/apache/spark/api/java/JavaSparkContext.html" title="class in org.apache.spark.api.java">JavaSparkContext</a> jsc, + byte[][] arr)</pre> +<div class="block">Create an RRDD given a sequence of byte arrays. Used to create RRDD when <code>parallelize</code> is + called from R.</div> +<dl><dt><span class="strong">Parameters:</span></dt><dd><code>jsc</code> - (undocumented)</dd><dd><code>arr</code> - (undocumented)</dd> +<dt><span class="strong">Returns:</span></dt><dd>(undocumented)</dd></dl> +</li> +</ul> +<a name="createRDDFromFile(org.apache.spark.api.java.JavaSparkContext, java.lang.String, int)"> +<!-- --> +</a> +<ul class="blockList"> +<li class="blockList"> +<h4>createRDDFromFile</h4> +<pre>public static <a href="../../../../../org/apache/spark/api/java/JavaRDD.html" title="class in org.apache.spark.api.java">JavaRDD</a><byte[]> createRDDFromFile(<a href="../../../../../org/apache/spark/api/java/JavaSparkContext.html" title="class in org.apache.spark.api.java">JavaSparkContext</a> jsc, + String fileName, + int parallelism)</pre> +<div class="block">Create an RRDD given a temporary file name. This is used to create RRDD when parallelize is + called on large R objects. + <p></div> +<dl><dt><span class="strong">Parameters:</span></dt><dd><code>fileName</code> - name of temporary file on driver machine</dd><dd><code>parallelism</code> - number of slices defaults to 4</dd><dd><code>jsc</code> - (undocumented)</dd> +<dt><span class="strong">Returns:</span></dt><dd>(undocumented)</dd></dl> +</li> +</ul> +<a name="partitioner()"> +<!-- --> +</a> +<ul class="blockList"> +<li class="blockList"> +<h4>partitioner</h4> +<pre>public static scala.Option<<a href="../../../../../org/apache/spark/Partitioner.html" title="class in org.apache.spark">Partitioner</a>> partitioner()</pre> +</li> +</ul> +<a name="sparkContext()"> +<!-- --> +</a> +<ul class="blockList"> +<li class="blockList"> +<h4>sparkContext</h4> +<pre>public static <a href="../../../../../org/apache/spark/SparkContext.html" title="class in org.apache.spark">SparkContext</a> sparkContext()</pre> +</li> +</ul> +<a name="id()"> +<!-- --> +</a> +<ul class="blockList"> +<li class="blockList"> +<h4>id</h4> +<pre>public static int id()</pre> +</li> +</ul> +<a name="name()"> +<!-- --> +</a> +<ul class="blockList"> +<li class="blockList"> +<h4>name</h4> +<pre>public static String name()</pre> +</li> +</ul> +<a name="name_$eq(java.lang.String)"> +<!-- --> +</a> +<ul class="blockList"> +<li class="blockList"> +<h4>name_$eq</h4> +<pre>public static void name_$eq(String x$1)</pre> +</li> +</ul> +<a name="setName(java.lang.String)"> +<!-- --> +</a> +<ul class="blockList"> +<li class="blockList"> +<h4>setName</h4> +<pre>public static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><T> setName(String _name)</pre> +</li> +</ul> +<a name="persist(org.apache.spark.storage.StorageLevel)"> +<!-- --> +</a> +<ul class="blockList"> +<li class="blockList"> +<h4>persist</h4> +<pre>public static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><T> persist(<a href="../../../../../org/apache/spark/storage/StorageLevel.html" title="class in org.apache.spark.storage">StorageLevel</a> newLevel)</pre> +</li> +</ul> +<a name="persist()"> +<!-- --> +</a> +<ul class="blockList"> +<li class="blockList"> +<h4>persist</h4> +<pre>public static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><T> persist()</pre> +</li> +</ul> +<a name="cache()"> +<!-- --> +</a> +<ul class="blockList"> +<li class="blockList"> +<h4>cache</h4> +<pre>public static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><T> cache()</pre> +</li> +</ul> +<a name="unpersist(boolean)"> +<!-- --> +</a> +<ul class="blockList"> +<li class="blockList"> +<h4>unpersist</h4> +<pre>public static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><T> unpersist(boolean blocking)</pre> +</li> +</ul> +<a name="getStorageLevel()"> +<!-- --> +</a> +<ul class="blockList"> +<li class="blockList"> +<h4>getStorageLevel</h4> +<pre>public static <a href="../../../../../org/apache/spark/storage/StorageLevel.html" title="class in org.apache.spark.storage">StorageLevel</a> getStorageLevel()</pre> +</li> +</ul> +<a name="dependencies()"> +<!-- --> +</a> +<ul class="blockList"> +<li class="blockList"> +<h4>dependencies</h4> +<pre>public static final scala.collection.Seq<<a href="../../../../../org/apache/spark/Dependency.html" title="class in org.apache.spark">Dependency</a><?>> dependencies()</pre> +</li> +</ul> +<a name="partitions()"> +<!-- --> +</a> +<ul class="blockList"> +<li class="blockList"> +<h4>partitions</h4> +<pre>public static final <a href="../../../../../org/apache/spark/Partition.html" title="interface in org.apache.spark">Partition</a>[] partitions()</pre> +</li> +</ul> +<a name="getNumPartitions()"> +<!-- --> +</a> +<ul class="blockList"> +<li class="blockList"> +<h4>getNumPartitions</h4> +<pre>public static final int getNumPartitions()</pre> +</li> +</ul> +<a name="preferredLocations(org.apache.spark.Partition)"> +<!-- --> +</a> +<ul class="blockList"> +<li class="blockList"> +<h4>preferredLocations</h4> +<pre>public static final scala.collection.Seq<String> preferredLocations(<a href="../../../../../org/apache/spark/Partition.html" title="interface in org.apache.spark">Partition</a> split)</pre> +</li> +</ul> +<a name="iterator(org.apache.spark.Partition, org.apache.spark.TaskContext)"> +<!-- --> +</a> +<ul class="blockList"> +<li class="blockList"> +<h4>iterator</h4> +<pre>public static final scala.collection.Iterator<T> iterator(<a href="../../../../../org/apache/spark/Partition.html" title="interface in org.apache.spark">Partition</a> split, + <a href="../../../../../org/apache/spark/TaskContext.html" title="class in org.apache.spark">TaskContext</a> context)</pre> +</li> +</ul> +<a name="map(scala.Function1, scala.reflect.ClassTag)"> +<!-- --> +</a> +<ul class="blockList"> +<li class="blockList"> +<h4>map</h4> +<pre>public static <U> <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><U> map(scala.Function1<T,U> f, + scala.reflect.ClassTag<U> evidence$3)</pre> +</li> +</ul> +<a name="flatMap(scala.Function1, scala.reflect.ClassTag)"> +<!-- --> +</a> +<ul class="blockList"> +<li class="blockList"> +<h4>flatMap</h4> +<pre>public static <U> <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><U> flatMap(scala.Function1<T,scala.collection.TraversableOnce<U>> f, + scala.reflect.ClassTag<U> evidence$4)</pre> +</li> +</ul> +<a name="filter(scala.Function1)"> +<!-- --> +</a> +<ul class="blockList"> +<li class="blockList"> +<h4>filter</h4> +<pre>public static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><T> filter(scala.Function1<T,Object> f)</pre> +</li> +</ul> +<a name="distinct(int, scala.math.Ordering)"> +<!-- --> +</a> +<ul class="blockList"> +<li class="blockList"> +<h4>distinct</h4> +<pre>public static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><T> distinct(int numPartitions, + scala.math.Ordering<T> ord)</pre> +</li> +</ul> +<a name="distinct()"> +<!-- --> +</a> +<ul class="blockList"> +<li class="blockList"> +<h4>distinct</h4> +<pre>public static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><T> distinct()</pre> +</li> +</ul> +<a name="repartition(int, scala.math.Ordering)"> +<!-- --> +</a> +<ul class="blockList"> +<li class="blockList"> +<h4>repartition</h4> +<pre>public static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><T> repartition(int numPartitions, + scala.math.Ordering<T> ord)</pre> +</li> +</ul> +<a name="coalesce(int, boolean, scala.Option, scala.math.Ordering)"> +<!-- --> +</a> +<ul class="blockList"> +<li class="blockList"> +<h4>coalesce</h4> +<pre>public static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><T> coalesce(int numPartitions, + boolean shuffle, + scala.Option<<a href="../../../../../org/apache/spark/rdd/PartitionCoalescer.html" title="interface in org.apache.spark.rdd">PartitionCoalescer</a>> partitionCoalescer, + scala.math.Ordering<T> ord)</pre> +</li> +</ul> +<a name="sample(boolean, double, long)"> +<!-- --> +</a> +<ul class="blockList"> +<li class="blockList"> +<h4>sample</h4> +<pre>public static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><T> sample(boolean withReplacement, + double fraction, + long seed)</pre> +</li> +</ul> +<a name="randomSplit(double[], long)"> +<!-- --> +</a> +<ul class="blockList"> +<li class="blockList"> +<h4>randomSplit</h4> +<pre>public static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><T>[] randomSplit(double[] weights, + long seed)</pre> +</li> +</ul> +<a name="takeSample(boolean, int, long)"> +<!-- --> +</a> +<ul class="blockList"> +<li class="blockList"> +<h4>takeSample</h4> +<pre>public static Object takeSample(boolean withReplacement, + int num, + long seed)</pre> +</li> +</ul> +<a name="union(org.apache.spark.rdd.RDD)"> +<!-- --> +</a> +<ul class="blockList"> +<li class="blockList"> +<h4>union</h4> +<pre>public static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><T> union(<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><T> other)</pre> +</li> +</ul> +<a name="$plus$plus(org.apache.spark.rdd.RDD)"> +<!-- --> +</a> +<ul class="blockList"> +<li class="blockList"> +<h4>$plus$plus</h4> +<pre>public static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><T> $plus$plus(<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><T> other)</pre> +</li> +</ul> +<a name="sortBy(scala.Function1, boolean, int, scala.math.Ordering, scala.reflect.ClassTag)"> +<!-- --> +</a> +<ul class="blockList"> +<li class="blockList"> +<h4>sortBy</h4> +<pre>public static <K> <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><T> sortBy(scala.Function1<T,K> f, + boolean ascending, + int numPartitions, + scala.math.Ordering<K> ord, + scala.reflect.ClassTag<K> ctag)</pre> +</li> +</ul> +<a name="intersection(org.apache.spark.rdd.RDD)"> +<!-- --> +</a> +<ul class="blockList"> +<li class="blockList"> +<h4>intersection</h4> +<pre>public static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><T> intersection(<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><T> other)</pre> +</li> +</ul> +<a name="intersection(org.apache.spark.rdd.RDD, org.apache.spark.Partitioner, scala.math.Ordering)"> +<!-- --> +</a> +<ul class="blockList"> +<li class="blockList"> +<h4>intersection</h4> +<pre>public static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><T> intersection(<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><T> other, + <a href="../../../../../org/apache/spark/Partitioner.html" title="class in org.apache.spark">Partitioner</a> partitioner, + scala.math.Ordering<T> ord)</pre> +</li> +</ul> +<a name="intersection(org.apache.spark.rdd.RDD, int)"> +<!-- --> +</a> +<ul class="blockList"> +<li class="blockList"> +<h4>intersection</h4> +<pre>public static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><T> intersection(<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><T> other, + int numPartitions)</pre> +</li> +</ul> +<a name="glom()"> +<!-- --> +</a> +<ul class="blockList"> +<li class="blockList"> +<h4>glom</h4> +<pre>public static <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><Object> glom()</pre> +</li> +</ul> +<a name="cartesian(org.apache.spark.rdd.RDD, scala.reflect.ClassTag)"> +<!-- --> +</a> +<ul class="blockList"> +<li class="blockList"> +<h4>cartesian</h4> +<pre>public static <U> <a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><scala.Tuple2<T,U>> cartesian(<a href="../../../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a><U> other, + scala.reflect.ClassTag<U <TRUNCATED> --------------------------------------------------------------------- To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org