Spark 4.0.0 ScalaDoc - org.apache.spark.rdd.RDD (original) (raw)
abstract class RDD[T] extends Serializable with Logging
Ordering
- Alphabetic
- By Inheritance
Inherited
RDD
Logging
Serializable
AnyRef
Any
Hide All
Show All
Visibility
- Public
- Protected
Instance Constructors
- new RDD(oneParent: RDD[_])(implicit arg0: ClassTag[T])
- new RDD(_sc: SparkContext, deps: Seq[Dependency[_]])(implicit arg0: ClassTag[T])
Type Members
- implicit class LogStringContext extends AnyRef
Concrete Value Members
- final def !=(arg0: Any): Boolean
- final def ##: Int
- def ++(other: RDD[T]): RDD[T]
- final def ==(arg0: Any): Boolean
- def aggregate[U](zeroValue: U)(seqOp: (U, T) => U, combOp: (U, U) => U)(implicit arg0: ClassTag[U]): U
- final def asInstanceOf[T0]: T0
- def barrier(): RDDBarrier[T]
- def cache(): RDD.this.type
- def cartesian[U](other: RDD[U])(implicit arg0: ClassTag[U]): RDD[(T, U)]
- def checkpoint(): Unit
- def cleanShuffleDependencies(blocking: Boolean = false): Unit
- def clearDependencies(): Unit
- def clone(): AnyRef
- def coalesce(numPartitions: Int, shuffle: Boolean = false, partitionCoalescer: Option[PartitionCoalescer] = Option.empty)(implicit ord: Ordering[T] = null): RDD[T]
- def collect[U](f: PartialFunction[T, U])(implicit arg0: ClassTag[U]): RDD[U]
- def collect(): Array[T]
- def context: SparkContext
- def count(): Long
- def countApprox(timeout: Long, confidence: Double = 0.95): PartialResult[BoundedDouble]
- def countApproxDistinct(relativeSD: Double = 0.05): Long
- def countApproxDistinct(p: Int, sp: Int): Long
- def countByValue()(implicit ord: Ordering[T] = null): Map[T, Long]
- def countByValueApprox(timeout: Long, confidence: Double = 0.95)(implicit ord: Ordering[T] = null): PartialResult[Map[T, BoundedDouble]]
- final def dependencies: Seq[Dependency[_]]
- def distinct(): RDD[T]
- def distinct(numPartitions: Int)(implicit ord: Ordering[T] = null): RDD[T]
- final def eq(arg0: AnyRef): Boolean
- def equals(arg0: AnyRef): Boolean
- def filter(f: (T) => Boolean): RDD[T]
- def first(): T
- def firstParent[U](implicit arg0: ClassTag[U]): RDD[U]
- def flatMap[U](f: (T) => IterableOnce[U])(implicit arg0: ClassTag[U]): RDD[U]
- def fold(zeroValue: T)(op: (T, T) => T): T
- def foreach(f: (T) => Unit): Unit
- def foreachPartition(f: (Iterator[T]) => Unit): Unit
- def getCheckpointFile: Option[String]
- final def getClass(): Class[_ <: AnyRef]
- def getDependencies: Seq[Dependency[_]]
- final def getNumPartitions: Int
- def getOutputDeterministicLevel: DeterministicLevel.Value
- def getPreferredLocations(split: Partition): Seq[String]
- def getResourceProfile(): ResourceProfile
- def getStorageLevel: StorageLevel
- def glom(): RDD[Array[T]]
- def groupBy[K](f: (T) => K, p: Partitioner)(implicit kt: ClassTag[K], ord: Ordering[K] = null): RDD[(K, Iterable[T])]
- def groupBy[K](f: (T) => K, numPartitions: Int)(implicit kt: ClassTag[K]): RDD[(K, Iterable[T])]
- def groupBy[K](f: (T) => K)(implicit kt: ClassTag[K]): RDD[(K, Iterable[T])]
- def hashCode(): Int
- val id: Int
- def initializeLogIfNecessary(isInterpreter: Boolean, silent: Boolean): Boolean
- def initializeLogIfNecessary(isInterpreter: Boolean): Unit
- def intersection(other: RDD[T], numPartitions: Int): RDD[T]
- def intersection(other: RDD[T], partitioner: Partitioner)(implicit ord: Ordering[T] = null): RDD[T]
- def intersection(other: RDD[T]): RDD[T]
- lazy val isBarrier_: Boolean
- def isCheckpointed: Boolean
- def isEmpty(): Boolean
- final def isInstanceOf[T0]: Boolean
- def isTraceEnabled(): Boolean
- final def iterator(split: Partition, context: TaskContext): Iterator[T]
- def keyBy[K](f: (T) => K): RDD[(K, T)]
- def localCheckpoint(): RDD.this.type
- def log: Logger
- def logDebug(msg: => String, throwable: Throwable): Unit
- def logDebug(entry: LogEntry, throwable: Throwable): Unit
- def logDebug(entry: LogEntry): Unit
- def logDebug(msg: => String): Unit
- def logError(msg: => String, throwable: Throwable): Unit
- def logError(entry: LogEntry, throwable: Throwable): Unit
- def logError(entry: LogEntry): Unit
- def logError(msg: => String): Unit
- def logInfo(msg: => String, throwable: Throwable): Unit
- def logInfo(entry: LogEntry, throwable: Throwable): Unit
- def logInfo(entry: LogEntry): Unit
- def logInfo(msg: => String): Unit
- def logName: String
- def logTrace(msg: => String, throwable: Throwable): Unit
- def logTrace(entry: LogEntry, throwable: Throwable): Unit
- def logTrace(entry: LogEntry): Unit
- def logTrace(msg: => String): Unit
- def logWarning(msg: => String, throwable: Throwable): Unit
- def logWarning(entry: LogEntry, throwable: Throwable): Unit
- def logWarning(entry: LogEntry): Unit
- def logWarning(msg: => String): Unit
- def map[U](f: (T) => U)(implicit arg0: ClassTag[U]): RDD[U]
- def mapPartitions[U](f: (Iterator[T]) => Iterator[U], preservesPartitioning: Boolean = false)(implicit arg0: ClassTag[U]): RDD[U]
- def mapPartitionsWithEvaluator[U](evaluatorFactory: PartitionEvaluatorFactory[T, U])(implicit arg0: ClassTag[U]): RDD[U]
- def mapPartitionsWithIndex[U](f: (Int, Iterator[T]) => Iterator[U], preservesPartitioning: Boolean = false)(implicit arg0: ClassTag[U]): RDD[U]
- def max()(implicit ord: Ordering[T]): T
- def min()(implicit ord: Ordering[T]): T
- var name: String
- final def ne(arg0: AnyRef): Boolean
- final def notify(): Unit
- final def notifyAll(): Unit
- def parent[U](j: Int)(implicit arg0: ClassTag[U]): RDD[U]
- val partitioner: Option[Partitioner]
- final def partitions: Array[Partition]
- def persist(): RDD.this.type
- def persist(newLevel: StorageLevel): RDD.this.type
- def pipe(command: Seq[String], env: Map[String, String] = Map(), printPipeContext: ((String) => Unit) => Unit = null, printRDDElement: (T, (String) => Unit) => Unit = null, separateWorkingDir: Boolean = false, bufferSize: Int = 8192, encoding: String = Codec.defaultCharsetCodec.name): RDD[String]
- def pipe(command: String, env: Map[String, String]): RDD[String]
- def pipe(command: String): RDD[String]
- final def preferredLocations(split: Partition): Seq[String]
- def randomSplit(weights: Array[Double], seed: Long = Utils.random.nextLong): Array[RDD[T]]
- def reduce(f: (T, T) => T): T
- def repartition(numPartitions: Int)(implicit ord: Ordering[T] = null): RDD[T]
- def sample(withReplacement: Boolean, fraction: Double, seed: Long = Utils.random.nextLong): RDD[T]
- def saveAsObjectFile(path: String): Unit
- def saveAsTextFile(path: String, codec: Class[_ <: CompressionCodec]): Unit
- def saveAsTextFile(path: String): Unit
- def setName(_name: String): RDD.this.type
- def sortBy[K](f: (T) => K, ascending: Boolean = true, numPartitions: Int = this.partitions.length)(implicit ord: Ordering[K], ctag: ClassTag[K]): RDD[T]
- def sparkContext: SparkContext
- def subtract(other: RDD[T], p: Partitioner)(implicit ord: Ordering[T] = null): RDD[T]
- def subtract(other: RDD[T], numPartitions: Int): RDD[T]
- def subtract(other: RDD[T]): RDD[T]
- final def synchronized[T0](arg0: => T0): T0
- def take(num: Int): Array[T]
- def takeOrdered(num: Int)(implicit ord: Ordering[T]): Array[T]
- def takeSample(withReplacement: Boolean, num: Int, seed: Long = Utils.random.nextLong): Array[T]
- def toDebugString: String
- def toJavaRDD(): JavaRDD[T]
- def toLocalIterator: Iterator[T]
- def toString(): String
- def top(num: Int)(implicit ord: Ordering[T]): Array[T]
- def treeAggregate[U](zeroValue: U, seqOp: (U, T) => U, combOp: (U, U) => U, depth: Int, finalAggregateOnExecutor: Boolean)(implicit arg0: ClassTag[U]): U
- def treeAggregate[U](zeroValue: U)(seqOp: (U, T) => U, combOp: (U, U) => U, depth: Int = 2)(implicit arg0: ClassTag[U]): U
- def treeReduce(f: (T, T) => T, depth: Int = 2): T
- def union(other: RDD[T]): RDD[T]
- def unpersist(blocking: Boolean = false): RDD.this.type
- final def wait(arg0: Long, arg1: Int): Unit
- final def wait(arg0: Long): Unit
- final def wait(): Unit
- def withLogContext(context: Map[String, String])(body: => Unit): Unit
- def withResources(rp: ResourceProfile): RDD.this.type
- def zip[U](other: RDD[U])(implicit arg0: ClassTag[U]): RDD[(T, U)]
- def zipPartitions[B, C, D, V](rdd2: RDD[B], rdd3: RDD[C], rdd4: RDD[D])(f: (Iterator[T], Iterator[B], Iterator[C], Iterator[D]) => Iterator[V])(implicit arg0: ClassTag[B], arg1: ClassTag[C], arg2: ClassTag[D], arg3: ClassTag[V]): RDD[V]
- def zipPartitions[B, C, D, V](rdd2: RDD[B], rdd3: RDD[C], rdd4: RDD[D], preservesPartitioning: Boolean)(f: (Iterator[T], Iterator[B], Iterator[C], Iterator[D]) => Iterator[V])(implicit arg0: ClassTag[B], arg1: ClassTag[C], arg2: ClassTag[D], arg3: ClassTag[V]): RDD[V]
- def zipPartitions[B, C, V](rdd2: RDD[B], rdd3: RDD[C])(f: (Iterator[T], Iterator[B], Iterator[C]) => Iterator[V])(implicit arg0: ClassTag[B], arg1: ClassTag[C], arg2: ClassTag[V]): RDD[V]
- def zipPartitions[B, C, V](rdd2: RDD[B], rdd3: RDD[C], preservesPartitioning: Boolean)(f: (Iterator[T], Iterator[B], Iterator[C]) => Iterator[V])(implicit arg0: ClassTag[B], arg1: ClassTag[C], arg2: ClassTag[V]): RDD[V]
- def zipPartitions[B, V](rdd2: RDD[B])(f: (Iterator[T], Iterator[B]) => Iterator[V])(implicit arg0: ClassTag[B], arg1: ClassTag[V]): RDD[V]
- def zipPartitions[B, V](rdd2: RDD[B], preservesPartitioning: Boolean)(f: (Iterator[T], Iterator[B]) => Iterator[V])(implicit arg0: ClassTag[B], arg1: ClassTag[V]): RDD[V]
- def zipPartitionsWithEvaluator[U](rdd2: RDD[T], evaluatorFactory: PartitionEvaluatorFactory[T, U])(implicit arg0: ClassTag[U]): RDD[U]
- def zipWithIndex(): RDD[(T, Long)]
- def zipWithUniqueId(): RDD[(T, Long)]
Deprecated Value Members
- def finalize(): Unit