Instance Constructors
-
new
BucketRDD
(prev: RDD[T], detector: BucketDetector[T])(implicit arg0: ClassManifest[T])
Value Members
-
def
!=
(arg0: AnyRef): Boolean
-
def
!=
(arg0: Any): Boolean
-
def
##
(): Int
-
def
++
(other: RDD[Array[T]]): RDD[Array[T]]
-
def
==
(arg0: AnyRef): Boolean
-
def
==
(arg0: Any): Boolean
-
def
asInstanceOf
[T0]
: T0
-
def
cache
(): RDD[Array[T]]
-
def
cartesian
[U]
(other: RDD[U])(implicit arg0: ClassManifest[U]): RDD[(Array[T], U)]
-
def
clone
(): AnyRef
-
def
collect
(): Array[Array[T]]
-
def
compute
(split: Split): Iterator[Array[T]]
-
def
context
: SparkContext
-
def
count
(): Long
-
val
dependencies
: List[OneToOneDependency[T]]
-
def
eq
(arg0: AnyRef): Boolean
-
def
equals
(arg0: Any): Boolean
-
def
filter
(f: (Array[T]) ⇒ Boolean): RDD[Array[T]]
-
def
finalize
(): Unit
-
def
first
(): Array[T]
-
def
flatMap
[U]
(f: (Array[T]) ⇒ Traversable[U])(implicit arg0: ClassManifest[U]): RDD[U]
-
def
foreach
(f: (Array[T]) ⇒ Unit): Unit
-
def
getClass
(): java.lang.Class[_]
-
def
glom
(): RDD[Array[Array[T]]]
-
def
groupBy
[K]
(f: (Array[T]) ⇒ K)(implicit arg0: ClassManifest[K]): RDD[(K, Seq[Array[T]])]
-
def
groupBy
[K]
(f: (Array[T]) ⇒ K, numSplits: Int)(implicit arg0: ClassManifest[K]): RDD[(K, Seq[Array[T]])]
-
def
hashCode
(): Int
-
val
id
: Int
-
def
isInstanceOf
[T0]
: Boolean
-
def
iterator
(split: Split): Iterator[Array[T]]
-
def
map
[U]
(f: (Array[T]) ⇒ U)(implicit arg0: ClassManifest[U]): RDD[U]
-
def
mapPartitions
[U]
(f: (Iterator[Array[T]]) ⇒ Iterator[U])(implicit arg0: ClassManifest[U]): RDD[U]
-
def
ne
(arg0: AnyRef): Boolean
-
def
notify
(): Unit
-
def
notifyAll
(): Unit
-
val
partitioner
: Option[Partitioner]
-
def
pipe
(command: Seq[String]): RDD[String]
-
def
pipe
(command: String): RDD[String]
-
def
preferredLocations
(split: Split): Seq[String]
-
val
prevSplits
: Array[Split]
-
def
reduce
(f: (Array[T], Array[T]) ⇒ Array[T]): Array[T]
-
def
sample
(withReplacement: Boolean, fraction: Double, seed: Int): RDD[Array[T]]
-
def
saveAsObjectFile
(path: String): Unit
-
def
saveAsTextFile
(path: String): Unit
-
def
splits
: Array[Split]
-
def
synchronized
[T0]
(arg0: ⇒ T0): T0
-
def
take
(num: Int): Array[Array[T]]
-
def
toArray
(): Array[Array[T]]
-
def
toString
(): String
-
def
union
(other: RDD[Array[T]]): RDD[Array[T]]
-
def
wait
(): Unit
-
def
wait
(arg0: Long, arg1: Int): Unit
-
def
wait
(arg0: Long): Unit
Inherited from RDD[Array[T]]
Inherited from Serializable
Inherited from Serializable
Inherited from AnyRef
Inherited from Any
A class that represents an RDD of Arrays of a specified type. Each Array is a partition or "bucket" of items that belong together. Usage:
See spark.timeseries.BucketLogsByHour for a more real example.
Accepts a BucketDetector to assign data items into buckets.