Packages

c

org.apache.spark

RangePartitioner

class RangePartitioner[K, V] extends Partitioner

A org.apache.spark.Partitioner that partitions sortable records by range into roughly equal ranges. The ranges are determined by sampling the content of the RDD passed in.

Source
Partitioner.scala
Note

The actual number of partitions created by the RangePartitioner might not be the same as the partitions parameter, in the case where the number of sampled records is less than the value of partitions.

Linear Supertypes
Partitioner, Serializable, Serializable, AnyRef, Any
Ordering
  1. Alphabetic
  2. By Inheritance
Inherited
  1. RangePartitioner
  2. Partitioner
  3. Serializable
  4. Serializable
  5. AnyRef
  6. Any
  1. Hide All
  2. Show All
Visibility
  1. Public
  2. All

Instance Constructors

  1. new RangePartitioner(partitions: Int, rdd: RDD[_ <: Product2[K, V]], ascending: Boolean)(implicit arg0: Ordering[K], arg1: ClassTag[K])
  2. new RangePartitioner(partitions: Int, rdd: RDD[_ <: Product2[K, V]], ascending: Boolean = true, samplePointsPerPartitionHint: Int = 20)(implicit arg0: Ordering[K], arg1: ClassTag[K])

Value Members

  1. def equals(other: Any): Boolean
    Definition Classes
    RangePartitioner → AnyRef → Any
  2. def getPartition(key: Any): Int
    Definition Classes
    RangePartitionerPartitioner
  3. def hashCode(): Int
    Definition Classes
    RangePartitioner → AnyRef → Any
  4. def numPartitions: Int
    Definition Classes
    RangePartitionerPartitioner
  5. val samplePointsPerPartitionHint: Int