public class JavaNewHadoopRDD<K,V> extends JavaPairRDD<K,V>
| Constructor and Description |
|---|
JavaNewHadoopRDD(NewHadoopRDD<K,V> rdd,
scala.reflect.ClassTag<K> kClassTag,
scala.reflect.ClassTag<V> vClassTag) |
| Modifier and Type | Method and Description |
|---|---|
scala.reflect.ClassTag<K> |
kClassTag() |
<R> JavaRDD<R> |
mapPartitionsWithInputSplit(Function2<org.apache.hadoop.mapreduce.InputSplit,java.util.Iterator<scala.Tuple2<K,V>>,java.util.Iterator<R>> f,
boolean preservesPartitioning)
Maps over a partition, providing the InputSplit that was used as the base of the partition.
|
scala.reflect.ClassTag<V> |
vClassTag() |
aggregateByKey, aggregateByKey, aggregateByKey, cache, classTag, coalesce, coalesce, cogroup, cogroup, cogroup, cogroup, cogroup, cogroup, cogroup, cogroup, cogroup, collectAsMap, combineByKey, combineByKey, combineByKey, combineByKey, countApproxDistinctByKey, countApproxDistinctByKey, countApproxDistinctByKey, countByKey, countByKeyApprox, countByKeyApprox, distinct, distinct, filter, filterByRange, filterByRange, first, flatMapValues, foldByKey, foldByKey, foldByKey, fromJavaRDD, fromRDD, fullOuterJoin, fullOuterJoin, fullOuterJoin, groupByKey, groupByKey, groupByKey, groupWith, groupWith, groupWith, intersection, join, join, join, keys, leftOuterJoin, leftOuterJoin, leftOuterJoin, lookup, mapValues, partitionBy, persist, rdd, reduceByKey, reduceByKey, reduceByKey, reduceByKeyLocally, repartition, repartitionAndSortWithinPartitions, repartitionAndSortWithinPartitions, rightOuterJoin, rightOuterJoin, rightOuterJoin, sample, sample, sampleByKey, sampleByKey, sampleByKeyExact, sampleByKeyExact, saveAsHadoopDataset, saveAsHadoopFile, saveAsHadoopFile, saveAsHadoopFile, saveAsNewAPIHadoopDataset, saveAsNewAPIHadoopFile, saveAsNewAPIHadoopFile, setName, sortByKey, sortByKey, sortByKey, sortByKey, sortByKey, sortByKey, subtract, subtract, subtract, subtractByKey, subtractByKey, subtractByKey, toRDD, union, unpersist, unpersist, values, wrapRDDequals, getClass, hashCode, notify, notifyAll, toString, wait, wait, waitaggregate, cartesian, checkpoint, collect, collectAsync, collectPartitions, context, count, countApprox, countApprox, countApproxDistinct, countAsync, countByValue, countByValueApprox, countByValueApprox, flatMap, flatMapToDouble, flatMapToPair, fold, foreach, foreachAsync, foreachPartition, foreachPartitionAsync, getCheckpointFile, getNumPartitions, getStorageLevel, glom, groupBy, groupBy, id, isCheckpointed, isEmpty, iterator, keyBy, map, mapPartitions, mapPartitions, mapPartitionsToDouble, mapPartitionsToDouble, mapPartitionsToPair, mapPartitionsToPair, mapPartitionsWithIndex, mapToDouble, mapToPair, max, min, name, partitioner, partitions, pipe, pipe, pipe, pipe, pipe, reduce, saveAsObjectFile, saveAsTextFile, saveAsTextFile, take, takeAsync, takeOrdered, takeOrdered, takeSample, takeSample, toDebugString, toLocalIterator, top, top, treeAggregate, treeAggregate, treeAggregate, treeReduce, treeReduce, zip, zipPartitions, zipWithIndex, zipWithUniqueIdpublic JavaNewHadoopRDD(NewHadoopRDD<K,V> rdd, scala.reflect.ClassTag<K> kClassTag, scala.reflect.ClassTag<V> vClassTag)
public scala.reflect.ClassTag<K> kClassTag()
kClassTag in class JavaPairRDD<K,V>public <R> JavaRDD<R> mapPartitionsWithInputSplit(Function2<org.apache.hadoop.mapreduce.InputSplit,java.util.Iterator<scala.Tuple2<K,V>>,java.util.Iterator<R>> f, boolean preservesPartitioning)
public scala.reflect.ClassTag<V> vClassTag()
vClassTag in class JavaPairRDD<K,V>