org.apache.spark.SparkContext.accumulable(R, AccumulableParam<R, T>)
|
org.apache.spark.SparkContext.accumulable(R, String, AccumulableParam<R, T>)
|
org.apache.spark.api.java.JavaSparkContext.accumulable(T, AccumulableParam<T, R>)
|
org.apache.spark.api.java.JavaSparkContext.accumulable(T, String, AccumulableParam<T, R>)
|
org.apache.spark.SparkContext.accumulableCollection(R, Function1<R, Growable<T>>, ClassTag<R>)
|
org.apache.spark.api.java.JavaSparkContext.accumulator(double)
|
org.apache.spark.api.java.JavaSparkContext.accumulator(double, String)
|
org.apache.spark.api.java.JavaSparkContext.accumulator(int)
|
org.apache.spark.api.java.JavaSparkContext.accumulator(int, String)
|
org.apache.spark.SparkContext.accumulator(T, AccumulatorParam<T>)
|
org.apache.spark.api.java.JavaSparkContext.accumulator(T, AccumulatorParam<T>)
|
org.apache.spark.SparkContext.accumulator(T, String, AccumulatorParam<T>)
|
org.apache.spark.api.java.JavaSparkContext.accumulator(T, String, AccumulatorParam<T>)
|
org.apache.spark.sql.streaming.ProcessingTime.apply(Duration)
|
org.apache.spark.scheduler.AccumulableInfo.apply(long, String, Option<String>, String)
|
org.apache.spark.scheduler.AccumulableInfo.apply(long, String, Option<String>, String, boolean)
|
org.apache.spark.scheduler.AccumulableInfo.apply(long, String, String)
|
org.apache.spark.sql.streaming.ProcessingTime.apply(String)
|
org.apache.spark.sql.SQLContext.applySchema(JavaRDD<?>, Class<?>)
|
org.apache.spark.sql.SQLContext.applySchema(JavaRDD<Row>, StructType)
|
org.apache.spark.sql.SQLContext.applySchema(RDD<?>, Class<?>)
|
org.apache.spark.sql.SQLContext.applySchema(RDD<Row>, StructType)
|
org.apache.spark.sql.functions.approxCountDistinct(Column)
|
org.apache.spark.sql.functions.approxCountDistinct(Column, double)
|
org.apache.spark.sql.functions.approxCountDistinct(String)
|
org.apache.spark.sql.functions.approxCountDistinct(String, double)
|
org.apache.spark.sql.SQLContext.clearActive()
|
org.apache.spark.sql.streaming.ProcessingTime.create(long, TimeUnit)
|
org.apache.spark.sql.streaming.ProcessingTime.create(String)
|
org.apache.spark.sql.SQLContext.createExternalTable(String, String)
|
org.apache.spark.sql.catalog.Catalog.createExternalTable(String, String)
|
org.apache.spark.sql.SQLContext.createExternalTable(String, String, Map<String, String>)
|
org.apache.spark.sql.SQLContext.createExternalTable(String, String, Map<String, String>)
|
org.apache.spark.sql.catalog.Catalog.createExternalTable(String, String, Map<String, String>)
|
org.apache.spark.sql.catalog.Catalog.createExternalTable(String, String, Map<String, String>)
|
org.apache.spark.sql.SQLContext.createExternalTable(String, String, String)
|
org.apache.spark.sql.catalog.Catalog.createExternalTable(String, String, String)
|
org.apache.spark.sql.SQLContext.createExternalTable(String, String, StructType, Map<String, String>)
|
org.apache.spark.sql.SQLContext.createExternalTable(String, String, StructType, Map<String, String>)
|
org.apache.spark.sql.catalog.Catalog.createExternalTable(String, String, StructType, Map<String, String>)
|
org.apache.spark.sql.catalog.Catalog.createExternalTable(String, String, StructType, Map<String, String>)
|
org.apache.spark.streaming.kinesis.KinesisUtils.createStream(JavaStreamingContext, String, String, String, String, InitialPositionInStream, Duration, StorageLevel)
|
org.apache.spark.streaming.kinesis.KinesisUtils.createStream(JavaStreamingContext, String, String, String, String, InitialPositionInStream, Duration, StorageLevel, Function<Record, T>, Class<T>)
|
org.apache.spark.streaming.kinesis.KinesisUtils.createStream(JavaStreamingContext, String, String, String, String, InitialPositionInStream, Duration, StorageLevel, Function<Record, T>, Class<T>, String, String)
|
org.apache.spark.streaming.kinesis.KinesisUtils.createStream(JavaStreamingContext, String, String, String, String, InitialPositionInStream, Duration, StorageLevel, Function<Record, T>, Class<T>, String, String, String, String, String)
|
org.apache.spark.streaming.kinesis.KinesisUtils.createStream(JavaStreamingContext, String, String, String, String, InitialPositionInStream, Duration, StorageLevel, String, String)
|
org.apache.spark.streaming.kinesis.KinesisUtils.createStream(StreamingContext, String, String, String, String, InitialPositionInStream, Duration, StorageLevel)
|
org.apache.spark.streaming.kinesis.KinesisUtils.createStream(StreamingContext, String, String, String, String, InitialPositionInStream, Duration, StorageLevel, Function1<Record, T>, ClassTag<T>)
|
org.apache.spark.streaming.kinesis.KinesisUtils.createStream(StreamingContext, String, String, String, String, InitialPositionInStream, Duration, StorageLevel, Function1<Record, T>, String, String, ClassTag<T>)
|
org.apache.spark.streaming.kinesis.KinesisUtils.createStream(StreamingContext, String, String, String, String, InitialPositionInStream, Duration, StorageLevel, Function1<Record, T>, String, String, String, String, String, ClassTag<T>)
|
org.apache.spark.streaming.kinesis.KinesisUtils.createStream(StreamingContext, String, String, String, String, InitialPositionInStream, Duration, StorageLevel, String, String)
|
org.apache.spark.api.java.JavaSparkContext.doubleAccumulator(double)
|
org.apache.spark.api.java.JavaSparkContext.doubleAccumulator(double, String)
|
org.apache.spark.sql.Dataset.explode(Seq<Column>, Function1<Row, TraversableOnce<A>>, TypeTags.TypeTag<A>)
|
org.apache.spark.sql.Dataset.explode(String, String, Function1<A, TraversableOnce<B>>, TypeTags.TypeTag<B>)
|
org.apache.spark.mllib.evaluation.MulticlassMetrics.fMeasure()
|
org.apache.spark.SparkContext.getExecutorStorageStatus()
|
org.apache.spark.sql.SQLContext.getOrCreate(SparkContext)
|
org.apache.spark.mllib.clustering.KMeans.getRuns()
|
org.apache.spark.api.java.JavaSparkContext.intAccumulator(int)
|
org.apache.spark.api.java.JavaSparkContext.intAccumulator(int, String)
|
org.apache.spark.TaskContext.isRunningLocally()
|
org.apache.spark.sql.SQLContext.jdbc(String, String)
|
org.apache.spark.sql.SQLContext.jdbc(String, String, String[])
|
org.apache.spark.sql.SQLContext.jdbc(String, String, String, long, long, int)
|
org.apache.spark.sql.DataFrameReader.json(JavaRDD<String>)
|
org.apache.spark.sql.DataFrameReader.json(RDD<String>)
|
org.apache.spark.sql.SQLContext.jsonFile(String)
|
org.apache.spark.sql.SQLContext.jsonFile(String, double)
|
org.apache.spark.sql.SQLContext.jsonFile(String, StructType)
|
org.apache.spark.sql.SQLContext.jsonRDD(JavaRDD<String>)
|
org.apache.spark.sql.SQLContext.jsonRDD(JavaRDD<String>, double)
|
org.apache.spark.sql.SQLContext.jsonRDD(JavaRDD<String>, StructType)
|
org.apache.spark.sql.SQLContext.jsonRDD(RDD<String>)
|
org.apache.spark.sql.SQLContext.jsonRDD(RDD<String>, double)
|
org.apache.spark.sql.SQLContext.jsonRDD(RDD<String>, StructType)
|
org.apache.spark.sql.SQLContext.load(String)
|
org.apache.spark.sql.SQLContext.load(String, Map<String, String>)
|
org.apache.spark.sql.SQLContext.load(String, Map<String, String>)
|
org.apache.spark.sql.SQLContext.load(String, String)
|
org.apache.spark.sql.SQLContext.load(String, StructType, Map<String, String>)
|
org.apache.spark.sql.SQLContext.load(String, StructType, Map<String, String>)
|
org.apache.spark.sql.functions.monotonicallyIncreasingId()
|
org.apache.spark.sql.SQLImplicits.newBooleanSeqEncoder()
|
org.apache.spark.sql.SQLImplicits.newByteSeqEncoder()
|
org.apache.spark.sql.SQLImplicits.newDoubleSeqEncoder()
|
org.apache.spark.sql.SQLImplicits.newFloatSeqEncoder()
|
org.apache.spark.sql.SQLImplicits.newIntSeqEncoder()
|
org.apache.spark.sql.SQLImplicits.newLongSeqEncoder()
|
org.apache.spark.sql.SQLImplicits.newProductSeqEncoder(TypeTags.TypeTag<A>)
|
org.apache.spark.sql.SQLImplicits.newShortSeqEncoder()
|
org.apache.spark.sql.SQLImplicits.newStringSeqEncoder()
|
org.apache.spark.sql.SQLContext.parquetFile(Seq<String>)
|
org.apache.spark.sql.SQLContext.parquetFile(String...)
|
org.apache.spark.mllib.evaluation.MulticlassMetrics.precision()
|
org.apache.spark.mllib.evaluation.MulticlassMetrics.recall()
|
org.apache.spark.sql.Dataset.registerTempTable(String)
|
org.apache.spark.sql.SQLContext.setActive(SQLContext)
|
org.apache.spark.mllib.clustering.KMeans.setRuns(int)
|
org.apache.spark.sql.functions.toDegrees(Column)
|
org.apache.spark.sql.functions.toDegrees(String)
|
org.apache.spark.sql.functions.toRadians(Column)
|
org.apache.spark.sql.functions.toRadians(String)
|
org.apache.spark.mllib.clustering.KMeans.train(RDD<Vector>, int, int, int)
|
org.apache.spark.mllib.clustering.KMeans.train(RDD<Vector>, int, int, int, String)
|
org.apache.spark.mllib.clustering.KMeans.train(RDD<Vector>, int, int, int, String, long)
|
org.apache.spark.sql.Dataset.unionAll(Dataset<T>)
|