org.apache.spark.sql.SQLContext.applySchema(JavaRDD<?>, Class<?>)
|
org.apache.spark.sql.SQLContext.applySchema(JavaRDD<Row>, StructType)
|
org.apache.spark.sql.SQLContext.applySchema(RDD<?>, Class<?>)
|
org.apache.spark.sql.SQLContext.applySchema(RDD<Row>, StructType)
|
org.apache.spark.sql.functions.approxCountDistinct(Column)
|
org.apache.spark.sql.functions.approxCountDistinct(Column, double)
|
org.apache.spark.sql.functions.approxCountDistinct(String)
|
org.apache.spark.sql.functions.approxCountDistinct(String, double)
|
org.apache.spark.sql.functions.bitwiseNOT(Column)
|
org.apache.spark.status.api.v1.ExecutorSummary.blacklistedInStages()
|
org.apache.spark.sql.functions.callUDF(String, Seq<Column>)
|
org.apache.spark.sql.SQLContext.clearActive()
|
org.apache.spark.sql.jdbc.JdbcDialect.compileAggregate(AggregateFunc)
|
org.apache.spark.ml.clustering.BisectingKMeansModel.computeCost(Dataset<?>)
|
org.apache.spark.sql.SQLContext.createExternalTable(String, String)
|
org.apache.spark.sql.catalog.Catalog.createExternalTable(String, String)
|
org.apache.spark.sql.SQLContext.createExternalTable(String, String, Map<String, String>)
|
org.apache.spark.sql.SQLContext.createExternalTable(String, String, Map<String, String>)
|
org.apache.spark.sql.catalog.Catalog.createExternalTable(String, String, Map<String, String>)
|
org.apache.spark.sql.catalog.Catalog.createExternalTable(String, String, Map<String, String>)
|
org.apache.spark.sql.SQLContext.createExternalTable(String, String, String)
|
org.apache.spark.sql.catalog.Catalog.createExternalTable(String, String, String)
|
org.apache.spark.sql.SQLContext.createExternalTable(String, String, StructType, Map<String, String>)
|
org.apache.spark.sql.SQLContext.createExternalTable(String, String, StructType, Map<String, String>)
|
org.apache.spark.sql.catalog.Catalog.createExternalTable(String, String, StructType, Map<String, String>)
|
org.apache.spark.sql.catalog.Catalog.createExternalTable(String, String, StructType, Map<String, String>)
|
org.apache.spark.sql.Dataset.explode(Seq<Column>, Function1<Row, TraversableOnce<A>>, TypeTags.TypeTag<A>)
|
org.apache.spark.sql.Dataset.explode(String, String, Function1<A, TraversableOnce<B>>, TypeTags.TypeTag<B>)
|
org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder.getAttributes() |
org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.getAttributes() |
org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder.getAttributes() |
org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgressOrBuilder.getCustomMetrics() |
org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.getCustomMetrics() |
org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder.getCustomMetrics() |
org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder.getDurationMs() |
org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.getDurationMs() |
org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder.getDurationMs() |
org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder.getEventTime() |
org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.getEventTime() |
org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder.getEventTime() |
org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder.getExecutorLogs() |
org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.getExecutorLogs() |
org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder.getExecutorLogs() |
org.apache.spark.status.protobuf.StoreTypes.TaskDataOrBuilder.getExecutorLogs() |
org.apache.spark.status.protobuf.StoreTypes.TaskData.getExecutorLogs() |
org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder.getExecutorLogs() |
org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfoOrBuilder.getExecutorResources() |
org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.getExecutorResources() |
org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.Builder.getExecutorResources() |
org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder.getExecutorSummary() |
org.apache.spark.status.protobuf.StoreTypes.StageData.getExecutorSummary() |
org.apache.spark.status.protobuf.StoreTypes.StageData.Builder.getExecutorSummary() |
org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder.getJobs() |
org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.getJobs() |
org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder.getJobs() |
org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder.getJobsValue() |
org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.getJobsValue() |
org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder.getJobsValue() |
org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder.getKilledTasksSummary() |
org.apache.spark.status.protobuf.StoreTypes.StageData.getKilledTasksSummary() |
org.apache.spark.status.protobuf.StoreTypes.StageData.Builder.getKilledTasksSummary() |
org.apache.spark.status.protobuf.StoreTypes.JobDataOrBuilder.getKillTasksSummary() |
org.apache.spark.status.protobuf.StoreTypes.JobData.getKillTasksSummary() |
org.apache.spark.status.protobuf.StoreTypes.JobData.Builder.getKillTasksSummary() |
org.apache.spark.status.protobuf.StoreTypes.StageDataWrapperOrBuilder.getLocality() |
org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper.getLocality() |
org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper.Builder.getLocality() |
org.apache.spark.status.protobuf.StoreTypes.ExecutorMetricsOrBuilder.getMetrics() |
org.apache.spark.status.protobuf.StoreTypes.ExecutorMetrics.getMetrics() |
org.apache.spark.status.protobuf.StoreTypes.ExecutorMetrics.Builder.getMetrics() |
org.apache.spark.status.protobuf.StoreTypes.SourceProgressOrBuilder.getMetrics() |
org.apache.spark.status.protobuf.StoreTypes.SourceProgress.getMetrics() |
org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder.getMetrics() |
org.apache.spark.status.protobuf.StoreTypes.SinkProgressOrBuilder.getMetrics() |
org.apache.spark.status.protobuf.StoreTypes.SinkProgress.getMetrics() |
org.apache.spark.status.protobuf.StoreTypes.SinkProgress.Builder.getMetrics() |
org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder.getMetricValues() |
org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.getMetricValues() |
org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder.getMetricValues() |
org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIDataOrBuilder.getModifiedConfigs() |
org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.getModifiedConfigs() |
org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder.getModifiedConfigs() |
org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder.getMutableAttributes() |
org.apache.spark.status.protobuf.StoreTypes.StateOperatorProgress.Builder.getMutableCustomMetrics() |
org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder.getMutableDurationMs() |
org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder.getMutableEventTime() |
org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder.getMutableExecutorLogs() |
org.apache.spark.status.protobuf.StoreTypes.TaskData.Builder.getMutableExecutorLogs() |
org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.Builder.getMutableExecutorResources() |
org.apache.spark.status.protobuf.StoreTypes.StageData.Builder.getMutableExecutorSummary() |
org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder.getMutableJobs() |
org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder.getMutableJobsValue() |
org.apache.spark.status.protobuf.StoreTypes.StageData.Builder.getMutableKilledTasksSummary() |
org.apache.spark.status.protobuf.StoreTypes.JobData.Builder.getMutableKillTasksSummary() |
org.apache.spark.status.protobuf.StoreTypes.StageDataWrapper.Builder.getMutableLocality() |
org.apache.spark.status.protobuf.StoreTypes.ExecutorMetrics.Builder.getMutableMetrics() |
org.apache.spark.status.protobuf.StoreTypes.SourceProgress.Builder.getMutableMetrics() |
org.apache.spark.status.protobuf.StoreTypes.SinkProgress.Builder.getMutableMetrics() |
org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder.getMutableMetricValues() |
org.apache.spark.status.protobuf.StoreTypes.SQLExecutionUIData.Builder.getMutableModifiedConfigs() |
org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder.getMutableObservedMetrics() |
org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder.getMutableProcessLogs() |
org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder.getMutableResources() |
org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.Builder.getMutableTaskResources() |
org.apache.spark.status.protobuf.StoreTypes.StageData.Builder.getMutableTasks() |
org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgressOrBuilder.getObservedMetrics() |
org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.getObservedMetrics() |
org.apache.spark.status.protobuf.StoreTypes.StreamingQueryProgress.Builder.getObservedMetrics() |
org.apache.spark.sql.SQLContext.getOrCreate(SparkContext)
|
org.apache.spark.status.protobuf.StoreTypes.ProcessSummaryOrBuilder.getProcessLogs() |
org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.getProcessLogs() |
org.apache.spark.status.protobuf.StoreTypes.ProcessSummary.Builder.getProcessLogs() |
org.apache.spark.status.protobuf.StoreTypes.ExecutorSummaryOrBuilder.getResources() |
org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.getResources() |
org.apache.spark.status.protobuf.StoreTypes.ExecutorSummary.Builder.getResources() |
org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfoOrBuilder.getTaskResources() |
org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.getTaskResources() |
org.apache.spark.status.protobuf.StoreTypes.ResourceProfileInfo.Builder.getTaskResources() |
org.apache.spark.status.protobuf.StoreTypes.StageDataOrBuilder.getTasks() |
org.apache.spark.status.protobuf.StoreTypes.StageData.getTasks() |
org.apache.spark.status.protobuf.StoreTypes.StageData.Builder.getTasks() |
org.apache.spark.status.api.v1.ExecutorSummary.isBlacklisted()
|
org.apache.spark.status.api.v1.ExecutorStageSummary.isBlacklistedForStage()
|
org.apache.spark.sql.SQLContext.jdbc(String, String)
|
org.apache.spark.sql.SQLContext.jdbc(String, String, String[])
|
org.apache.spark.sql.SQLContext.jdbc(String, String, String, long, long, int)
|
org.apache.spark.sql.DataFrameReader.json(JavaRDD<String>)
|
org.apache.spark.sql.DataFrameReader.json(RDD<String>)
|
org.apache.spark.sql.SQLContext.jsonFile(String)
|
org.apache.spark.sql.SQLContext.jsonFile(String, double)
|
org.apache.spark.sql.SQLContext.jsonFile(String, StructType)
|
org.apache.spark.sql.SQLContext.jsonRDD(JavaRDD<String>)
|
org.apache.spark.sql.SQLContext.jsonRDD(JavaRDD<String>, double)
|
org.apache.spark.sql.SQLContext.jsonRDD(JavaRDD<String>, StructType)
|
org.apache.spark.sql.SQLContext.jsonRDD(RDD<String>)
|
org.apache.spark.sql.SQLContext.jsonRDD(RDD<String>, double)
|
org.apache.spark.sql.SQLContext.jsonRDD(RDD<String>, StructType)
|
org.apache.spark.ml.feature.StringIndexerModel.labels()
|
org.apache.spark.sql.SQLContext.load(String)
|
org.apache.spark.sql.SQLContext.load(String, Map<String, String>)
|
org.apache.spark.sql.SQLContext.load(String, Map<String, String>)
|
org.apache.spark.sql.SQLContext.load(String, String)
|
org.apache.spark.sql.SQLContext.load(String, StructType, Map<String, String>)
|
org.apache.spark.sql.SQLContext.load(String, StructType, Map<String, String>)
|
org.apache.spark.sql.functions.monotonicallyIncreasingId()
|
org.apache.spark.sql.SQLImplicits.newBooleanSeqEncoder()
|
org.apache.spark.sql.SQLImplicits.newByteSeqEncoder()
|
org.apache.spark.sql.SQLImplicits.newDoubleSeqEncoder()
|
org.apache.spark.sql.SQLImplicits.newFloatSeqEncoder()
|
org.apache.spark.sql.SQLImplicits.newIntSeqEncoder()
|
org.apache.spark.sql.SQLImplicits.newLongSeqEncoder()
|
org.apache.spark.sql.SQLImplicits.newProductSeqEncoder(TypeTags.TypeTag<A>)
|
org.apache.spark.sql.SQLImplicits.newShortSeqEncoder()
|
org.apache.spark.sql.SQLImplicits.newStringSeqEncoder()
|
org.apache.spark.sql.streaming.Trigger.Once()
|
org.apache.spark.scheduler.SparkListenerInterface.onExecutorBlacklisted(SparkListenerExecutorBlacklisted)
|
org.apache.spark.scheduler.SparkListenerInterface.onExecutorBlacklistedForStage(SparkListenerExecutorBlacklistedForStage)
|
org.apache.spark.scheduler.SparkListenerInterface.onExecutorUnblacklisted(SparkListenerExecutorUnblacklisted)
|
org.apache.spark.scheduler.SparkListenerInterface.onNodeBlacklisted(SparkListenerNodeBlacklisted)
|
org.apache.spark.scheduler.SparkListenerInterface.onNodeBlacklistedForStage(SparkListenerNodeBlacklistedForStage)
|
org.apache.spark.scheduler.SparkListenerInterface.onNodeUnblacklisted(SparkListenerNodeUnblacklisted)
|
org.apache.spark.sql.SQLContext.parquetFile(Seq<String>)
|
org.apache.spark.sql.SQLContext.parquetFile(String...)
|
org.apache.spark.sql.UDFRegistration.register(String, UserDefinedAggregateFunction)
|
org.apache.spark.sql.Dataset.registerTempTable(String)
|
org.apache.spark.mllib.evaluation.BinaryClassificationMetrics.scoreLabelsWeight()
|
org.apache.spark.sql.SQLContext.setActive(SQLContext)
|
org.apache.spark.sql.functions.shiftLeft(Column, int)
|
org.apache.spark.sql.functions.shiftRight(Column, int)
|
org.apache.spark.sql.functions.shiftRightUnsigned(Column, int)
|
org.apache.spark.sql.functions.sumDistinct(Column)
|
org.apache.spark.sql.functions.sumDistinct(String)
|
org.apache.spark.sql.functions.toDegrees(Column)
|
org.apache.spark.sql.functions.toDegrees(String)
|
org.apache.spark.sql.functions.toRadians(Column)
|
org.apache.spark.sql.functions.toRadians(String)
|
org.apache.spark.sql.functions.udf(Object, DataType)
|
org.apache.spark.status.protobuf.StoreTypes.JobExecutionStatus.valueOf(int)
|
org.apache.spark.status.protobuf.StoreTypes.DeterministicLevel.valueOf(int)
|
org.apache.spark.status.protobuf.StoreTypes.StageStatus.valueOf(int)
|
org.apache.spark.status.protobuf.StoreTypes.SparkPlanGraphNodeWrapper.WrapperCase.valueOf(int)
|